Oct 31 02:42:30.046663 kernel: Linux version 6.6.113-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Oct 30 22:59:39 -00 2025 Oct 31 02:42:30.046701 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=950876ad7bc3e9634b7585a81697da4ef03ac6558969e5c002165369dd7c7885 Oct 31 02:42:30.046715 kernel: BIOS-provided physical RAM map: Oct 31 02:42:30.046732 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Oct 31 02:42:30.046742 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Oct 31 02:42:30.046752 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Oct 31 02:42:30.046764 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdbfff] usable Oct 31 02:42:30.046775 kernel: BIOS-e820: [mem 0x000000007ffdc000-0x000000007fffffff] reserved Oct 31 02:42:30.046786 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Oct 31 02:42:30.046796 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Oct 31 02:42:30.046807 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Oct 31 02:42:30.046817 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Oct 31 02:42:30.046833 kernel: NX (Execute Disable) protection: active Oct 31 02:42:30.046844 kernel: APIC: Static calls initialized Oct 31 02:42:30.046857 kernel: SMBIOS 2.8 present. Oct 31 02:42:30.046869 kernel: DMI: Red Hat KVM/RHEL-AV, BIOS 1.13.0-2.module_el8.5.0+2608+72063365 04/01/2014 Oct 31 02:42:30.046880 kernel: Hypervisor detected: KVM Oct 31 02:42:30.046896 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Oct 31 02:42:30.046908 kernel: kvm-clock: using sched offset of 4438629266 cycles Oct 31 02:42:30.046921 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Oct 31 02:42:30.046933 kernel: tsc: Detected 2500.032 MHz processor Oct 31 02:42:30.046944 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Oct 31 02:42:30.046956 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Oct 31 02:42:30.046968 kernel: last_pfn = 0x7ffdc max_arch_pfn = 0x400000000 Oct 31 02:42:30.046979 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Oct 31 02:42:30.046991 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Oct 31 02:42:30.047007 kernel: Using GB pages for direct mapping Oct 31 02:42:30.047050 kernel: ACPI: Early table checksum verification disabled Oct 31 02:42:30.047063 kernel: ACPI: RSDP 0x00000000000F5AA0 000014 (v00 BOCHS ) Oct 31 02:42:30.047075 kernel: ACPI: RSDT 0x000000007FFE47A5 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047086 kernel: ACPI: FACP 0x000000007FFE438D 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047098 kernel: ACPI: DSDT 0x000000007FFDFD80 00460D (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047109 kernel: ACPI: FACS 0x000000007FFDFD40 000040 Oct 31 02:42:30.047120 kernel: ACPI: APIC 0x000000007FFE4481 0000F0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047132 kernel: ACPI: SRAT 0x000000007FFE4571 0001D0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047149 kernel: ACPI: MCFG 0x000000007FFE4741 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047161 kernel: ACPI: WAET 0x000000007FFE477D 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 31 02:42:30.047172 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe438d-0x7ffe4480] Oct 31 02:42:30.047183 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffdfd80-0x7ffe438c] Oct 31 02:42:30.047194 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffdfd40-0x7ffdfd7f] Oct 31 02:42:30.047212 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe4481-0x7ffe4570] Oct 31 02:42:30.047224 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe4571-0x7ffe4740] Oct 31 02:42:30.047240 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe4741-0x7ffe477c] Oct 31 02:42:30.047252 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe477d-0x7ffe47a4] Oct 31 02:42:30.047264 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Oct 31 02:42:30.047287 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Oct 31 02:42:30.047311 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 Oct 31 02:42:30.047323 kernel: SRAT: PXM 0 -> APIC 0x03 -> Node 0 Oct 31 02:42:30.047335 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 Oct 31 02:42:30.047347 kernel: SRAT: PXM 0 -> APIC 0x05 -> Node 0 Oct 31 02:42:30.047364 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 Oct 31 02:42:30.047377 kernel: SRAT: PXM 0 -> APIC 0x07 -> Node 0 Oct 31 02:42:30.047389 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 Oct 31 02:42:30.047400 kernel: SRAT: PXM 0 -> APIC 0x09 -> Node 0 Oct 31 02:42:30.047412 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 Oct 31 02:42:30.047424 kernel: SRAT: PXM 0 -> APIC 0x0b -> Node 0 Oct 31 02:42:30.047436 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 Oct 31 02:42:30.047448 kernel: SRAT: PXM 0 -> APIC 0x0d -> Node 0 Oct 31 02:42:30.047460 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 Oct 31 02:42:30.047477 kernel: SRAT: PXM 0 -> APIC 0x0f -> Node 0 Oct 31 02:42:30.047489 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Oct 31 02:42:30.047501 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Oct 31 02:42:30.047513 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x20800fffff] hotplug Oct 31 02:42:30.047525 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdbfff] -> [mem 0x00000000-0x7ffdbfff] Oct 31 02:42:30.047538 kernel: NODE_DATA(0) allocated [mem 0x7ffd6000-0x7ffdbfff] Oct 31 02:42:30.047550 kernel: Zone ranges: Oct 31 02:42:30.047562 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Oct 31 02:42:30.047574 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdbfff] Oct 31 02:42:30.047591 kernel: Normal empty Oct 31 02:42:30.047603 kernel: Movable zone start for each node Oct 31 02:42:30.047615 kernel: Early memory node ranges Oct 31 02:42:30.047627 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Oct 31 02:42:30.047639 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdbfff] Oct 31 02:42:30.047652 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdbfff] Oct 31 02:42:30.047664 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 31 02:42:30.047676 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Oct 31 02:42:30.047688 kernel: On node 0, zone DMA32: 36 pages in unavailable ranges Oct 31 02:42:30.047700 kernel: ACPI: PM-Timer IO Port: 0x608 Oct 31 02:42:30.047717 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Oct 31 02:42:30.047729 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Oct 31 02:42:30.047741 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Oct 31 02:42:30.047753 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Oct 31 02:42:30.047765 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Oct 31 02:42:30.047777 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Oct 31 02:42:30.047789 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Oct 31 02:42:30.047801 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Oct 31 02:42:30.047813 kernel: TSC deadline timer available Oct 31 02:42:30.047830 kernel: smpboot: Allowing 16 CPUs, 14 hotplug CPUs Oct 31 02:42:30.047843 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Oct 31 02:42:30.047855 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Oct 31 02:42:30.047867 kernel: Booting paravirtualized kernel on KVM Oct 31 02:42:30.047879 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Oct 31 02:42:30.047892 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Oct 31 02:42:30.047904 kernel: percpu: Embedded 58 pages/cpu s196712 r8192 d32664 u262144 Oct 31 02:42:30.047916 kernel: pcpu-alloc: s196712 r8192 d32664 u262144 alloc=1*2097152 Oct 31 02:42:30.047928 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Oct 31 02:42:30.047945 kernel: kvm-guest: PV spinlocks enabled Oct 31 02:42:30.047957 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Oct 31 02:42:30.047971 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=950876ad7bc3e9634b7585a81697da4ef03ac6558969e5c002165369dd7c7885 Oct 31 02:42:30.047983 kernel: random: crng init done Oct 31 02:42:30.047996 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 31 02:42:30.048008 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Oct 31 02:42:30.048144 kernel: Fallback order for Node 0: 0 Oct 31 02:42:30.048177 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515804 Oct 31 02:42:30.048209 kernel: Policy zone: DMA32 Oct 31 02:42:30.048222 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 31 02:42:30.048234 kernel: software IO TLB: area num 16. Oct 31 02:42:30.048247 kernel: Memory: 1901528K/2096616K available (12288K kernel code, 2288K rwdata, 22748K rodata, 42880K init, 2320K bss, 194828K reserved, 0K cma-reserved) Oct 31 02:42:30.048259 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Oct 31 02:42:30.048272 kernel: Kernel/User page tables isolation: enabled Oct 31 02:42:30.048284 kernel: ftrace: allocating 37980 entries in 149 pages Oct 31 02:42:30.048308 kernel: ftrace: allocated 149 pages with 4 groups Oct 31 02:42:30.048321 kernel: Dynamic Preempt: voluntary Oct 31 02:42:30.048339 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 31 02:42:30.048352 kernel: rcu: RCU event tracing is enabled. Oct 31 02:42:30.048365 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Oct 31 02:42:30.048377 kernel: Trampoline variant of Tasks RCU enabled. Oct 31 02:42:30.048390 kernel: Rude variant of Tasks RCU enabled. Oct 31 02:42:30.048415 kernel: Tracing variant of Tasks RCU enabled. Oct 31 02:42:30.048432 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 31 02:42:30.048445 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Oct 31 02:42:30.048458 kernel: NR_IRQS: 33024, nr_irqs: 552, preallocated irqs: 16 Oct 31 02:42:30.048471 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 31 02:42:30.048484 kernel: Console: colour VGA+ 80x25 Oct 31 02:42:30.048496 kernel: printk: console [tty0] enabled Oct 31 02:42:30.048515 kernel: printk: console [ttyS0] enabled Oct 31 02:42:30.048528 kernel: ACPI: Core revision 20230628 Oct 31 02:42:30.048541 kernel: APIC: Switch to symmetric I/O mode setup Oct 31 02:42:30.048554 kernel: x2apic enabled Oct 31 02:42:30.048566 kernel: APIC: Switched APIC routing to: physical x2apic Oct 31 02:42:30.048584 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x240957bf147, max_idle_ns: 440795216753 ns Oct 31 02:42:30.048598 kernel: Calibrating delay loop (skipped) preset value.. 5000.06 BogoMIPS (lpj=2500032) Oct 31 02:42:30.048611 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Oct 31 02:42:30.048624 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Oct 31 02:42:30.048637 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Oct 31 02:42:30.048649 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Oct 31 02:42:30.048662 kernel: Spectre V2 : Mitigation: Retpolines Oct 31 02:42:30.048675 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Oct 31 02:42:30.048688 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Oct 31 02:42:30.048701 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Oct 31 02:42:30.048719 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Oct 31 02:42:30.048731 kernel: MDS: Mitigation: Clear CPU buffers Oct 31 02:42:30.048744 kernel: MMIO Stale Data: Unknown: No mitigations Oct 31 02:42:30.048757 kernel: SRBDS: Unknown: Dependent on hypervisor status Oct 31 02:42:30.048770 kernel: active return thunk: its_return_thunk Oct 31 02:42:30.048782 kernel: ITS: Mitigation: Aligned branch/return thunks Oct 31 02:42:30.048795 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Oct 31 02:42:30.048808 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Oct 31 02:42:30.048821 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Oct 31 02:42:30.048834 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Oct 31 02:42:30.048847 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Oct 31 02:42:30.048865 kernel: Freeing SMP alternatives memory: 32K Oct 31 02:42:30.048877 kernel: pid_max: default: 32768 minimum: 301 Oct 31 02:42:30.048890 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Oct 31 02:42:30.048903 kernel: landlock: Up and running. Oct 31 02:42:30.048916 kernel: SELinux: Initializing. Oct 31 02:42:30.048929 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 31 02:42:30.048941 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 31 02:42:30.048954 kernel: smpboot: CPU0: Intel Xeon E3-12xx v2 (Ivy Bridge, IBRS) (family: 0x6, model: 0x3a, stepping: 0x9) Oct 31 02:42:30.048967 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Oct 31 02:42:30.048980 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Oct 31 02:42:30.049002 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Oct 31 02:42:30.049016 kernel: Performance Events: unsupported p6 CPU model 58 no PMU driver, software events only. Oct 31 02:42:30.049041 kernel: signal: max sigframe size: 1776 Oct 31 02:42:30.049054 kernel: rcu: Hierarchical SRCU implementation. Oct 31 02:42:30.049067 kernel: rcu: Max phase no-delay instances is 400. Oct 31 02:42:30.049080 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Oct 31 02:42:30.049093 kernel: smp: Bringing up secondary CPUs ... Oct 31 02:42:30.049106 kernel: smpboot: x86: Booting SMP configuration: Oct 31 02:42:30.049119 kernel: .... node #0, CPUs: #1 Oct 31 02:42:30.049131 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 Oct 31 02:42:30.049150 kernel: smp: Brought up 1 node, 2 CPUs Oct 31 02:42:30.049164 kernel: smpboot: Max logical packages: 16 Oct 31 02:42:30.049177 kernel: smpboot: Total of 2 processors activated (10000.12 BogoMIPS) Oct 31 02:42:30.049189 kernel: devtmpfs: initialized Oct 31 02:42:30.049202 kernel: x86/mm: Memory block size: 128MB Oct 31 02:42:30.049215 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 31 02:42:30.049228 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Oct 31 02:42:30.049241 kernel: pinctrl core: initialized pinctrl subsystem Oct 31 02:42:30.049254 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 31 02:42:30.049271 kernel: audit: initializing netlink subsys (disabled) Oct 31 02:42:30.049285 kernel: audit: type=2000 audit(1761878548.220:1): state=initialized audit_enabled=0 res=1 Oct 31 02:42:30.049308 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 31 02:42:30.049322 kernel: thermal_sys: Registered thermal governor 'user_space' Oct 31 02:42:30.049335 kernel: cpuidle: using governor menu Oct 31 02:42:30.049348 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 31 02:42:30.049361 kernel: dca service started, version 1.12.1 Oct 31 02:42:30.049373 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Oct 31 02:42:30.049386 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Oct 31 02:42:30.049405 kernel: PCI: Using configuration type 1 for base access Oct 31 02:42:30.049418 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Oct 31 02:42:30.049431 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Oct 31 02:42:30.049444 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Oct 31 02:42:30.049457 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 31 02:42:30.049470 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Oct 31 02:42:30.049483 kernel: ACPI: Added _OSI(Module Device) Oct 31 02:42:30.049496 kernel: ACPI: Added _OSI(Processor Device) Oct 31 02:42:30.049513 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 31 02:42:30.049527 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Oct 31 02:42:30.049539 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Oct 31 02:42:30.049552 kernel: ACPI: Interpreter enabled Oct 31 02:42:30.049565 kernel: ACPI: PM: (supports S0 S5) Oct 31 02:42:30.049578 kernel: ACPI: Using IOAPIC for interrupt routing Oct 31 02:42:30.049591 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Oct 31 02:42:30.049604 kernel: PCI: Using E820 reservations for host bridge windows Oct 31 02:42:30.049617 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Oct 31 02:42:30.049630 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 31 02:42:30.049900 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Oct 31 02:42:30.050636 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Oct 31 02:42:30.050816 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Oct 31 02:42:30.050837 kernel: PCI host bridge to bus 0000:00 Oct 31 02:42:30.051042 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Oct 31 02:42:30.051208 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Oct 31 02:42:30.051392 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Oct 31 02:42:30.051551 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Oct 31 02:42:30.051732 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Oct 31 02:42:30.051890 kernel: pci_bus 0000:00: root bus resource [mem 0x20c0000000-0x28bfffffff window] Oct 31 02:42:30.052061 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 31 02:42:30.052277 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Oct 31 02:42:30.052499 kernel: pci 0000:00:01.0: [1013:00b8] type 00 class 0x030000 Oct 31 02:42:30.052683 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfa000000-0xfbffffff pref] Oct 31 02:42:30.052861 kernel: pci 0000:00:01.0: reg 0x14: [mem 0xfea50000-0xfea50fff] Oct 31 02:42:30.053093 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfea40000-0xfea4ffff pref] Oct 31 02:42:30.053410 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Oct 31 02:42:30.053625 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.053804 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfea51000-0xfea51fff] Oct 31 02:42:30.054014 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.054211 kernel: pci 0000:00:02.1: reg 0x10: [mem 0xfea52000-0xfea52fff] Oct 31 02:42:30.054425 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.054602 kernel: pci 0000:00:02.2: reg 0x10: [mem 0xfea53000-0xfea53fff] Oct 31 02:42:30.054842 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.055048 kernel: pci 0000:00:02.3: reg 0x10: [mem 0xfea54000-0xfea54fff] Oct 31 02:42:30.055303 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.055515 kernel: pci 0000:00:02.4: reg 0x10: [mem 0xfea55000-0xfea55fff] Oct 31 02:42:30.055755 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.055959 kernel: pci 0000:00:02.5: reg 0x10: [mem 0xfea56000-0xfea56fff] Oct 31 02:42:30.056224 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.056425 kernel: pci 0000:00:02.6: reg 0x10: [mem 0xfea57000-0xfea57fff] Oct 31 02:42:30.056634 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Oct 31 02:42:30.056828 kernel: pci 0000:00:02.7: reg 0x10: [mem 0xfea58000-0xfea58fff] Oct 31 02:42:30.057119 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Oct 31 02:42:30.057345 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc0c0-0xc0df] Oct 31 02:42:30.059333 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfea59000-0xfea59fff] Oct 31 02:42:30.059514 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfd000000-0xfd003fff 64bit pref] Oct 31 02:42:30.059690 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfea00000-0xfea3ffff pref] Oct 31 02:42:30.059920 kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 Oct 31 02:42:30.060123 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] Oct 31 02:42:30.060308 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfea5a000-0xfea5afff] Oct 31 02:42:30.060482 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfd004000-0xfd007fff 64bit pref] Oct 31 02:42:30.060693 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Oct 31 02:42:30.060888 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Oct 31 02:42:30.062177 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Oct 31 02:42:30.062407 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc0e0-0xc0ff] Oct 31 02:42:30.062582 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfea5b000-0xfea5bfff] Oct 31 02:42:30.062802 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Oct 31 02:42:30.062993 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Oct 31 02:42:30.064288 kernel: pci 0000:01:00.0: [1b36:000e] type 01 class 0x060400 Oct 31 02:42:30.064485 kernel: pci 0000:01:00.0: reg 0x10: [mem 0xfda00000-0xfda000ff 64bit] Oct 31 02:42:30.064668 kernel: pci 0000:00:02.0: PCI bridge to [bus 01-02] Oct 31 02:42:30.064862 kernel: pci 0000:00:02.0: bridge window [mem 0xfd800000-0xfdbfffff] Oct 31 02:42:30.065036 kernel: pci 0000:00:02.0: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Oct 31 02:42:30.066525 kernel: pci_bus 0000:02: extended config space not accessible Oct 31 02:42:30.066768 kernel: pci 0000:02:01.0: [8086:25ab] type 00 class 0x088000 Oct 31 02:42:30.066969 kernel: pci 0000:02:01.0: reg 0x10: [mem 0xfd800000-0xfd80000f] Oct 31 02:42:30.067178 kernel: pci 0000:01:00.0: PCI bridge to [bus 02] Oct 31 02:42:30.067372 kernel: pci 0000:01:00.0: bridge window [mem 0xfd800000-0xfd9fffff] Oct 31 02:42:30.067583 kernel: pci 0000:03:00.0: [1b36:000d] type 00 class 0x0c0330 Oct 31 02:42:30.067766 kernel: pci 0000:03:00.0: reg 0x10: [mem 0xfe800000-0xfe803fff 64bit] Oct 31 02:42:30.067943 kernel: pci 0000:00:02.1: PCI bridge to [bus 03] Oct 31 02:42:30.070196 kernel: pci 0000:00:02.1: bridge window [mem 0xfe800000-0xfe9fffff] Oct 31 02:42:30.070400 kernel: pci 0000:00:02.1: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Oct 31 02:42:30.070615 kernel: pci 0000:04:00.0: [1af4:1044] type 00 class 0x00ff00 Oct 31 02:42:30.070808 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xfca00000-0xfca03fff 64bit pref] Oct 31 02:42:30.070986 kernel: pci 0000:00:02.2: PCI bridge to [bus 04] Oct 31 02:42:30.071204 kernel: pci 0000:00:02.2: bridge window [mem 0xfe600000-0xfe7fffff] Oct 31 02:42:30.071394 kernel: pci 0000:00:02.2: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Oct 31 02:42:30.071568 kernel: pci 0000:00:02.3: PCI bridge to [bus 05] Oct 31 02:42:30.071738 kernel: pci 0000:00:02.3: bridge window [mem 0xfe400000-0xfe5fffff] Oct 31 02:42:30.071932 kernel: pci 0000:00:02.3: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Oct 31 02:42:30.074184 kernel: pci 0000:00:02.4: PCI bridge to [bus 06] Oct 31 02:42:30.074385 kernel: pci 0000:00:02.4: bridge window [mem 0xfe200000-0xfe3fffff] Oct 31 02:42:30.074563 kernel: pci 0000:00:02.4: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Oct 31 02:42:30.074739 kernel: pci 0000:00:02.5: PCI bridge to [bus 07] Oct 31 02:42:30.074912 kernel: pci 0000:00:02.5: bridge window [mem 0xfe000000-0xfe1fffff] Oct 31 02:42:30.075099 kernel: pci 0000:00:02.5: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Oct 31 02:42:30.075274 kernel: pci 0000:00:02.6: PCI bridge to [bus 08] Oct 31 02:42:30.075461 kernel: pci 0000:00:02.6: bridge window [mem 0xfde00000-0xfdffffff] Oct 31 02:42:30.075650 kernel: pci 0000:00:02.6: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Oct 31 02:42:30.075819 kernel: pci 0000:00:02.7: PCI bridge to [bus 09] Oct 31 02:42:30.076003 kernel: pci 0000:00:02.7: bridge window [mem 0xfdc00000-0xfddfffff] Oct 31 02:42:30.078215 kernel: pci 0000:00:02.7: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Oct 31 02:42:30.078240 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Oct 31 02:42:30.078255 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Oct 31 02:42:30.078269 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Oct 31 02:42:30.078282 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Oct 31 02:42:30.078308 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Oct 31 02:42:30.078330 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Oct 31 02:42:30.078343 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Oct 31 02:42:30.078356 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Oct 31 02:42:30.078369 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Oct 31 02:42:30.078383 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Oct 31 02:42:30.078396 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Oct 31 02:42:30.078409 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Oct 31 02:42:30.078422 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Oct 31 02:42:30.078435 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Oct 31 02:42:30.078453 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Oct 31 02:42:30.078466 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Oct 31 02:42:30.078480 kernel: iommu: Default domain type: Translated Oct 31 02:42:30.078493 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Oct 31 02:42:30.078506 kernel: PCI: Using ACPI for IRQ routing Oct 31 02:42:30.078519 kernel: PCI: pci_cache_line_size set to 64 bytes Oct 31 02:42:30.078532 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Oct 31 02:42:30.078545 kernel: e820: reserve RAM buffer [mem 0x7ffdc000-0x7fffffff] Oct 31 02:42:30.078718 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Oct 31 02:42:30.078905 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Oct 31 02:42:30.079103 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Oct 31 02:42:30.079123 kernel: vgaarb: loaded Oct 31 02:42:30.079146 kernel: clocksource: Switched to clocksource kvm-clock Oct 31 02:42:30.079159 kernel: VFS: Disk quotas dquot_6.6.0 Oct 31 02:42:30.079172 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 31 02:42:30.079185 kernel: pnp: PnP ACPI init Oct 31 02:42:30.079403 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Oct 31 02:42:30.079432 kernel: pnp: PnP ACPI: found 5 devices Oct 31 02:42:30.079446 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Oct 31 02:42:30.079459 kernel: NET: Registered PF_INET protocol family Oct 31 02:42:30.079472 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 31 02:42:30.079486 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Oct 31 02:42:30.079499 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 31 02:42:30.079512 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Oct 31 02:42:30.079525 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Oct 31 02:42:30.079543 kernel: TCP: Hash tables configured (established 16384 bind 16384) Oct 31 02:42:30.079557 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 31 02:42:30.079570 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 31 02:42:30.079583 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 31 02:42:30.079596 kernel: NET: Registered PF_XDP protocol family Oct 31 02:42:30.079763 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01-02] add_size 1000 Oct 31 02:42:30.079937 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Oct 31 02:42:30.082185 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Oct 31 02:42:30.082389 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Oct 31 02:42:30.082562 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Oct 31 02:42:30.082763 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Oct 31 02:42:30.082949 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Oct 31 02:42:30.083141 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Oct 31 02:42:30.083330 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Oct 31 02:42:30.083510 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Oct 31 02:42:30.083679 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Oct 31 02:42:30.083848 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Oct 31 02:42:30.085059 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Oct 31 02:42:30.085246 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Oct 31 02:42:30.085431 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Oct 31 02:42:30.085601 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Oct 31 02:42:30.085778 kernel: pci 0000:01:00.0: PCI bridge to [bus 02] Oct 31 02:42:30.085990 kernel: pci 0000:01:00.0: bridge window [mem 0xfd800000-0xfd9fffff] Oct 31 02:42:30.087207 kernel: pci 0000:00:02.0: PCI bridge to [bus 01-02] Oct 31 02:42:30.087396 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Oct 31 02:42:30.087569 kernel: pci 0000:00:02.0: bridge window [mem 0xfd800000-0xfdbfffff] Oct 31 02:42:30.087740 kernel: pci 0000:00:02.0: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Oct 31 02:42:30.087922 kernel: pci 0000:00:02.1: PCI bridge to [bus 03] Oct 31 02:42:30.089143 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Oct 31 02:42:30.089344 kernel: pci 0000:00:02.1: bridge window [mem 0xfe800000-0xfe9fffff] Oct 31 02:42:30.089516 kernel: pci 0000:00:02.1: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Oct 31 02:42:30.089695 kernel: pci 0000:00:02.2: PCI bridge to [bus 04] Oct 31 02:42:30.089874 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Oct 31 02:42:30.092076 kernel: pci 0000:00:02.2: bridge window [mem 0xfe600000-0xfe7fffff] Oct 31 02:42:30.092302 kernel: pci 0000:00:02.2: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Oct 31 02:42:30.092476 kernel: pci 0000:00:02.3: PCI bridge to [bus 05] Oct 31 02:42:30.092654 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Oct 31 02:42:30.092826 kernel: pci 0000:00:02.3: bridge window [mem 0xfe400000-0xfe5fffff] Oct 31 02:42:30.093007 kernel: pci 0000:00:02.3: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Oct 31 02:42:30.093239 kernel: pci 0000:00:02.4: PCI bridge to [bus 06] Oct 31 02:42:30.093456 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Oct 31 02:42:30.093628 kernel: pci 0000:00:02.4: bridge window [mem 0xfe200000-0xfe3fffff] Oct 31 02:42:30.093798 kernel: pci 0000:00:02.4: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Oct 31 02:42:30.093968 kernel: pci 0000:00:02.5: PCI bridge to [bus 07] Oct 31 02:42:30.094163 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Oct 31 02:42:30.094358 kernel: pci 0000:00:02.5: bridge window [mem 0xfe000000-0xfe1fffff] Oct 31 02:42:30.094528 kernel: pci 0000:00:02.5: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Oct 31 02:42:30.094704 kernel: pci 0000:00:02.6: PCI bridge to [bus 08] Oct 31 02:42:30.094875 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Oct 31 02:42:30.097079 kernel: pci 0000:00:02.6: bridge window [mem 0xfde00000-0xfdffffff] Oct 31 02:42:30.097301 kernel: pci 0000:00:02.6: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Oct 31 02:42:30.097479 kernel: pci 0000:00:02.7: PCI bridge to [bus 09] Oct 31 02:42:30.097651 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Oct 31 02:42:30.097826 kernel: pci 0000:00:02.7: bridge window [mem 0xfdc00000-0xfddfffff] Oct 31 02:42:30.097998 kernel: pci 0000:00:02.7: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Oct 31 02:42:30.098176 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Oct 31 02:42:30.098378 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Oct 31 02:42:30.098538 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Oct 31 02:42:30.098700 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Oct 31 02:42:30.098866 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Oct 31 02:42:30.101040 kernel: pci_bus 0000:00: resource 9 [mem 0x20c0000000-0x28bfffffff window] Oct 31 02:42:30.101379 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Oct 31 02:42:30.101552 kernel: pci_bus 0000:01: resource 1 [mem 0xfd800000-0xfdbfffff] Oct 31 02:42:30.101719 kernel: pci_bus 0000:01: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Oct 31 02:42:30.101898 kernel: pci_bus 0000:02: resource 1 [mem 0xfd800000-0xfd9fffff] Oct 31 02:42:30.104116 kernel: pci_bus 0000:03: resource 0 [io 0x2000-0x2fff] Oct 31 02:42:30.104310 kernel: pci_bus 0000:03: resource 1 [mem 0xfe800000-0xfe9fffff] Oct 31 02:42:30.104478 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Oct 31 02:42:30.104675 kernel: pci_bus 0000:04: resource 0 [io 0x3000-0x3fff] Oct 31 02:42:30.104841 kernel: pci_bus 0000:04: resource 1 [mem 0xfe600000-0xfe7fffff] Oct 31 02:42:30.105006 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Oct 31 02:42:30.105270 kernel: pci_bus 0000:05: resource 0 [io 0x4000-0x4fff] Oct 31 02:42:30.105456 kernel: pci_bus 0000:05: resource 1 [mem 0xfe400000-0xfe5fffff] Oct 31 02:42:30.105617 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Oct 31 02:42:30.105819 kernel: pci_bus 0000:06: resource 0 [io 0x5000-0x5fff] Oct 31 02:42:30.105981 kernel: pci_bus 0000:06: resource 1 [mem 0xfe200000-0xfe3fffff] Oct 31 02:42:30.106174 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Oct 31 02:42:30.106384 kernel: pci_bus 0000:07: resource 0 [io 0x6000-0x6fff] Oct 31 02:42:30.106546 kernel: pci_bus 0000:07: resource 1 [mem 0xfe000000-0xfe1fffff] Oct 31 02:42:30.106736 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Oct 31 02:42:30.106916 kernel: pci_bus 0000:08: resource 0 [io 0x7000-0x7fff] Oct 31 02:42:30.107097 kernel: pci_bus 0000:08: resource 1 [mem 0xfde00000-0xfdffffff] Oct 31 02:42:30.107271 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Oct 31 02:42:30.107480 kernel: pci_bus 0000:09: resource 0 [io 0x8000-0x8fff] Oct 31 02:42:30.107668 kernel: pci_bus 0000:09: resource 1 [mem 0xfdc00000-0xfddfffff] Oct 31 02:42:30.107830 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Oct 31 02:42:30.107859 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Oct 31 02:42:30.107873 kernel: PCI: CLS 0 bytes, default 64 Oct 31 02:42:30.107887 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Oct 31 02:42:30.107901 kernel: software IO TLB: mapped [mem 0x0000000079800000-0x000000007d800000] (64MB) Oct 31 02:42:30.107915 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Oct 31 02:42:30.107929 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x240957bf147, max_idle_ns: 440795216753 ns Oct 31 02:42:30.107943 kernel: Initialise system trusted keyrings Oct 31 02:42:30.107957 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Oct 31 02:42:30.107970 kernel: Key type asymmetric registered Oct 31 02:42:30.107989 kernel: Asymmetric key parser 'x509' registered Oct 31 02:42:30.108003 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Oct 31 02:42:30.108017 kernel: io scheduler mq-deadline registered Oct 31 02:42:30.108031 kernel: io scheduler kyber registered Oct 31 02:42:30.108072 kernel: io scheduler bfq registered Oct 31 02:42:30.108249 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Oct 31 02:42:30.108439 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Oct 31 02:42:30.108623 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.108815 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Oct 31 02:42:30.108988 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Oct 31 02:42:30.109214 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.109400 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Oct 31 02:42:30.109573 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Oct 31 02:42:30.109743 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.109926 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Oct 31 02:42:30.110130 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Oct 31 02:42:30.110316 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.110501 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Oct 31 02:42:30.110685 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Oct 31 02:42:30.110866 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.111082 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Oct 31 02:42:30.111258 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Oct 31 02:42:30.111441 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.111624 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Oct 31 02:42:30.111802 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Oct 31 02:42:30.111972 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.112203 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Oct 31 02:42:30.112389 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Oct 31 02:42:30.112572 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Oct 31 02:42:30.112593 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Oct 31 02:42:30.112607 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Oct 31 02:42:30.112621 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Oct 31 02:42:30.112642 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 31 02:42:30.112660 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Oct 31 02:42:30.112674 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Oct 31 02:42:30.112687 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Oct 31 02:42:30.112700 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Oct 31 02:42:30.112879 kernel: rtc_cmos 00:03: RTC can wake from S4 Oct 31 02:42:30.112914 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Oct 31 02:42:30.113088 kernel: rtc_cmos 00:03: registered as rtc0 Oct 31 02:42:30.113250 kernel: rtc_cmos 00:03: setting system clock to 2025-10-31T02:42:29 UTC (1761878549) Oct 31 02:42:30.113434 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Oct 31 02:42:30.113455 kernel: intel_pstate: CPU model not supported Oct 31 02:42:30.113469 kernel: NET: Registered PF_INET6 protocol family Oct 31 02:42:30.113483 kernel: Segment Routing with IPv6 Oct 31 02:42:30.113497 kernel: In-situ OAM (IOAM) with IPv6 Oct 31 02:42:30.113510 kernel: NET: Registered PF_PACKET protocol family Oct 31 02:42:30.113524 kernel: Key type dns_resolver registered Oct 31 02:42:30.113537 kernel: IPI shorthand broadcast: enabled Oct 31 02:42:30.113551 kernel: sched_clock: Marking stable (1144013067, 229267139)->(1616318518, -243038312) Oct 31 02:42:30.113572 kernel: registered taskstats version 1 Oct 31 02:42:30.113586 kernel: Loading compiled-in X.509 certificates Oct 31 02:42:30.113600 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.113-flatcar: 3640cadef2ce00a652278ae302be325ebb54a228' Oct 31 02:42:30.113614 kernel: Key type .fscrypt registered Oct 31 02:42:30.113627 kernel: Key type fscrypt-provisioning registered Oct 31 02:42:30.113641 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 31 02:42:30.113655 kernel: ima: Allocated hash algorithm: sha1 Oct 31 02:42:30.113668 kernel: ima: No architecture policies found Oct 31 02:42:30.113682 kernel: clk: Disabling unused clocks Oct 31 02:42:30.113701 kernel: Freeing unused kernel image (initmem) memory: 42880K Oct 31 02:42:30.113715 kernel: Write protecting the kernel read-only data: 36864k Oct 31 02:42:30.113728 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Oct 31 02:42:30.113742 kernel: Run /init as init process Oct 31 02:42:30.113756 kernel: with arguments: Oct 31 02:42:30.113770 kernel: /init Oct 31 02:42:30.113783 kernel: with environment: Oct 31 02:42:30.113797 kernel: HOME=/ Oct 31 02:42:30.113810 kernel: TERM=linux Oct 31 02:42:30.113832 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Oct 31 02:42:30.113849 systemd[1]: Detected virtualization kvm. Oct 31 02:42:30.113864 systemd[1]: Detected architecture x86-64. Oct 31 02:42:30.113878 systemd[1]: Running in initrd. Oct 31 02:42:30.113892 systemd[1]: No hostname configured, using default hostname. Oct 31 02:42:30.113906 systemd[1]: Hostname set to . Oct 31 02:42:30.113920 systemd[1]: Initializing machine ID from VM UUID. Oct 31 02:42:30.113941 systemd[1]: Queued start job for default target initrd.target. Oct 31 02:42:30.113955 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 31 02:42:30.113970 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 31 02:42:30.113985 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 31 02:42:30.113999 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 31 02:42:30.114014 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 31 02:42:30.114058 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 31 02:42:30.114082 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Oct 31 02:42:30.114098 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Oct 31 02:42:30.114112 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 31 02:42:30.114127 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 31 02:42:30.114141 systemd[1]: Reached target paths.target - Path Units. Oct 31 02:42:30.114156 systemd[1]: Reached target slices.target - Slice Units. Oct 31 02:42:30.114170 systemd[1]: Reached target swap.target - Swaps. Oct 31 02:42:30.114185 systemd[1]: Reached target timers.target - Timer Units. Oct 31 02:42:30.114205 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 31 02:42:30.114224 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 31 02:42:30.114239 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 31 02:42:30.114254 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Oct 31 02:42:30.114268 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 31 02:42:30.114283 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 31 02:42:30.114317 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 31 02:42:30.114333 systemd[1]: Reached target sockets.target - Socket Units. Oct 31 02:42:30.114347 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 31 02:42:30.114369 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 31 02:42:30.114383 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 31 02:42:30.114398 systemd[1]: Starting systemd-fsck-usr.service... Oct 31 02:42:30.114412 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 31 02:42:30.114427 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 31 02:42:30.114441 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 31 02:42:30.114497 systemd-journald[202]: Collecting audit messages is disabled. Oct 31 02:42:30.114535 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 31 02:42:30.114550 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 31 02:42:30.114565 systemd[1]: Finished systemd-fsck-usr.service. Oct 31 02:42:30.114586 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 31 02:42:30.114601 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 31 02:42:30.114615 kernel: Bridge firewalling registered Oct 31 02:42:30.114630 systemd-journald[202]: Journal started Oct 31 02:42:30.114673 systemd-journald[202]: Runtime Journal (/run/log/journal/78198dc696834c4fa3e789d144cf0e99) is 4.7M, max 38.0M, 33.2M free. Oct 31 02:42:30.053365 systemd-modules-load[203]: Inserted module 'overlay' Oct 31 02:42:30.103696 systemd-modules-load[203]: Inserted module 'br_netfilter' Oct 31 02:42:30.171040 systemd[1]: Started systemd-journald.service - Journal Service. Oct 31 02:42:30.171466 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 31 02:42:30.172505 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 31 02:42:30.180227 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 31 02:42:30.184258 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 31 02:42:30.198887 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 31 02:42:30.200689 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 31 02:42:30.210270 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 31 02:42:30.219992 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 31 02:42:30.223798 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 31 02:42:30.226216 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 31 02:42:30.228790 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 31 02:42:30.239236 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 31 02:42:30.242706 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 31 02:42:30.258012 dracut-cmdline[232]: dracut-dracut-053 Oct 31 02:42:30.264301 dracut-cmdline[232]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=950876ad7bc3e9634b7585a81697da4ef03ac6558969e5c002165369dd7c7885 Oct 31 02:42:30.284455 systemd-resolved[234]: Positive Trust Anchors: Oct 31 02:42:30.284472 systemd-resolved[234]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 31 02:42:30.284518 systemd-resolved[234]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 31 02:42:30.294055 systemd-resolved[234]: Defaulting to hostname 'linux'. Oct 31 02:42:30.296204 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 31 02:42:30.298004 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 31 02:42:30.368297 kernel: SCSI subsystem initialized Oct 31 02:42:30.379085 kernel: Loading iSCSI transport class v2.0-870. Oct 31 02:42:30.393054 kernel: iscsi: registered transport (tcp) Oct 31 02:42:30.420742 kernel: iscsi: registered transport (qla4xxx) Oct 31 02:42:30.420800 kernel: QLogic iSCSI HBA Driver Oct 31 02:42:30.477125 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 31 02:42:30.483225 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 31 02:42:30.517565 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 31 02:42:30.517622 kernel: device-mapper: uevent: version 1.0.3 Oct 31 02:42:30.519962 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Oct 31 02:42:30.568141 kernel: raid6: sse2x4 gen() 12804 MB/s Oct 31 02:42:30.586111 kernel: raid6: sse2x2 gen() 8956 MB/s Oct 31 02:42:30.604923 kernel: raid6: sse2x1 gen() 9568 MB/s Oct 31 02:42:30.604978 kernel: raid6: using algorithm sse2x4 gen() 12804 MB/s Oct 31 02:42:30.626825 kernel: raid6: .... xor() 7520 MB/s, rmw enabled Oct 31 02:42:30.627199 kernel: raid6: using ssse3x2 recovery algorithm Oct 31 02:42:30.654059 kernel: xor: automatically using best checksumming function avx Oct 31 02:42:30.857084 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 31 02:42:30.873631 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 31 02:42:30.881322 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 31 02:42:30.913275 systemd-udevd[419]: Using default interface naming scheme 'v255'. Oct 31 02:42:30.921994 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 31 02:42:30.930600 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 31 02:42:30.953476 dracut-pre-trigger[424]: rd.md=0: removing MD RAID activation Oct 31 02:42:30.994549 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 31 02:42:31.002302 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 31 02:42:31.118096 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 31 02:42:31.130285 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 31 02:42:31.160491 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 31 02:42:31.162478 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 31 02:42:31.164479 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 31 02:42:31.166642 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 31 02:42:31.174225 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 31 02:42:31.209757 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 31 02:42:31.254047 kernel: virtio_blk virtio1: 2/0/0 default/read/poll queues Oct 31 02:42:31.278230 kernel: virtio_blk virtio1: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Oct 31 02:42:31.280049 kernel: cryptd: max_cpu_qlen set to 1000 Oct 31 02:42:31.301586 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 31 02:42:31.301659 kernel: GPT:17805311 != 125829119 Oct 31 02:42:31.301679 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 31 02:42:31.301696 kernel: GPT:17805311 != 125829119 Oct 31 02:42:31.301725 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 31 02:42:31.301744 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 31 02:42:31.304163 kernel: AVX version of gcm_enc/dec engaged. Oct 31 02:42:31.304200 kernel: AES CTR mode by8 optimization enabled Oct 31 02:42:31.319819 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 31 02:42:31.320977 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 31 02:42:31.322955 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 31 02:42:31.324876 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 31 02:42:31.325081 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 31 02:42:31.329966 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 31 02:42:31.344752 kernel: ACPI: bus type USB registered Oct 31 02:42:31.344805 kernel: usbcore: registered new interface driver usbfs Oct 31 02:42:31.346275 kernel: usbcore: registered new interface driver hub Oct 31 02:42:31.346357 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 31 02:42:31.350060 kernel: usbcore: registered new device driver usb Oct 31 02:42:31.362055 kernel: libata version 3.00 loaded. Oct 31 02:42:31.408780 kernel: ahci 0000:00:1f.2: version 3.0 Oct 31 02:42:31.413036 kernel: BTRFS: device fsid 1021cdf2-f4a0-46ed-8fe0-b31d3115a6e0 devid 1 transid 37 /dev/vda3 scanned by (udev-worker) (479) Oct 31 02:42:31.416042 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (469) Oct 31 02:42:31.430051 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Oct 31 02:42:31.431043 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Oct 31 02:42:31.431332 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Oct 31 02:42:31.434935 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Oct 31 02:42:31.472607 kernel: scsi host0: ahci Oct 31 02:42:31.472885 kernel: scsi host1: ahci Oct 31 02:42:31.473144 kernel: scsi host2: ahci Oct 31 02:42:31.473391 kernel: scsi host3: ahci Oct 31 02:42:31.473597 kernel: scsi host4: ahci Oct 31 02:42:31.473816 kernel: scsi host5: ahci Oct 31 02:42:31.471929 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 31 02:42:31.511944 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b100 irq 38 Oct 31 02:42:31.511979 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b180 irq 38 Oct 31 02:42:31.512000 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b200 irq 38 Oct 31 02:42:31.512039 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b280 irq 38 Oct 31 02:42:31.512060 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b300 irq 38 Oct 31 02:42:31.512079 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b380 irq 38 Oct 31 02:42:31.512097 kernel: xhci_hcd 0000:03:00.0: xHCI Host Controller Oct 31 02:42:31.512372 kernel: xhci_hcd 0000:03:00.0: new USB bus registered, assigned bus number 1 Oct 31 02:42:31.512622 kernel: xhci_hcd 0000:03:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Oct 31 02:42:31.512823 kernel: xhci_hcd 0000:03:00.0: xHCI Host Controller Oct 31 02:42:31.513039 kernel: xhci_hcd 0000:03:00.0: new USB bus registered, assigned bus number 2 Oct 31 02:42:31.514849 kernel: xhci_hcd 0000:03:00.0: Host supports USB 3.0 SuperSpeed Oct 31 02:42:31.515126 kernel: hub 1-0:1.0: USB hub found Oct 31 02:42:31.515389 kernel: hub 1-0:1.0: 4 ports detected Oct 31 02:42:31.515615 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Oct 31 02:42:31.515837 kernel: hub 2-0:1.0: USB hub found Oct 31 02:42:31.516068 kernel: hub 2-0:1.0: 4 ports detected Oct 31 02:42:31.510948 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Oct 31 02:42:31.528149 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Oct 31 02:42:31.530056 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Oct 31 02:42:31.538664 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 31 02:42:31.553511 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 31 02:42:31.557231 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 31 02:42:31.561849 disk-uuid[561]: Primary Header is updated. Oct 31 02:42:31.561849 disk-uuid[561]: Secondary Entries is updated. Oct 31 02:42:31.561849 disk-uuid[561]: Secondary Header is updated. Oct 31 02:42:31.570070 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 31 02:42:31.575163 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 31 02:42:31.600530 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 31 02:42:31.754080 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Oct 31 02:42:31.784097 kernel: ata2: SATA link down (SStatus 0 SControl 300) Oct 31 02:42:31.792099 kernel: ata3: SATA link down (SStatus 0 SControl 300) Oct 31 02:42:31.800102 kernel: ata6: SATA link down (SStatus 0 SControl 300) Oct 31 02:42:31.800141 kernel: ata1: SATA link down (SStatus 0 SControl 300) Oct 31 02:42:31.802059 kernel: ata4: SATA link down (SStatus 0 SControl 300) Oct 31 02:42:31.802102 kernel: ata5: SATA link down (SStatus 0 SControl 300) Oct 31 02:42:31.900080 kernel: hid: raw HID events driver (C) Jiri Kosina Oct 31 02:42:31.906660 kernel: usbcore: registered new interface driver usbhid Oct 31 02:42:31.906710 kernel: usbhid: USB HID core driver Oct 31 02:42:31.916159 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:03:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input2 Oct 31 02:42:31.916208 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:03:00.0-1/input0 Oct 31 02:42:32.583070 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 31 02:42:32.585393 disk-uuid[562]: The operation has completed successfully. Oct 31 02:42:32.635878 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 31 02:42:32.636094 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 31 02:42:32.660268 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Oct 31 02:42:32.666737 sh[583]: Success Oct 31 02:42:32.684223 kernel: device-mapper: verity: sha256 using implementation "sha256-avx" Oct 31 02:42:32.743690 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Oct 31 02:42:32.753397 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Oct 31 02:42:32.755339 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Oct 31 02:42:32.787519 kernel: BTRFS info (device dm-0): first mount of filesystem 1021cdf2-f4a0-46ed-8fe0-b31d3115a6e0 Oct 31 02:42:32.787588 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Oct 31 02:42:32.787608 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Oct 31 02:42:32.790535 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 31 02:42:32.792263 kernel: BTRFS info (device dm-0): using free space tree Oct 31 02:42:32.802581 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Oct 31 02:42:32.804793 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 31 02:42:32.811251 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 31 02:42:32.814459 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 31 02:42:32.828166 kernel: BTRFS info (device vda6): first mount of filesystem 1a1fe00d-a5e5-45c6-a30a-fcc91f19f9ea Oct 31 02:42:32.828227 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 31 02:42:32.828249 kernel: BTRFS info (device vda6): using free space tree Oct 31 02:42:32.836188 kernel: BTRFS info (device vda6): auto enabling async discard Oct 31 02:42:32.853915 systemd[1]: mnt-oem.mount: Deactivated successfully. Oct 31 02:42:32.854913 kernel: BTRFS info (device vda6): last unmount of filesystem 1a1fe00d-a5e5-45c6-a30a-fcc91f19f9ea Oct 31 02:42:32.866412 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 31 02:42:32.873271 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 31 02:42:32.967642 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 31 02:42:32.977288 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 31 02:42:33.020632 systemd-networkd[766]: lo: Link UP Oct 31 02:42:33.021353 systemd-networkd[766]: lo: Gained carrier Oct 31 02:42:33.025690 systemd-networkd[766]: Enumeration completed Oct 31 02:42:33.026385 systemd-networkd[766]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 31 02:42:33.026391 systemd-networkd[766]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 31 02:42:33.027223 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 31 02:42:33.030494 systemd-networkd[766]: eth0: Link UP Oct 31 02:42:33.030501 systemd-networkd[766]: eth0: Gained carrier Oct 31 02:42:33.035056 ignition[671]: Ignition 2.19.0 Oct 31 02:42:33.030513 systemd-networkd[766]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 31 02:42:33.035075 ignition[671]: Stage: fetch-offline Oct 31 02:42:33.031173 systemd[1]: Reached target network.target - Network. Oct 31 02:42:33.035159 ignition[671]: no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:33.037693 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 31 02:42:33.035184 ignition[671]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:33.035390 ignition[671]: parsed url from cmdline: "" Oct 31 02:42:33.035397 ignition[671]: no config URL provided Oct 31 02:42:33.035407 ignition[671]: reading system config file "/usr/lib/ignition/user.ign" Oct 31 02:42:33.035423 ignition[671]: no config at "/usr/lib/ignition/user.ign" Oct 31 02:42:33.035433 ignition[671]: failed to fetch config: resource requires networking Oct 31 02:42:33.035716 ignition[671]: Ignition finished successfully Oct 31 02:42:33.048203 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Oct 31 02:42:33.063239 systemd-networkd[766]: eth0: DHCPv4 address 10.230.25.230/30, gateway 10.230.25.229 acquired from 10.230.25.229 Oct 31 02:42:33.066288 ignition[773]: Ignition 2.19.0 Oct 31 02:42:33.066308 ignition[773]: Stage: fetch Oct 31 02:42:33.066605 ignition[773]: no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:33.066639 ignition[773]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:33.066772 ignition[773]: parsed url from cmdline: "" Oct 31 02:42:33.066780 ignition[773]: no config URL provided Oct 31 02:42:33.066789 ignition[773]: reading system config file "/usr/lib/ignition/user.ign" Oct 31 02:42:33.066805 ignition[773]: no config at "/usr/lib/ignition/user.ign" Oct 31 02:42:33.067057 ignition[773]: GET http://169.254.169.254/openstack/latest/user_data: attempt #1 Oct 31 02:42:33.067413 ignition[773]: config drive ("/dev/disk/by-label/config-2") not found. Waiting... Oct 31 02:42:33.067429 ignition[773]: GET error: Get "http://169.254.169.254/openstack/latest/user_data": dial tcp 169.254.169.254:80: connect: network is unreachable Oct 31 02:42:33.067459 ignition[773]: config drive ("/dev/disk/by-label/CONFIG-2") not found. Waiting... Oct 31 02:42:33.267626 ignition[773]: GET http://169.254.169.254/openstack/latest/user_data: attempt #2 Oct 31 02:42:33.285993 ignition[773]: GET result: OK Oct 31 02:42:33.287009 ignition[773]: parsing config with SHA512: 6b6266735a9296eebec2b41622d3906274d2274472bef70f67c1bb26482ff8877bf5a0d0b7dcd4a4170776663de317b509c2d8bf8209488baf79a30e48618b6f Oct 31 02:42:33.293434 unknown[773]: fetched base config from "system" Oct 31 02:42:33.294099 unknown[773]: fetched base config from "system" Oct 31 02:42:33.294111 unknown[773]: fetched user config from "openstack" Oct 31 02:42:33.294801 ignition[773]: fetch: fetch complete Oct 31 02:42:33.294809 ignition[773]: fetch: fetch passed Oct 31 02:42:33.294879 ignition[773]: Ignition finished successfully Oct 31 02:42:33.297470 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Oct 31 02:42:33.309334 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 31 02:42:33.339913 ignition[780]: Ignition 2.19.0 Oct 31 02:42:33.339931 ignition[780]: Stage: kargs Oct 31 02:42:33.341288 ignition[780]: no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:33.341309 ignition[780]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:33.345524 ignition[780]: kargs: kargs passed Oct 31 02:42:33.346308 ignition[780]: Ignition finished successfully Oct 31 02:42:33.348278 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 31 02:42:33.355357 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 31 02:42:33.378276 ignition[786]: Ignition 2.19.0 Oct 31 02:42:33.378296 ignition[786]: Stage: disks Oct 31 02:42:33.378542 ignition[786]: no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:33.378563 ignition[786]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:33.383835 ignition[786]: disks: disks passed Oct 31 02:42:33.383921 ignition[786]: Ignition finished successfully Oct 31 02:42:33.385124 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 31 02:42:33.386745 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 31 02:42:33.388350 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 31 02:42:33.389935 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 31 02:42:33.391651 systemd[1]: Reached target sysinit.target - System Initialization. Oct 31 02:42:33.393173 systemd[1]: Reached target basic.target - Basic System. Oct 31 02:42:33.399313 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 31 02:42:33.421520 systemd-fsck[794]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Oct 31 02:42:33.424903 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 31 02:42:33.435312 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 31 02:42:33.561048 kernel: EXT4-fs (vda9): mounted filesystem 044ea9d4-3e15-48f6-be3f-240ec74f6b62 r/w with ordered data mode. Quota mode: none. Oct 31 02:42:33.561740 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 31 02:42:33.563298 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 31 02:42:33.571184 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 31 02:42:33.575152 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 31 02:42:33.576413 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Oct 31 02:42:33.580319 systemd[1]: Starting flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent... Oct 31 02:42:33.581882 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 31 02:42:33.581937 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 31 02:42:33.589790 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 31 02:42:33.594236 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 31 02:42:33.602081 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (802) Oct 31 02:42:33.602112 kernel: BTRFS info (device vda6): first mount of filesystem 1a1fe00d-a5e5-45c6-a30a-fcc91f19f9ea Oct 31 02:42:33.602139 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 31 02:42:33.602169 kernel: BTRFS info (device vda6): using free space tree Oct 31 02:42:33.621044 kernel: BTRFS info (device vda6): auto enabling async discard Oct 31 02:42:33.623790 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 31 02:42:33.705143 initrd-setup-root[831]: cut: /sysroot/etc/passwd: No such file or directory Oct 31 02:42:33.717425 initrd-setup-root[838]: cut: /sysroot/etc/group: No such file or directory Oct 31 02:42:33.722194 initrd-setup-root[845]: cut: /sysroot/etc/shadow: No such file or directory Oct 31 02:42:33.729388 initrd-setup-root[852]: cut: /sysroot/etc/gshadow: No such file or directory Oct 31 02:42:33.839145 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 31 02:42:33.845244 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 31 02:42:33.849236 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 31 02:42:33.860246 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 31 02:42:33.863467 kernel: BTRFS info (device vda6): last unmount of filesystem 1a1fe00d-a5e5-45c6-a30a-fcc91f19f9ea Oct 31 02:42:33.897182 ignition[919]: INFO : Ignition 2.19.0 Oct 31 02:42:33.897182 ignition[919]: INFO : Stage: mount Oct 31 02:42:33.896336 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 31 02:42:33.900606 ignition[919]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:33.900606 ignition[919]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:33.900606 ignition[919]: INFO : mount: mount passed Oct 31 02:42:33.900606 ignition[919]: INFO : Ignition finished successfully Oct 31 02:42:33.900106 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 31 02:42:34.429477 systemd-networkd[766]: eth0: Gained IPv6LL Oct 31 02:42:36.989986 systemd-networkd[766]: eth0: Ignoring DHCPv6 address 2a02:1348:179:8679:24:19ff:fee6:19e6/128 (valid for 59min 59s, preferred for 59min 59s) which conflicts with 2a02:1348:179:8679:24:19ff:fee6:19e6/64 assigned by NDisc. Oct 31 02:42:36.990005 systemd-networkd[766]: eth0: Hint: use IPv6Token= setting to change the address generated by NDisc or set UseAutonomousPrefix=no. Oct 31 02:42:40.791560 coreos-metadata[804]: Oct 31 02:42:40.791 WARN failed to locate config-drive, using the metadata service API instead Oct 31 02:42:40.815365 coreos-metadata[804]: Oct 31 02:42:40.815 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Oct 31 02:42:40.828362 coreos-metadata[804]: Oct 31 02:42:40.828 INFO Fetch successful Oct 31 02:42:40.829273 coreos-metadata[804]: Oct 31 02:42:40.828 INFO wrote hostname srv-4lokn.gb1.brightbox.com to /sysroot/etc/hostname Oct 31 02:42:40.831782 systemd[1]: flatcar-openstack-hostname.service: Deactivated successfully. Oct 31 02:42:40.832037 systemd[1]: Finished flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent. Oct 31 02:42:40.840155 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 31 02:42:40.862282 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 31 02:42:40.874139 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (936) Oct 31 02:42:40.880062 kernel: BTRFS info (device vda6): first mount of filesystem 1a1fe00d-a5e5-45c6-a30a-fcc91f19f9ea Oct 31 02:42:40.880106 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 31 02:42:40.880148 kernel: BTRFS info (device vda6): using free space tree Oct 31 02:42:40.885051 kernel: BTRFS info (device vda6): auto enabling async discard Oct 31 02:42:40.888258 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 31 02:42:40.919314 ignition[954]: INFO : Ignition 2.19.0 Oct 31 02:42:40.919314 ignition[954]: INFO : Stage: files Oct 31 02:42:40.921217 ignition[954]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:40.921217 ignition[954]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:40.921217 ignition[954]: DEBUG : files: compiled without relabeling support, skipping Oct 31 02:42:40.924044 ignition[954]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 31 02:42:40.924044 ignition[954]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 31 02:42:40.926160 ignition[954]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 31 02:42:40.926160 ignition[954]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 31 02:42:40.928500 ignition[954]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 31 02:42:40.928209 unknown[954]: wrote ssh authorized keys file for user: core Oct 31 02:42:40.930508 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Oct 31 02:42:40.930508 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Oct 31 02:42:40.930508 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 31 02:42:40.930508 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Oct 31 02:42:41.107475 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Oct 31 02:42:41.335066 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 31 02:42:41.335066 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 31 02:42:41.343331 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Oct 31 02:42:41.711307 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Oct 31 02:42:44.218518 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 31 02:42:44.218518 ignition[954]: INFO : files: op(c): [started] processing unit "containerd.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(c): [finished] processing unit "containerd.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Oct 31 02:42:44.223389 ignition[954]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 31 02:42:44.223389 ignition[954]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 31 02:42:44.223389 ignition[954]: INFO : files: files passed Oct 31 02:42:44.223389 ignition[954]: INFO : Ignition finished successfully Oct 31 02:42:44.223847 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 31 02:42:44.235266 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 31 02:42:44.243299 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 31 02:42:44.252011 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 31 02:42:44.252268 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 31 02:42:44.262322 initrd-setup-root-after-ignition[982]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 31 02:42:44.262322 initrd-setup-root-after-ignition[982]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 31 02:42:44.265651 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 31 02:42:44.267811 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 31 02:42:44.269506 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 31 02:42:44.279259 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 31 02:42:44.312977 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 31 02:42:44.313174 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 31 02:42:44.314980 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 31 02:42:44.316390 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 31 02:42:44.318144 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 31 02:42:44.329226 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 31 02:42:44.346406 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 31 02:42:44.361488 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 31 02:42:44.374684 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 31 02:42:44.376700 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 31 02:42:44.378294 systemd[1]: Stopped target timers.target - Timer Units. Oct 31 02:42:44.379343 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 31 02:42:44.379590 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 31 02:42:44.381302 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 31 02:42:44.382278 systemd[1]: Stopped target basic.target - Basic System. Oct 31 02:42:44.383854 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 31 02:42:44.385376 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 31 02:42:44.386855 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 31 02:42:44.388543 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 31 02:42:44.390229 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 31 02:42:44.391845 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 31 02:42:44.393437 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 31 02:42:44.394981 systemd[1]: Stopped target swap.target - Swaps. Oct 31 02:42:44.396399 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 31 02:42:44.396614 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 31 02:42:44.398482 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 31 02:42:44.399436 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 31 02:42:44.400971 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 31 02:42:44.401175 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 31 02:42:44.402616 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 31 02:42:44.402817 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 31 02:42:44.405029 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 31 02:42:44.405238 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 31 02:42:44.406929 systemd[1]: ignition-files.service: Deactivated successfully. Oct 31 02:42:44.407204 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 31 02:42:44.428330 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 31 02:42:44.434280 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 31 02:42:44.441730 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 31 02:42:44.442602 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 31 02:42:44.444038 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 31 02:42:44.444217 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 31 02:42:44.452935 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 31 02:42:44.453874 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 31 02:42:44.467473 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 31 02:42:44.469975 ignition[1006]: INFO : Ignition 2.19.0 Oct 31 02:42:44.469975 ignition[1006]: INFO : Stage: umount Oct 31 02:42:44.469975 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 31 02:42:44.469975 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Oct 31 02:42:44.470179 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 31 02:42:44.476095 ignition[1006]: INFO : umount: umount passed Oct 31 02:42:44.476095 ignition[1006]: INFO : Ignition finished successfully Oct 31 02:42:44.470378 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 31 02:42:44.473062 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 31 02:42:44.473192 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 31 02:42:44.475135 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 31 02:42:44.475271 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 31 02:42:44.476844 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 31 02:42:44.476923 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 31 02:42:44.478247 systemd[1]: ignition-fetch.service: Deactivated successfully. Oct 31 02:42:44.478313 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Oct 31 02:42:44.479613 systemd[1]: Stopped target network.target - Network. Oct 31 02:42:44.480978 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 31 02:42:44.481066 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 31 02:42:44.482485 systemd[1]: Stopped target paths.target - Path Units. Oct 31 02:42:44.483770 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 31 02:42:44.487109 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 31 02:42:44.488555 systemd[1]: Stopped target slices.target - Slice Units. Oct 31 02:42:44.489939 systemd[1]: Stopped target sockets.target - Socket Units. Oct 31 02:42:44.491336 systemd[1]: iscsid.socket: Deactivated successfully. Oct 31 02:42:44.491401 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 31 02:42:44.492864 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 31 02:42:44.492928 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 31 02:42:44.494454 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 31 02:42:44.494535 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 31 02:42:44.495942 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 31 02:42:44.496013 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 31 02:42:44.497605 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 31 02:42:44.497682 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 31 02:42:44.499676 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 31 02:42:44.501939 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 31 02:42:44.505131 systemd-networkd[766]: eth0: DHCPv6 lease lost Oct 31 02:42:44.508484 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 31 02:42:44.508687 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 31 02:42:44.510531 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 31 02:42:44.510645 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 31 02:42:44.521159 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 31 02:42:44.521975 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 31 02:42:44.522067 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 31 02:42:44.525653 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 31 02:42:44.527879 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 31 02:42:44.528092 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 31 02:42:44.536476 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 31 02:42:44.536717 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 31 02:42:44.544100 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 31 02:42:44.545254 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 31 02:42:44.547248 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 31 02:42:44.547325 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 31 02:42:44.548994 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 31 02:42:44.549105 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 31 02:42:44.551284 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 31 02:42:44.551363 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 31 02:42:44.552814 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 31 02:42:44.552923 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 31 02:42:44.566276 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 31 02:42:44.569413 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 31 02:42:44.569488 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 31 02:42:44.571132 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 31 02:42:44.571212 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 31 02:42:44.572602 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 31 02:42:44.572671 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 31 02:42:44.577090 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Oct 31 02:42:44.577173 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 31 02:42:44.578956 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 31 02:42:44.579048 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 31 02:42:44.580709 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 31 02:42:44.580800 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 31 02:42:44.582364 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 31 02:42:44.582437 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 31 02:42:44.585606 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 31 02:42:44.585755 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 31 02:42:44.587007 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 31 02:42:44.587163 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 31 02:42:44.589724 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 31 02:42:44.597237 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 31 02:42:44.607295 systemd[1]: Switching root. Oct 31 02:42:44.644826 systemd-journald[202]: Journal stopped Oct 31 02:42:46.067142 systemd-journald[202]: Received SIGTERM from PID 1 (systemd). Oct 31 02:42:46.067255 kernel: SELinux: policy capability network_peer_controls=1 Oct 31 02:42:46.067282 kernel: SELinux: policy capability open_perms=1 Oct 31 02:42:46.067309 kernel: SELinux: policy capability extended_socket_class=1 Oct 31 02:42:46.067346 kernel: SELinux: policy capability always_check_network=0 Oct 31 02:42:46.067375 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 31 02:42:46.067396 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 31 02:42:46.067421 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 31 02:42:46.067441 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 31 02:42:46.067461 kernel: audit: type=1403 audit(1761878564.846:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 31 02:42:46.067488 systemd[1]: Successfully loaded SELinux policy in 49.551ms. Oct 31 02:42:46.067524 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 21.122ms. Oct 31 02:42:46.067548 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Oct 31 02:42:46.067584 systemd[1]: Detected virtualization kvm. Oct 31 02:42:46.067614 systemd[1]: Detected architecture x86-64. Oct 31 02:42:46.067636 systemd[1]: Detected first boot. Oct 31 02:42:46.067662 systemd[1]: Hostname set to . Oct 31 02:42:46.067683 systemd[1]: Initializing machine ID from VM UUID. Oct 31 02:42:46.067713 zram_generator::config[1066]: No configuration found. Oct 31 02:42:46.067741 systemd[1]: Populated /etc with preset unit settings. Oct 31 02:42:46.067792 systemd[1]: Queued start job for default target multi-user.target. Oct 31 02:42:46.067817 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Oct 31 02:42:46.067839 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 31 02:42:46.067860 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 31 02:42:46.067880 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 31 02:42:46.067901 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 31 02:42:46.067921 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 31 02:42:46.067943 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 31 02:42:46.067964 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 31 02:42:46.067999 systemd[1]: Created slice user.slice - User and Session Slice. Oct 31 02:42:46.068034 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 31 02:42:46.068058 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 31 02:42:46.068080 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 31 02:42:46.068100 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 31 02:42:46.068121 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 31 02:42:46.068143 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 31 02:42:46.068165 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Oct 31 02:42:46.068185 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 31 02:42:46.068221 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 31 02:42:46.068244 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 31 02:42:46.068265 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 31 02:42:46.068285 systemd[1]: Reached target slices.target - Slice Units. Oct 31 02:42:46.068305 systemd[1]: Reached target swap.target - Swaps. Oct 31 02:42:46.068326 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 31 02:42:46.068359 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 31 02:42:46.068400 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 31 02:42:46.068435 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Oct 31 02:42:46.068457 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 31 02:42:46.068477 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 31 02:42:46.068497 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 31 02:42:46.068518 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 31 02:42:46.068550 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 31 02:42:46.068573 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 31 02:42:46.068602 systemd[1]: Mounting media.mount - External Media Directory... Oct 31 02:42:46.068624 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:46.068645 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 31 02:42:46.068667 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 31 02:42:46.068688 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 31 02:42:46.068709 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 31 02:42:46.068742 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 31 02:42:46.068782 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 31 02:42:46.068806 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 31 02:42:46.068827 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 31 02:42:46.068848 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 31 02:42:46.068870 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 31 02:42:46.068891 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 31 02:42:46.068912 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 31 02:42:46.068933 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 31 02:42:46.068967 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Oct 31 02:42:46.068997 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Oct 31 02:42:46.069152 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 31 02:42:46.069181 kernel: loop: module loaded Oct 31 02:42:46.069201 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 31 02:42:46.069221 kernel: fuse: init (API version 7.39) Oct 31 02:42:46.069241 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 31 02:42:46.069261 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 31 02:42:46.069296 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 31 02:42:46.069319 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:46.069339 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 31 02:42:46.069359 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 31 02:42:46.069414 kernel: ACPI: bus type drm_connector registered Oct 31 02:42:46.069435 systemd[1]: Mounted media.mount - External Media Directory. Oct 31 02:42:46.069457 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 31 02:42:46.069478 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 31 02:42:46.069535 systemd-journald[1184]: Collecting audit messages is disabled. Oct 31 02:42:46.069588 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 31 02:42:46.069611 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 31 02:42:46.069632 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 31 02:42:46.069653 systemd-journald[1184]: Journal started Oct 31 02:42:46.069693 systemd-journald[1184]: Runtime Journal (/run/log/journal/78198dc696834c4fa3e789d144cf0e99) is 4.7M, max 38.0M, 33.2M free. Oct 31 02:42:46.075082 systemd[1]: Started systemd-journald.service - Journal Service. Oct 31 02:42:46.076702 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 31 02:42:46.077016 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 31 02:42:46.078345 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 31 02:42:46.078598 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 31 02:42:46.079796 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 31 02:42:46.080075 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 31 02:42:46.081368 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 31 02:42:46.081617 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 31 02:42:46.082868 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 31 02:42:46.083297 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 31 02:42:46.084457 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 31 02:42:46.086381 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 31 02:42:46.088599 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 31 02:42:46.091879 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 31 02:42:46.094715 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 31 02:42:46.109979 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 31 02:42:46.117134 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 31 02:42:46.127133 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 31 02:42:46.129132 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 31 02:42:46.138215 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 31 02:42:46.152183 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 31 02:42:46.154199 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 31 02:42:46.159635 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 31 02:42:46.162190 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 31 02:42:46.175894 systemd-journald[1184]: Time spent on flushing to /var/log/journal/78198dc696834c4fa3e789d144cf0e99 is 65.602ms for 1124 entries. Oct 31 02:42:46.175894 systemd-journald[1184]: System Journal (/var/log/journal/78198dc696834c4fa3e789d144cf0e99) is 8.0M, max 584.8M, 576.8M free. Oct 31 02:42:46.259996 systemd-journald[1184]: Received client request to flush runtime journal. Oct 31 02:42:46.179252 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 31 02:42:46.189224 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 31 02:42:46.196540 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 31 02:42:46.200457 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 31 02:42:46.209564 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 31 02:42:46.212775 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 31 02:42:46.232651 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 31 02:42:46.263629 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 31 02:42:46.284720 systemd-tmpfiles[1222]: ACLs are not supported, ignoring. Oct 31 02:42:46.284746 systemd-tmpfiles[1222]: ACLs are not supported, ignoring. Oct 31 02:42:46.303411 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 31 02:42:46.319339 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 31 02:42:46.344635 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 31 02:42:46.358712 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Oct 31 02:42:46.380476 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 31 02:42:46.388314 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 31 02:42:46.396091 udevadm[1242]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Oct 31 02:42:46.419744 systemd-tmpfiles[1245]: ACLs are not supported, ignoring. Oct 31 02:42:46.419796 systemd-tmpfiles[1245]: ACLs are not supported, ignoring. Oct 31 02:42:46.427256 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 31 02:42:46.927857 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 31 02:42:46.935298 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 31 02:42:46.982794 systemd-udevd[1251]: Using default interface naming scheme 'v255'. Oct 31 02:42:47.011567 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 31 02:42:47.024289 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 31 02:42:47.057237 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 31 02:42:47.081401 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Oct 31 02:42:47.168131 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 31 02:42:47.229065 kernel: mousedev: PS/2 mouse device common for all mice Oct 31 02:42:47.240048 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1265) Oct 31 02:42:47.279915 systemd-networkd[1257]: lo: Link UP Oct 31 02:42:47.279928 systemd-networkd[1257]: lo: Gained carrier Oct 31 02:42:47.282418 systemd-networkd[1257]: Enumeration completed Oct 31 02:42:47.282925 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 31 02:42:47.283869 systemd-networkd[1257]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 31 02:42:47.283882 systemd-networkd[1257]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 31 02:42:47.285229 systemd-networkd[1257]: eth0: Link UP Oct 31 02:42:47.285235 systemd-networkd[1257]: eth0: Gained carrier Oct 31 02:42:47.285253 systemd-networkd[1257]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 31 02:42:47.296224 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Oct 31 02:42:47.300216 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 31 02:42:47.305177 systemd-networkd[1257]: eth0: DHCPv4 address 10.230.25.230/30, gateway 10.230.25.229 acquired from 10.230.25.229 Oct 31 02:42:47.334071 kernel: ACPI: button: Power Button [PWRF] Oct 31 02:42:47.370058 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Oct 31 02:42:47.375692 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Oct 31 02:42:47.376064 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Oct 31 02:42:47.382385 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 31 02:42:47.417055 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Oct 31 02:42:47.454319 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 31 02:42:47.605317 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 31 02:42:47.651437 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Oct 31 02:42:47.660291 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Oct 31 02:42:47.678040 lvm[1291]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Oct 31 02:42:47.713575 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Oct 31 02:42:47.715342 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 31 02:42:47.723324 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Oct 31 02:42:47.731181 lvm[1294]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Oct 31 02:42:47.760673 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Oct 31 02:42:47.762423 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 31 02:42:47.763402 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 31 02:42:47.763548 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 31 02:42:47.764373 systemd[1]: Reached target machines.target - Containers. Oct 31 02:42:47.767052 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Oct 31 02:42:47.773307 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 31 02:42:47.781434 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 31 02:42:47.783072 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 31 02:42:47.786623 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 31 02:42:47.799117 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Oct 31 02:42:47.804370 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 31 02:42:47.823901 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 31 02:42:47.835370 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 31 02:42:47.854052 kernel: loop0: detected capacity change from 0 to 8 Oct 31 02:42:47.867304 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 31 02:42:47.867839 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 31 02:42:47.871532 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Oct 31 02:42:47.900797 kernel: loop1: detected capacity change from 0 to 140768 Oct 31 02:42:47.951142 kernel: loop2: detected capacity change from 0 to 142488 Oct 31 02:42:48.002234 kernel: loop3: detected capacity change from 0 to 224512 Oct 31 02:42:48.046052 kernel: loop4: detected capacity change from 0 to 8 Oct 31 02:42:48.051176 kernel: loop5: detected capacity change from 0 to 140768 Oct 31 02:42:48.078065 kernel: loop6: detected capacity change from 0 to 142488 Oct 31 02:42:48.096059 kernel: loop7: detected capacity change from 0 to 224512 Oct 31 02:42:48.109163 (sd-merge)[1316]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-openstack'. Oct 31 02:42:48.110566 (sd-merge)[1316]: Merged extensions into '/usr'. Oct 31 02:42:48.134857 systemd[1]: Reloading requested from client PID 1302 ('systemd-sysext') (unit systemd-sysext.service)... Oct 31 02:42:48.135377 systemd[1]: Reloading... Oct 31 02:42:48.209129 zram_generator::config[1341]: No configuration found. Oct 31 02:42:48.450704 ldconfig[1298]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 31 02:42:48.485555 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 31 02:42:48.577749 systemd[1]: Reloading finished in 441 ms. Oct 31 02:42:48.597077 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 31 02:42:48.598407 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 31 02:42:48.610268 systemd[1]: Starting ensure-sysext.service... Oct 31 02:42:48.614212 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 31 02:42:48.631124 systemd[1]: Reloading requested from client PID 1407 ('systemctl') (unit ensure-sysext.service)... Oct 31 02:42:48.631158 systemd[1]: Reloading... Oct 31 02:42:48.671871 systemd-tmpfiles[1408]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 31 02:42:48.672504 systemd-tmpfiles[1408]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 31 02:42:48.674714 systemd-tmpfiles[1408]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 31 02:42:48.675298 systemd-tmpfiles[1408]: ACLs are not supported, ignoring. Oct 31 02:42:48.675522 systemd-tmpfiles[1408]: ACLs are not supported, ignoring. Oct 31 02:42:48.680598 systemd-tmpfiles[1408]: Detected autofs mount point /boot during canonicalization of boot. Oct 31 02:42:48.681066 systemd-tmpfiles[1408]: Skipping /boot Oct 31 02:42:48.699561 systemd-tmpfiles[1408]: Detected autofs mount point /boot during canonicalization of boot. Oct 31 02:42:48.699754 systemd-tmpfiles[1408]: Skipping /boot Oct 31 02:42:48.742062 zram_generator::config[1439]: No configuration found. Oct 31 02:42:48.917818 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 31 02:42:49.009450 systemd[1]: Reloading finished in 377 ms. Oct 31 02:42:49.037749 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 31 02:42:49.045413 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Oct 31 02:42:49.057441 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 31 02:42:49.063231 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 31 02:42:49.071679 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 31 02:42:49.085285 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 31 02:42:49.086282 systemd-networkd[1257]: eth0: Gained IPv6LL Oct 31 02:42:49.100589 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 31 02:42:49.108999 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:49.110393 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 31 02:42:49.114660 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 31 02:42:49.126255 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 31 02:42:49.137860 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 31 02:42:49.140083 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 31 02:42:49.140274 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:49.150147 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:49.150911 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 31 02:42:49.151880 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 31 02:42:49.152112 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:49.158473 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:49.159288 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 31 02:42:49.166353 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 31 02:42:49.167647 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 31 02:42:49.167843 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 31 02:42:49.170627 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 31 02:42:49.179627 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 31 02:42:49.180032 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 31 02:42:49.182433 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 31 02:42:49.182678 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 31 02:42:49.189816 systemd[1]: Finished ensure-sysext.service. Oct 31 02:42:49.195417 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 31 02:42:49.201849 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Oct 31 02:42:49.204937 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 31 02:42:49.205328 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 31 02:42:49.208959 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 31 02:42:49.215548 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 31 02:42:49.217303 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 31 02:42:49.225744 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 31 02:42:49.235317 augenrules[1543]: No rules Oct 31 02:42:49.241411 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 31 02:42:49.248749 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Oct 31 02:42:49.258462 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 31 02:42:49.265658 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 31 02:42:49.275346 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 31 02:42:49.321568 systemd-resolved[1505]: Positive Trust Anchors: Oct 31 02:42:49.321601 systemd-resolved[1505]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 31 02:42:49.321648 systemd-resolved[1505]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 31 02:42:49.326204 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Oct 31 02:42:49.334122 systemd[1]: Reached target time-set.target - System Time Set. Oct 31 02:42:49.336406 systemd-resolved[1505]: Using system hostname 'srv-4lokn.gb1.brightbox.com'. Oct 31 02:42:49.340364 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 31 02:42:49.341458 systemd[1]: Reached target network.target - Network. Oct 31 02:42:49.342255 systemd[1]: Reached target network-online.target - Network is Online. Oct 31 02:42:49.343136 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 31 02:42:49.344015 systemd[1]: Reached target sysinit.target - System Initialization. Oct 31 02:42:49.344913 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 31 02:42:49.345831 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 31 02:42:49.346988 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 31 02:42:49.347871 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 31 02:42:49.348689 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 31 02:42:49.349642 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 31 02:42:49.349703 systemd[1]: Reached target paths.target - Path Units. Oct 31 02:42:49.350359 systemd[1]: Reached target timers.target - Timer Units. Oct 31 02:42:49.352630 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 31 02:42:49.356844 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 31 02:42:49.360806 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 31 02:42:49.367574 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 31 02:42:49.368378 systemd[1]: Reached target sockets.target - Socket Units. Oct 31 02:42:49.369062 systemd[1]: Reached target basic.target - Basic System. Oct 31 02:42:49.370004 systemd[1]: System is tainted: cgroupsv1 Oct 31 02:42:49.370117 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 31 02:42:49.370161 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 31 02:42:49.374212 systemd[1]: Starting containerd.service - containerd container runtime... Oct 31 02:42:49.379211 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Oct 31 02:42:49.389218 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 31 02:42:49.395133 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 31 02:42:49.402213 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 31 02:42:49.405111 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 31 02:42:49.413554 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:42:49.419679 jq[1563]: false Oct 31 02:42:49.427300 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 31 02:42:49.434225 dbus-daemon[1561]: [system] SELinux support is enabled Oct 31 02:42:49.439612 dbus-daemon[1561]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.2' (uid=244 pid=1257 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Oct 31 02:42:49.446611 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 31 02:42:49.453223 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 31 02:42:49.468269 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 31 02:42:49.472345 extend-filesystems[1565]: Found loop4 Oct 31 02:42:49.472345 extend-filesystems[1565]: Found loop5 Oct 31 02:42:49.472345 extend-filesystems[1565]: Found loop6 Oct 31 02:42:49.472345 extend-filesystems[1565]: Found loop7 Oct 31 02:42:49.472345 extend-filesystems[1565]: Found vda Oct 31 02:42:49.472345 extend-filesystems[1565]: Found vda1 Oct 31 02:42:49.472345 extend-filesystems[1565]: Found vda2 Oct 31 02:42:49.517480 extend-filesystems[1565]: Found vda3 Oct 31 02:42:49.517480 extend-filesystems[1565]: Found usr Oct 31 02:42:49.517480 extend-filesystems[1565]: Found vda4 Oct 31 02:42:49.517480 extend-filesystems[1565]: Found vda6 Oct 31 02:42:49.517480 extend-filesystems[1565]: Found vda7 Oct 31 02:42:49.517480 extend-filesystems[1565]: Found vda9 Oct 31 02:42:49.517480 extend-filesystems[1565]: Checking size of /dev/vda9 Oct 31 02:42:49.475228 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 31 02:42:49.538150 extend-filesystems[1565]: Resized partition /dev/vda9 Oct 31 02:42:49.511206 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 31 02:42:49.521903 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 31 02:42:49.528281 systemd[1]: Starting update-engine.service - Update Engine... Oct 31 02:42:49.541837 extend-filesystems[1597]: resize2fs 1.47.1 (20-May-2024) Oct 31 02:42:49.544145 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 31 02:42:49.549733 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 31 02:42:49.557320 kernel: EXT4-fs (vda9): resizing filesystem from 1617920 to 15121403 blocks Oct 31 02:42:49.558398 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 31 02:42:49.558799 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 31 02:42:49.568729 systemd[1]: motdgen.service: Deactivated successfully. Oct 31 02:42:49.570318 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 31 02:42:49.574637 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 31 02:42:49.584837 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 31 02:42:49.585236 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 31 02:42:49.586534 jq[1599]: true Oct 31 02:42:49.593154 update_engine[1595]: I20251031 02:42:49.593043 1595 main.cc:92] Flatcar Update Engine starting Oct 31 02:42:49.595891 update_engine[1595]: I20251031 02:42:49.595645 1595 update_check_scheduler.cc:74] Next update check in 7m21s Oct 31 02:42:49.633283 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1265) Oct 31 02:42:49.645746 (ntainerd)[1609]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Oct 31 02:42:49.657527 dbus-daemon[1561]: [system] Successfully activated service 'org.freedesktop.systemd1' Oct 31 02:42:49.663931 tar[1604]: linux-amd64/LICENSE Oct 31 02:42:49.663931 tar[1604]: linux-amd64/helm Oct 31 02:42:49.665477 jq[1606]: true Oct 31 02:42:49.691467 systemd-timesyncd[1534]: Contacted time server 139.162.219.252:123 (0.flatcar.pool.ntp.org). Oct 31 02:42:49.691546 systemd-timesyncd[1534]: Initial clock synchronization to Fri 2025-10-31 02:42:49.910057 UTC. Oct 31 02:42:49.698071 systemd[1]: Started update-engine.service - Update Engine. Oct 31 02:42:49.702813 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 31 02:42:49.702854 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 31 02:42:49.716668 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Oct 31 02:42:49.718610 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 31 02:42:49.718646 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 31 02:42:49.720464 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 31 02:42:49.735232 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 31 02:42:49.961064 systemd-logind[1591]: Watching system buttons on /dev/input/event2 (Power Button) Oct 31 02:42:49.961165 systemd-logind[1591]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Oct 31 02:42:49.970518 systemd-logind[1591]: New seat seat0. Oct 31 02:42:49.982251 systemd[1]: Started systemd-logind.service - User Login Management. Oct 31 02:42:50.003556 bash[1640]: Updated "/home/core/.ssh/authorized_keys" Oct 31 02:42:50.009244 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 31 02:42:50.033378 systemd[1]: Starting sshkeys.service... Oct 31 02:42:50.044252 systemd-networkd[1257]: eth0: Ignoring DHCPv6 address 2a02:1348:179:8679:24:19ff:fee6:19e6/128 (valid for 59min 59s, preferred for 59min 59s) which conflicts with 2a02:1348:179:8679:24:19ff:fee6:19e6/64 assigned by NDisc. Oct 31 02:42:50.044665 systemd-networkd[1257]: eth0: Hint: use IPv6Token= setting to change the address generated by NDisc or set UseAutonomousPrefix=no. Oct 31 02:42:50.071273 locksmithd[1624]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 31 02:42:50.115070 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Oct 31 02:42:50.119078 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Oct 31 02:42:50.126985 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Oct 31 02:42:50.137863 extend-filesystems[1597]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Oct 31 02:42:50.137863 extend-filesystems[1597]: old_desc_blocks = 1, new_desc_blocks = 8 Oct 31 02:42:50.137863 extend-filesystems[1597]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Oct 31 02:42:50.150166 extend-filesystems[1565]: Resized filesystem in /dev/vda9 Oct 31 02:42:50.142479 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 31 02:42:50.142849 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 31 02:42:50.266955 dbus-daemon[1561]: [system] Successfully activated service 'org.freedesktop.hostname1' Oct 31 02:42:50.269884 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Oct 31 02:42:50.272315 dbus-daemon[1561]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.8' (uid=0 pid=1623 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Oct 31 02:42:50.283559 systemd[1]: Starting polkit.service - Authorization Manager... Oct 31 02:42:50.325568 polkitd[1661]: Started polkitd version 121 Oct 31 02:42:50.337677 polkitd[1661]: Loading rules from directory /etc/polkit-1/rules.d Oct 31 02:42:50.337776 polkitd[1661]: Loading rules from directory /usr/share/polkit-1/rules.d Oct 31 02:42:50.342094 containerd[1609]: time="2025-10-31T02:42:50.340893787Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Oct 31 02:42:50.344789 polkitd[1661]: Finished loading, compiling and executing 2 rules Oct 31 02:42:50.353107 dbus-daemon[1561]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Oct 31 02:42:50.353330 systemd[1]: Started polkit.service - Authorization Manager. Oct 31 02:42:50.355238 polkitd[1661]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Oct 31 02:42:50.393223 systemd-hostnamed[1623]: Hostname set to (static) Oct 31 02:42:50.456394 containerd[1609]: time="2025-10-31T02:42:50.456330746Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.470532 containerd[1609]: time="2025-10-31T02:42:50.470283458Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.113-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Oct 31 02:42:50.470532 containerd[1609]: time="2025-10-31T02:42:50.470344877Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Oct 31 02:42:50.470532 containerd[1609]: time="2025-10-31T02:42:50.470372707Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Oct 31 02:42:50.470787 containerd[1609]: time="2025-10-31T02:42:50.470668535Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Oct 31 02:42:50.470787 containerd[1609]: time="2025-10-31T02:42:50.470705347Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.470862 containerd[1609]: time="2025-10-31T02:42:50.470814483Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Oct 31 02:42:50.470862 containerd[1609]: time="2025-10-31T02:42:50.470839111Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471146416Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471192476Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471217882Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471235741Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471363672Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471760736Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471959733Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Oct 31 02:42:50.472706 containerd[1609]: time="2025-10-31T02:42:50.471985064Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Oct 31 02:42:50.476387 containerd[1609]: time="2025-10-31T02:42:50.475659063Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Oct 31 02:42:50.476387 containerd[1609]: time="2025-10-31T02:42:50.475769584Z" level=info msg="metadata content store policy set" policy=shared Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.483735128Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.483824779Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.483854809Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.483890547Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.483914517Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.484144817Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Oct 31 02:42:50.484762 containerd[1609]: time="2025-10-31T02:42:50.484607737Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484789239Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484816289Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484836369Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484857207Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484879681Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484909529Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484934418Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484956474Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.484980062Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.485005731Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485072 containerd[1609]: time="2025-10-31T02:42:50.485028111Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485099409Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485126303Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485146164Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485175897Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485197732Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485241227Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485276779Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485309441Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485335578Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485359228Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485380802Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485401036Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485421677Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485446199Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Oct 31 02:42:50.485516 containerd[1609]: time="2025-10-31T02:42:50.485486084Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485510129Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485538154Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485632750Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485668517Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485688682Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485709489Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485726020Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485747619Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485788129Z" level=info msg="NRI interface is disabled by configuration." Oct 31 02:42:50.486109 containerd[1609]: time="2025-10-31T02:42:50.485808631Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Oct 31 02:42:50.489086 containerd[1609]: time="2025-10-31T02:42:50.487477612Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Oct 31 02:42:50.489086 containerd[1609]: time="2025-10-31T02:42:50.487585528Z" level=info msg="Connect containerd service" Oct 31 02:42:50.489086 containerd[1609]: time="2025-10-31T02:42:50.487662713Z" level=info msg="using legacy CRI server" Oct 31 02:42:50.489086 containerd[1609]: time="2025-10-31T02:42:50.487680881Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 31 02:42:50.489086 containerd[1609]: time="2025-10-31T02:42:50.487850647Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Oct 31 02:42:50.489086 containerd[1609]: time="2025-10-31T02:42:50.488855265Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.490712153Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.490802821Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.490894797Z" level=info msg="Start subscribing containerd event" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.490956119Z" level=info msg="Start recovering state" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.491079630Z" level=info msg="Start event monitor" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.491110590Z" level=info msg="Start snapshots syncer" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.491131424Z" level=info msg="Start cni network conf syncer for default" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.491149877Z" level=info msg="Start streaming server" Oct 31 02:42:50.492356 containerd[1609]: time="2025-10-31T02:42:50.491272909Z" level=info msg="containerd successfully booted in 0.157293s" Oct 31 02:42:50.491436 systemd[1]: Started containerd.service - containerd container runtime. Oct 31 02:42:51.142786 sshd_keygen[1603]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 31 02:42:51.176837 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 31 02:42:51.184471 tar[1604]: linux-amd64/README.md Oct 31 02:42:51.197994 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 31 02:42:51.204195 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 31 02:42:51.222987 systemd[1]: issuegen.service: Deactivated successfully. Oct 31 02:42:51.223436 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 31 02:42:51.233119 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 31 02:42:51.256387 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 31 02:42:51.267889 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 31 02:42:51.278295 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Oct 31 02:42:51.280579 systemd[1]: Reached target getty.target - Login Prompts. Oct 31 02:42:51.309269 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:42:51.312009 (kubelet)[1709]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 31 02:42:51.669923 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 31 02:42:51.682012 systemd[1]: Started sshd@0-10.230.25.230:22-147.75.109.163:32860.service - OpenSSH per-connection server daemon (147.75.109.163:32860). Oct 31 02:42:51.942124 kubelet[1709]: E1031 02:42:51.941955 1709 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 31 02:42:51.947325 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 31 02:42:51.947684 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 31 02:42:52.615374 sshd[1714]: Accepted publickey for core from 147.75.109.163 port 32860 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:42:52.617129 sshd[1714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:42:52.635856 systemd-logind[1591]: New session 1 of user core. Oct 31 02:42:52.638963 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 31 02:42:52.659634 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 31 02:42:52.685853 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 31 02:42:52.696832 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 31 02:42:52.714278 (systemd)[1724]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 31 02:42:52.859834 systemd[1724]: Queued start job for default target default.target. Oct 31 02:42:52.860425 systemd[1724]: Created slice app.slice - User Application Slice. Oct 31 02:42:52.860457 systemd[1724]: Reached target paths.target - Paths. Oct 31 02:42:52.860480 systemd[1724]: Reached target timers.target - Timers. Oct 31 02:42:52.867184 systemd[1724]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 31 02:42:52.878752 systemd[1724]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 31 02:42:52.878843 systemd[1724]: Reached target sockets.target - Sockets. Oct 31 02:42:52.878870 systemd[1724]: Reached target basic.target - Basic System. Oct 31 02:42:52.878937 systemd[1724]: Reached target default.target - Main User Target. Oct 31 02:42:52.879002 systemd[1724]: Startup finished in 155ms. Oct 31 02:42:52.879695 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 31 02:42:52.889897 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 31 02:42:53.545617 systemd[1]: Started sshd@1-10.230.25.230:22-147.75.109.163:32866.service - OpenSSH per-connection server daemon (147.75.109.163:32866). Oct 31 02:42:54.458925 sshd[1737]: Accepted publickey for core from 147.75.109.163 port 32866 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:42:54.461381 sshd[1737]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:42:54.469203 systemd-logind[1591]: New session 2 of user core. Oct 31 02:42:54.477770 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 31 02:42:55.098514 sshd[1737]: pam_unix(sshd:session): session closed for user core Oct 31 02:42:55.103795 systemd[1]: sshd@1-10.230.25.230:22-147.75.109.163:32866.service: Deactivated successfully. Oct 31 02:42:55.105194 systemd-logind[1591]: Session 2 logged out. Waiting for processes to exit. Oct 31 02:42:55.110119 systemd[1]: session-2.scope: Deactivated successfully. Oct 31 02:42:55.111334 systemd-logind[1591]: Removed session 2. Oct 31 02:42:55.258652 systemd[1]: Started sshd@2-10.230.25.230:22-147.75.109.163:32876.service - OpenSSH per-connection server daemon (147.75.109.163:32876). Oct 31 02:42:56.161461 sshd[1745]: Accepted publickey for core from 147.75.109.163 port 32876 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:42:56.164525 sshd[1745]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:42:56.171783 systemd-logind[1591]: New session 3 of user core. Oct 31 02:42:56.184588 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 31 02:42:56.314511 login[1700]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Oct 31 02:42:56.323639 systemd-logind[1591]: New session 4 of user core. Oct 31 02:42:56.331635 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 31 02:42:56.353869 login[1699]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Oct 31 02:42:56.361768 systemd-logind[1591]: New session 5 of user core. Oct 31 02:42:56.373396 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 31 02:42:56.514111 coreos-metadata[1560]: Oct 31 02:42:56.513 WARN failed to locate config-drive, using the metadata service API instead Oct 31 02:42:56.541926 coreos-metadata[1560]: Oct 31 02:42:56.541 INFO Fetching http://169.254.169.254/openstack/2012-08-10/meta_data.json: Attempt #1 Oct 31 02:42:56.549920 coreos-metadata[1560]: Oct 31 02:42:56.549 INFO Fetch failed with 404: resource not found Oct 31 02:42:56.549920 coreos-metadata[1560]: Oct 31 02:42:56.549 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Oct 31 02:42:56.550626 coreos-metadata[1560]: Oct 31 02:42:56.550 INFO Fetch successful Oct 31 02:42:56.550843 coreos-metadata[1560]: Oct 31 02:42:56.550 INFO Fetching http://169.254.169.254/latest/meta-data/instance-id: Attempt #1 Oct 31 02:42:56.565363 coreos-metadata[1560]: Oct 31 02:42:56.565 INFO Fetch successful Oct 31 02:42:56.565589 coreos-metadata[1560]: Oct 31 02:42:56.565 INFO Fetching http://169.254.169.254/latest/meta-data/instance-type: Attempt #1 Oct 31 02:42:56.579675 coreos-metadata[1560]: Oct 31 02:42:56.579 INFO Fetch successful Oct 31 02:42:56.579945 coreos-metadata[1560]: Oct 31 02:42:56.579 INFO Fetching http://169.254.169.254/latest/meta-data/local-ipv4: Attempt #1 Oct 31 02:42:56.671128 coreos-metadata[1560]: Oct 31 02:42:56.671 INFO Fetch successful Oct 31 02:42:56.671428 coreos-metadata[1560]: Oct 31 02:42:56.671 INFO Fetching http://169.254.169.254/latest/meta-data/public-ipv4: Attempt #1 Oct 31 02:42:56.691143 coreos-metadata[1560]: Oct 31 02:42:56.691 INFO Fetch successful Oct 31 02:42:56.718904 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Oct 31 02:42:56.720385 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 31 02:42:56.796576 sshd[1745]: pam_unix(sshd:session): session closed for user core Oct 31 02:42:56.801539 systemd[1]: sshd@2-10.230.25.230:22-147.75.109.163:32876.service: Deactivated successfully. Oct 31 02:42:56.805956 systemd[1]: session-3.scope: Deactivated successfully. Oct 31 02:42:56.806311 systemd-logind[1591]: Session 3 logged out. Waiting for processes to exit. Oct 31 02:42:56.809381 systemd-logind[1591]: Removed session 3. Oct 31 02:42:57.308600 coreos-metadata[1655]: Oct 31 02:42:57.308 WARN failed to locate config-drive, using the metadata service API instead Oct 31 02:42:57.332811 coreos-metadata[1655]: Oct 31 02:42:57.332 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys: Attempt #1 Oct 31 02:42:57.357149 coreos-metadata[1655]: Oct 31 02:42:57.357 INFO Fetch successful Oct 31 02:42:57.357665 coreos-metadata[1655]: Oct 31 02:42:57.357 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys/0/openssh-key: Attempt #1 Oct 31 02:42:57.390470 coreos-metadata[1655]: Oct 31 02:42:57.390 INFO Fetch successful Oct 31 02:42:57.393080 unknown[1655]: wrote ssh authorized keys file for user: core Oct 31 02:42:57.412278 update-ssh-keys[1792]: Updated "/home/core/.ssh/authorized_keys" Oct 31 02:42:57.413328 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Oct 31 02:42:57.420891 systemd[1]: Finished sshkeys.service. Oct 31 02:42:57.426785 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 31 02:42:57.427231 systemd[1]: Startup finished in 16.414s (kernel) + 12.629s (userspace) = 29.044s. Oct 31 02:43:02.016359 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 31 02:43:02.024268 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:02.210243 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:02.216640 (kubelet)[1810]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 31 02:43:02.323078 kubelet[1810]: E1031 02:43:02.322401 1810 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 31 02:43:02.326345 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 31 02:43:02.326680 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 31 02:43:07.021396 systemd[1]: Started sshd@3-10.230.25.230:22-147.75.109.163:52638.service - OpenSSH per-connection server daemon (147.75.109.163:52638). Oct 31 02:43:07.929280 sshd[1819]: Accepted publickey for core from 147.75.109.163 port 52638 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:43:07.932176 sshd[1819]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:43:07.939279 systemd-logind[1591]: New session 6 of user core. Oct 31 02:43:07.953747 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 31 02:43:08.564389 sshd[1819]: pam_unix(sshd:session): session closed for user core Oct 31 02:43:08.568755 systemd[1]: sshd@3-10.230.25.230:22-147.75.109.163:52638.service: Deactivated successfully. Oct 31 02:43:08.569379 systemd-logind[1591]: Session 6 logged out. Waiting for processes to exit. Oct 31 02:43:08.572674 systemd[1]: session-6.scope: Deactivated successfully. Oct 31 02:43:08.573495 systemd-logind[1591]: Removed session 6. Oct 31 02:43:08.716395 systemd[1]: Started sshd@4-10.230.25.230:22-147.75.109.163:52646.service - OpenSSH per-connection server daemon (147.75.109.163:52646). Oct 31 02:43:09.611108 sshd[1827]: Accepted publickey for core from 147.75.109.163 port 52646 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:43:09.613354 sshd[1827]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:43:09.621015 systemd-logind[1591]: New session 7 of user core. Oct 31 02:43:09.628699 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 31 02:43:10.232519 sshd[1827]: pam_unix(sshd:session): session closed for user core Oct 31 02:43:10.238791 systemd[1]: sshd@4-10.230.25.230:22-147.75.109.163:52646.service: Deactivated successfully. Oct 31 02:43:10.238942 systemd-logind[1591]: Session 7 logged out. Waiting for processes to exit. Oct 31 02:43:10.242506 systemd[1]: session-7.scope: Deactivated successfully. Oct 31 02:43:10.243849 systemd-logind[1591]: Removed session 7. Oct 31 02:43:10.388434 systemd[1]: Started sshd@5-10.230.25.230:22-147.75.109.163:41736.service - OpenSSH per-connection server daemon (147.75.109.163:41736). Oct 31 02:43:11.288430 sshd[1835]: Accepted publickey for core from 147.75.109.163 port 41736 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:43:11.290445 sshd[1835]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:43:11.296948 systemd-logind[1591]: New session 8 of user core. Oct 31 02:43:11.305603 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 31 02:43:11.923342 sshd[1835]: pam_unix(sshd:session): session closed for user core Oct 31 02:43:11.927576 systemd-logind[1591]: Session 8 logged out. Waiting for processes to exit. Oct 31 02:43:11.928963 systemd[1]: sshd@5-10.230.25.230:22-147.75.109.163:41736.service: Deactivated successfully. Oct 31 02:43:11.933944 systemd[1]: session-8.scope: Deactivated successfully. Oct 31 02:43:11.934864 systemd-logind[1591]: Removed session 8. Oct 31 02:43:12.076392 systemd[1]: Started sshd@6-10.230.25.230:22-147.75.109.163:41752.service - OpenSSH per-connection server daemon (147.75.109.163:41752). Oct 31 02:43:12.515987 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Oct 31 02:43:12.526473 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:12.687261 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:12.701678 (kubelet)[1856]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 31 02:43:12.764973 kubelet[1856]: E1031 02:43:12.764877 1856 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 31 02:43:12.769301 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 31 02:43:12.769652 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 31 02:43:12.984693 sshd[1843]: Accepted publickey for core from 147.75.109.163 port 41752 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:43:12.986755 sshd[1843]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:43:12.995153 systemd-logind[1591]: New session 9 of user core. Oct 31 02:43:13.002552 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 31 02:43:13.480422 sudo[1867]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 31 02:43:13.480954 sudo[1867]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 31 02:43:13.495808 sudo[1867]: pam_unix(sudo:session): session closed for user root Oct 31 02:43:13.642240 sshd[1843]: pam_unix(sshd:session): session closed for user core Oct 31 02:43:13.647655 systemd[1]: sshd@6-10.230.25.230:22-147.75.109.163:41752.service: Deactivated successfully. Oct 31 02:43:13.652307 systemd[1]: session-9.scope: Deactivated successfully. Oct 31 02:43:13.653498 systemd-logind[1591]: Session 9 logged out. Waiting for processes to exit. Oct 31 02:43:13.654908 systemd-logind[1591]: Removed session 9. Oct 31 02:43:13.796367 systemd[1]: Started sshd@7-10.230.25.230:22-147.75.109.163:41762.service - OpenSSH per-connection server daemon (147.75.109.163:41762). Oct 31 02:43:14.693208 sshd[1872]: Accepted publickey for core from 147.75.109.163 port 41762 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:43:14.695310 sshd[1872]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:43:14.702439 systemd-logind[1591]: New session 10 of user core. Oct 31 02:43:14.709442 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 31 02:43:15.176177 sudo[1877]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 31 02:43:15.176646 sudo[1877]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 31 02:43:15.181766 sudo[1877]: pam_unix(sudo:session): session closed for user root Oct 31 02:43:15.189665 sudo[1876]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Oct 31 02:43:15.190123 sudo[1876]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 31 02:43:15.208349 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Oct 31 02:43:15.214054 auditctl[1880]: No rules Oct 31 02:43:15.214608 systemd[1]: audit-rules.service: Deactivated successfully. Oct 31 02:43:15.214953 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Oct 31 02:43:15.224487 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Oct 31 02:43:15.259660 augenrules[1899]: No rules Oct 31 02:43:15.260570 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Oct 31 02:43:15.262158 sudo[1876]: pam_unix(sudo:session): session closed for user root Oct 31 02:43:15.410239 sshd[1872]: pam_unix(sshd:session): session closed for user core Oct 31 02:43:15.414465 systemd-logind[1591]: Session 10 logged out. Waiting for processes to exit. Oct 31 02:43:15.415706 systemd[1]: sshd@7-10.230.25.230:22-147.75.109.163:41762.service: Deactivated successfully. Oct 31 02:43:15.419554 systemd[1]: session-10.scope: Deactivated successfully. Oct 31 02:43:15.421944 systemd-logind[1591]: Removed session 10. Oct 31 02:43:15.565458 systemd[1]: Started sshd@8-10.230.25.230:22-147.75.109.163:41778.service - OpenSSH per-connection server daemon (147.75.109.163:41778). Oct 31 02:43:16.469193 sshd[1908]: Accepted publickey for core from 147.75.109.163 port 41778 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:43:16.471161 sshd[1908]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:43:16.477336 systemd-logind[1591]: New session 11 of user core. Oct 31 02:43:16.485503 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 31 02:43:16.957002 sudo[1912]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 31 02:43:16.958304 sudo[1912]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 31 02:43:17.424483 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 31 02:43:17.424880 (dockerd)[1928]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 31 02:43:17.865257 dockerd[1928]: time="2025-10-31T02:43:17.864304977Z" level=info msg="Starting up" Oct 31 02:43:17.988249 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2595510470-merged.mount: Deactivated successfully. Oct 31 02:43:18.159428 dockerd[1928]: time="2025-10-31T02:43:18.159222663Z" level=info msg="Loading containers: start." Oct 31 02:43:18.304177 kernel: Initializing XFRM netlink socket Oct 31 02:43:18.412264 systemd-networkd[1257]: docker0: Link UP Oct 31 02:43:18.434907 dockerd[1928]: time="2025-10-31T02:43:18.434863755Z" level=info msg="Loading containers: done." Oct 31 02:43:18.455061 dockerd[1928]: time="2025-10-31T02:43:18.454737871Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 31 02:43:18.455061 dockerd[1928]: time="2025-10-31T02:43:18.454925822Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Oct 31 02:43:18.455326 dockerd[1928]: time="2025-10-31T02:43:18.455148627Z" level=info msg="Daemon has completed initialization" Oct 31 02:43:18.492554 dockerd[1928]: time="2025-10-31T02:43:18.491760257Z" level=info msg="API listen on /run/docker.sock" Oct 31 02:43:18.492415 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 31 02:43:19.637206 containerd[1609]: time="2025-10-31T02:43:19.637095982Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Oct 31 02:43:20.420104 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Oct 31 02:43:20.557534 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount156030028.mount: Deactivated successfully. Oct 31 02:43:22.940562 containerd[1609]: time="2025-10-31T02:43:22.940287177Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:22.944006 containerd[1609]: time="2025-10-31T02:43:22.943944108Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837924" Oct 31 02:43:22.946652 containerd[1609]: time="2025-10-31T02:43:22.945300474Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:22.950342 containerd[1609]: time="2025-10-31T02:43:22.950256436Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:22.952468 containerd[1609]: time="2025-10-31T02:43:22.951331424Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 3.314111773s" Oct 31 02:43:22.952468 containerd[1609]: time="2025-10-31T02:43:22.951406104Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Oct 31 02:43:22.954930 containerd[1609]: time="2025-10-31T02:43:22.954627955Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Oct 31 02:43:23.016151 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Oct 31 02:43:23.033349 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:23.217307 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:23.222941 (kubelet)[2147]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 31 02:43:23.331910 kubelet[2147]: E1031 02:43:23.331822 2147 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 31 02:43:23.334590 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 31 02:43:23.335016 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 31 02:43:25.333108 containerd[1609]: time="2025-10-31T02:43:25.332242670Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:25.334819 containerd[1609]: time="2025-10-31T02:43:25.334488558Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787035" Oct 31 02:43:25.336512 containerd[1609]: time="2025-10-31T02:43:25.335714765Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:25.340059 containerd[1609]: time="2025-10-31T02:43:25.340003775Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:25.341803 containerd[1609]: time="2025-10-31T02:43:25.341759476Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 2.387058716s" Oct 31 02:43:25.341927 containerd[1609]: time="2025-10-31T02:43:25.341855923Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Oct 31 02:43:25.343060 containerd[1609]: time="2025-10-31T02:43:25.342993788Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Oct 31 02:43:28.921453 containerd[1609]: time="2025-10-31T02:43:28.921331340Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:28.923477 containerd[1609]: time="2025-10-31T02:43:28.923115543Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176297" Oct 31 02:43:28.924947 containerd[1609]: time="2025-10-31T02:43:28.924274491Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:28.928476 containerd[1609]: time="2025-10-31T02:43:28.928438099Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:28.930400 containerd[1609]: time="2025-10-31T02:43:28.930361228Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 3.587150482s" Oct 31 02:43:28.930542 containerd[1609]: time="2025-10-31T02:43:28.930513821Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Oct 31 02:43:28.932480 containerd[1609]: time="2025-10-31T02:43:28.932443610Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Oct 31 02:43:30.827114 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4129280428.mount: Deactivated successfully. Oct 31 02:43:31.527226 containerd[1609]: time="2025-10-31T02:43:31.527046449Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:31.528791 containerd[1609]: time="2025-10-31T02:43:31.528605442Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924214" Oct 31 02:43:31.529967 containerd[1609]: time="2025-10-31T02:43:31.529900020Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:31.534853 containerd[1609]: time="2025-10-31T02:43:31.534435182Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:31.535594 containerd[1609]: time="2025-10-31T02:43:31.535541987Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 2.602952542s" Oct 31 02:43:31.535681 containerd[1609]: time="2025-10-31T02:43:31.535631629Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Oct 31 02:43:31.537432 containerd[1609]: time="2025-10-31T02:43:31.537390343Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Oct 31 02:43:32.292155 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount721070686.mount: Deactivated successfully. Oct 31 02:43:33.516275 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Oct 31 02:43:33.528459 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:33.856811 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:33.860516 containerd[1609]: time="2025-10-31T02:43:33.860459104Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:33.865074 containerd[1609]: time="2025-10-31T02:43:33.864991515Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565249" Oct 31 02:43:33.866078 containerd[1609]: time="2025-10-31T02:43:33.865938189Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:33.869628 (kubelet)[2235]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 31 02:43:33.872944 containerd[1609]: time="2025-10-31T02:43:33.872313859Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:33.881848 containerd[1609]: time="2025-10-31T02:43:33.881794639Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 2.344356598s" Oct 31 02:43:33.882746 containerd[1609]: time="2025-10-31T02:43:33.882714582Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Oct 31 02:43:33.885536 containerd[1609]: time="2025-10-31T02:43:33.885498588Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Oct 31 02:43:33.945798 kubelet[2235]: E1031 02:43:33.945732 2235 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 31 02:43:33.948741 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 31 02:43:33.949164 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 31 02:43:34.398625 update_engine[1595]: I20251031 02:43:34.398404 1595 update_attempter.cc:509] Updating boot flags... Oct 31 02:43:34.447378 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2250) Oct 31 02:43:34.530182 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2248) Oct 31 02:43:34.607333 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1436190245.mount: Deactivated successfully. Oct 31 02:43:34.628246 containerd[1609]: time="2025-10-31T02:43:34.626985706Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:34.628246 containerd[1609]: time="2025-10-31T02:43:34.628183204Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321146" Oct 31 02:43:34.628686 containerd[1609]: time="2025-10-31T02:43:34.628652003Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:34.631682 containerd[1609]: time="2025-10-31T02:43:34.631645003Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:34.633462 containerd[1609]: time="2025-10-31T02:43:34.633427953Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 747.882468ms" Oct 31 02:43:34.634115 containerd[1609]: time="2025-10-31T02:43:34.634085345Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Oct 31 02:43:34.635053 containerd[1609]: time="2025-10-31T02:43:34.635001122Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Oct 31 02:43:35.472110 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount923240046.mount: Deactivated successfully. Oct 31 02:43:44.016896 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Oct 31 02:43:44.031110 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:44.346282 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:44.353638 (kubelet)[2325]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 31 02:43:44.456052 kubelet[2325]: E1031 02:43:44.455705 2325 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 31 02:43:44.459769 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 31 02:43:44.460124 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 31 02:43:44.644149 containerd[1609]: time="2025-10-31T02:43:44.643949928Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:44.645895 containerd[1609]: time="2025-10-31T02:43:44.645846482Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682064" Oct 31 02:43:44.648043 containerd[1609]: time="2025-10-31T02:43:44.646369804Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:44.650737 containerd[1609]: time="2025-10-31T02:43:44.650700441Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:43:44.652757 containerd[1609]: time="2025-10-31T02:43:44.652719683Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 10.017642759s" Oct 31 02:43:44.652934 containerd[1609]: time="2025-10-31T02:43:44.652903500Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Oct 31 02:43:49.424608 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:49.437823 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:49.483779 systemd[1]: Reloading requested from client PID 2363 ('systemctl') (unit session-11.scope)... Oct 31 02:43:49.484138 systemd[1]: Reloading... Oct 31 02:43:49.635090 zram_generator::config[2399]: No configuration found. Oct 31 02:43:49.843819 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 31 02:43:49.951973 systemd[1]: Reloading finished in 466 ms. Oct 31 02:43:50.015479 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 31 02:43:50.015888 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 31 02:43:50.016570 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:50.029578 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:50.413267 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:50.423671 (kubelet)[2479]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 31 02:43:50.520430 kubelet[2479]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 31 02:43:50.521865 kubelet[2479]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 31 02:43:50.521865 kubelet[2479]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 31 02:43:50.521865 kubelet[2479]: I1031 02:43:50.521225 2479 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 31 02:43:51.331722 kubelet[2479]: I1031 02:43:51.331633 2479 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 31 02:43:51.331722 kubelet[2479]: I1031 02:43:51.331681 2479 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 31 02:43:51.332089 kubelet[2479]: I1031 02:43:51.332060 2479 server.go:954] "Client rotation is on, will bootstrap in background" Oct 31 02:43:51.372055 kubelet[2479]: E1031 02:43:51.371008 2479 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.230.25.230:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:51.372332 kubelet[2479]: I1031 02:43:51.372298 2479 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 31 02:43:51.390404 kubelet[2479]: E1031 02:43:51.390365 2479 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Oct 31 02:43:51.390599 kubelet[2479]: I1031 02:43:51.390574 2479 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Oct 31 02:43:51.398597 kubelet[2479]: I1031 02:43:51.398573 2479 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 31 02:43:51.401756 kubelet[2479]: I1031 02:43:51.401709 2479 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 31 02:43:51.402247 kubelet[2479]: I1031 02:43:51.401877 2479 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"srv-4lokn.gb1.brightbox.com","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Oct 31 02:43:51.402591 kubelet[2479]: I1031 02:43:51.402568 2479 topology_manager.go:138] "Creating topology manager with none policy" Oct 31 02:43:51.402698 kubelet[2479]: I1031 02:43:51.402681 2479 container_manager_linux.go:304] "Creating device plugin manager" Oct 31 02:43:51.403053 kubelet[2479]: I1031 02:43:51.403004 2479 state_mem.go:36] "Initialized new in-memory state store" Oct 31 02:43:51.407087 kubelet[2479]: I1031 02:43:51.407065 2479 kubelet.go:446] "Attempting to sync node with API server" Oct 31 02:43:51.407324 kubelet[2479]: I1031 02:43:51.407280 2479 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 31 02:43:51.407396 kubelet[2479]: I1031 02:43:51.407336 2479 kubelet.go:352] "Adding apiserver pod source" Oct 31 02:43:51.407396 kubelet[2479]: I1031 02:43:51.407360 2479 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 31 02:43:51.423345 kubelet[2479]: I1031 02:43:51.423153 2479 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Oct 31 02:43:51.428056 kubelet[2479]: W1031 02:43:51.426649 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.230.25.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:51.428056 kubelet[2479]: E1031 02:43:51.426727 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.230.25.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:51.428056 kubelet[2479]: W1031 02:43:51.426832 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.230.25.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-4lokn.gb1.brightbox.com&limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:51.428056 kubelet[2479]: E1031 02:43:51.426882 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.230.25.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-4lokn.gb1.brightbox.com&limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:51.428056 kubelet[2479]: I1031 02:43:51.427873 2479 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 31 02:43:51.428766 kubelet[2479]: W1031 02:43:51.428742 2479 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 31 02:43:51.429944 kubelet[2479]: I1031 02:43:51.429921 2479 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 31 02:43:51.430116 kubelet[2479]: I1031 02:43:51.430096 2479 server.go:1287] "Started kubelet" Oct 31 02:43:51.431554 kubelet[2479]: I1031 02:43:51.431336 2479 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 31 02:43:51.433782 kubelet[2479]: I1031 02:43:51.433383 2479 server.go:479] "Adding debug handlers to kubelet server" Oct 31 02:43:51.436844 kubelet[2479]: I1031 02:43:51.436760 2479 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 31 02:43:51.437358 kubelet[2479]: I1031 02:43:51.437335 2479 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 31 02:43:51.441878 kubelet[2479]: I1031 02:43:51.441733 2479 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 31 02:43:51.444422 kubelet[2479]: E1031 02:43:51.438881 2479 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.230.25.230:6443/api/v1/namespaces/default/events\": dial tcp 10.230.25.230:6443: connect: connection refused" event="&Event{ObjectMeta:{srv-4lokn.gb1.brightbox.com.18737347cbbd7c76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:srv-4lokn.gb1.brightbox.com,UID:srv-4lokn.gb1.brightbox.com,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:srv-4lokn.gb1.brightbox.com,},FirstTimestamp:2025-10-31 02:43:51.430061174 +0000 UTC m=+0.999957967,LastTimestamp:2025-10-31 02:43:51.430061174 +0000 UTC m=+0.999957967,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:srv-4lokn.gb1.brightbox.com,}" Oct 31 02:43:51.446449 kubelet[2479]: I1031 02:43:51.446421 2479 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 31 02:43:51.450704 kubelet[2479]: I1031 02:43:51.450673 2479 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 31 02:43:51.451043 kubelet[2479]: E1031 02:43:51.450993 2479 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"srv-4lokn.gb1.brightbox.com\" not found" Oct 31 02:43:51.451503 kubelet[2479]: I1031 02:43:51.451478 2479 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 31 02:43:51.451588 kubelet[2479]: I1031 02:43:51.451569 2479 reconciler.go:26] "Reconciler: start to sync state" Oct 31 02:43:51.453497 kubelet[2479]: W1031 02:43:51.453444 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.230.25.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:51.454403 kubelet[2479]: E1031 02:43:51.454233 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.230.25.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:51.455347 kubelet[2479]: E1031 02:43:51.455310 2479 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.25.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-4lokn.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.25.230:6443: connect: connection refused" interval="200ms" Oct 31 02:43:51.456312 kubelet[2479]: I1031 02:43:51.456280 2479 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 31 02:43:51.460490 kubelet[2479]: E1031 02:43:51.460453 2479 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 31 02:43:51.461205 kubelet[2479]: I1031 02:43:51.461168 2479 factory.go:221] Registration of the containerd container factory successfully Oct 31 02:43:51.461205 kubelet[2479]: I1031 02:43:51.461203 2479 factory.go:221] Registration of the systemd container factory successfully Oct 31 02:43:51.484057 kubelet[2479]: I1031 02:43:51.483816 2479 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 31 02:43:51.487792 kubelet[2479]: I1031 02:43:51.487752 2479 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 31 02:43:51.487884 kubelet[2479]: I1031 02:43:51.487849 2479 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 31 02:43:51.487928 kubelet[2479]: I1031 02:43:51.487893 2479 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 31 02:43:51.487928 kubelet[2479]: I1031 02:43:51.487909 2479 kubelet.go:2382] "Starting kubelet main sync loop" Oct 31 02:43:51.488066 kubelet[2479]: E1031 02:43:51.488005 2479 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 31 02:43:51.491852 kubelet[2479]: W1031 02:43:51.490489 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.230.25.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:51.491852 kubelet[2479]: E1031 02:43:51.490555 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.230.25.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:51.524493 kubelet[2479]: I1031 02:43:51.524458 2479 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 31 02:43:51.524493 kubelet[2479]: I1031 02:43:51.524486 2479 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 31 02:43:51.525133 kubelet[2479]: I1031 02:43:51.524521 2479 state_mem.go:36] "Initialized new in-memory state store" Oct 31 02:43:51.526885 kubelet[2479]: I1031 02:43:51.526853 2479 policy_none.go:49] "None policy: Start" Oct 31 02:43:51.526968 kubelet[2479]: I1031 02:43:51.526903 2479 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 31 02:43:51.526968 kubelet[2479]: I1031 02:43:51.526942 2479 state_mem.go:35] "Initializing new in-memory state store" Oct 31 02:43:51.537824 kubelet[2479]: I1031 02:43:51.537705 2479 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 31 02:43:51.542098 kubelet[2479]: I1031 02:43:51.542073 2479 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 31 02:43:51.542202 kubelet[2479]: I1031 02:43:51.542131 2479 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 31 02:43:51.542703 kubelet[2479]: I1031 02:43:51.542681 2479 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 31 02:43:51.544960 kubelet[2479]: E1031 02:43:51.544922 2479 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 31 02:43:51.545267 kubelet[2479]: E1031 02:43:51.545012 2479 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"srv-4lokn.gb1.brightbox.com\" not found" Oct 31 02:43:51.602002 kubelet[2479]: E1031 02:43:51.599674 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.604568 kubelet[2479]: E1031 02:43:51.603856 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.607725 kubelet[2479]: E1031 02:43:51.607401 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.645233 kubelet[2479]: I1031 02:43:51.645185 2479 kubelet_node_status.go:75] "Attempting to register node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.646041 kubelet[2479]: E1031 02:43:51.645979 2479 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.25.230:6443/api/v1/nodes\": dial tcp 10.230.25.230:6443: connect: connection refused" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654239 kubelet[2479]: I1031 02:43:51.654057 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-usr-share-ca-certificates\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654239 kubelet[2479]: I1031 02:43:51.654110 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3ae349f67b9e50e769a8f49fcbc622f1-kubeconfig\") pod \"kube-scheduler-srv-4lokn.gb1.brightbox.com\" (UID: \"3ae349f67b9e50e769a8f49fcbc622f1\") " pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654239 kubelet[2479]: I1031 02:43:51.654170 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3e2368e63ed376cc7b2900b2238f34c7-k8s-certs\") pod \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" (UID: \"3e2368e63ed376cc7b2900b2238f34c7\") " pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654239 kubelet[2479]: I1031 02:43:51.654199 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-k8s-certs\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654239 kubelet[2479]: I1031 02:43:51.654242 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-kubeconfig\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654537 kubelet[2479]: I1031 02:43:51.654271 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-flexvolume-dir\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654537 kubelet[2479]: I1031 02:43:51.654296 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3e2368e63ed376cc7b2900b2238f34c7-ca-certs\") pod \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" (UID: \"3e2368e63ed376cc7b2900b2238f34c7\") " pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654537 kubelet[2479]: I1031 02:43:51.654343 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3e2368e63ed376cc7b2900b2238f34c7-usr-share-ca-certificates\") pod \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" (UID: \"3e2368e63ed376cc7b2900b2238f34c7\") " pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.654537 kubelet[2479]: I1031 02:43:51.654370 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-ca-certs\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.656138 kubelet[2479]: E1031 02:43:51.656091 2479 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.25.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-4lokn.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.25.230:6443: connect: connection refused" interval="400ms" Oct 31 02:43:51.853981 kubelet[2479]: I1031 02:43:51.852605 2479 kubelet_node_status.go:75] "Attempting to register node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.853981 kubelet[2479]: E1031 02:43:51.853035 2479 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.25.230:6443/api/v1/nodes\": dial tcp 10.230.25.230:6443: connect: connection refused" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:51.902694 containerd[1609]: time="2025-10-31T02:43:51.902057148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-srv-4lokn.gb1.brightbox.com,Uid:3e2368e63ed376cc7b2900b2238f34c7,Namespace:kube-system,Attempt:0,}" Oct 31 02:43:51.909777 containerd[1609]: time="2025-10-31T02:43:51.909728598Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-srv-4lokn.gb1.brightbox.com,Uid:3ae349f67b9e50e769a8f49fcbc622f1,Namespace:kube-system,Attempt:0,}" Oct 31 02:43:51.910161 containerd[1609]: time="2025-10-31T02:43:51.909733760Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-srv-4lokn.gb1.brightbox.com,Uid:459c05ce4d751e27798a6b5386a0f2aa,Namespace:kube-system,Attempt:0,}" Oct 31 02:43:52.057428 kubelet[2479]: E1031 02:43:52.057362 2479 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.25.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-4lokn.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.25.230:6443: connect: connection refused" interval="800ms" Oct 31 02:43:52.258620 kubelet[2479]: I1031 02:43:52.257293 2479 kubelet_node_status.go:75] "Attempting to register node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:52.258620 kubelet[2479]: E1031 02:43:52.258008 2479 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.25.230:6443/api/v1/nodes\": dial tcp 10.230.25.230:6443: connect: connection refused" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:52.621994 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3777190910.mount: Deactivated successfully. Oct 31 02:43:52.627054 containerd[1609]: time="2025-10-31T02:43:52.626762743Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 31 02:43:52.628341 containerd[1609]: time="2025-10-31T02:43:52.628239695Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Oct 31 02:43:52.628845 containerd[1609]: time="2025-10-31T02:43:52.628798651Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 31 02:43:52.629897 containerd[1609]: time="2025-10-31T02:43:52.629852079Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 31 02:43:52.631247 containerd[1609]: time="2025-10-31T02:43:52.631118470Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Oct 31 02:43:52.632125 containerd[1609]: time="2025-10-31T02:43:52.632086054Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312064" Oct 31 02:43:52.632362 containerd[1609]: time="2025-10-31T02:43:52.632299853Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 31 02:43:52.635152 containerd[1609]: time="2025-10-31T02:43:52.635093299Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 31 02:43:52.638855 containerd[1609]: time="2025-10-31T02:43:52.638563894Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 728.752234ms" Oct 31 02:43:52.643064 containerd[1609]: time="2025-10-31T02:43:52.642590984Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 732.69129ms" Oct 31 02:43:52.643401 containerd[1609]: time="2025-10-31T02:43:52.643318740Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 741.045104ms" Oct 31 02:43:52.687218 kubelet[2479]: W1031 02:43:52.685913 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.230.25.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:52.688005 kubelet[2479]: E1031 02:43:52.687840 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.230.25.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:52.721136 kubelet[2479]: W1031 02:43:52.719189 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.230.25.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-4lokn.gb1.brightbox.com&limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:52.721136 kubelet[2479]: E1031 02:43:52.719284 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.230.25.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-4lokn.gb1.brightbox.com&limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:52.829499 containerd[1609]: time="2025-10-31T02:43:52.826987602Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:43:52.829499 containerd[1609]: time="2025-10-31T02:43:52.827102960Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:43:52.829499 containerd[1609]: time="2025-10-31T02:43:52.827141956Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:43:52.829499 containerd[1609]: time="2025-10-31T02:43:52.827427390Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:43:52.835159 kubelet[2479]: W1031 02:43:52.835093 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.230.25.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:52.836107 kubelet[2479]: E1031 02:43:52.836069 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.230.25.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:52.850773 containerd[1609]: time="2025-10-31T02:43:52.850632033Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:43:52.850773 containerd[1609]: time="2025-10-31T02:43:52.850717518Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:43:52.850773 containerd[1609]: time="2025-10-31T02:43:52.850737524Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:43:52.851768 containerd[1609]: time="2025-10-31T02:43:52.851710036Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:43:52.853593 containerd[1609]: time="2025-10-31T02:43:52.852934287Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:43:52.853593 containerd[1609]: time="2025-10-31T02:43:52.853010089Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:43:52.853593 containerd[1609]: time="2025-10-31T02:43:52.853234837Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:43:52.854895 containerd[1609]: time="2025-10-31T02:43:52.854754494Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:43:52.858569 kubelet[2479]: E1031 02:43:52.858512 2479 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.25.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-4lokn.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.25.230:6443: connect: connection refused" interval="1.6s" Oct 31 02:43:52.979099 containerd[1609]: time="2025-10-31T02:43:52.977826806Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-srv-4lokn.gb1.brightbox.com,Uid:3e2368e63ed376cc7b2900b2238f34c7,Namespace:kube-system,Attempt:0,} returns sandbox id \"bec014eeab45122ab075e8fa68aa6bd1ad7fb0a2d11dd4db5586248626a287a4\"" Oct 31 02:43:52.995488 containerd[1609]: time="2025-10-31T02:43:52.995109319Z" level=info msg="CreateContainer within sandbox \"bec014eeab45122ab075e8fa68aa6bd1ad7fb0a2d11dd4db5586248626a287a4\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 31 02:43:53.013658 containerd[1609]: time="2025-10-31T02:43:53.013609484Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-srv-4lokn.gb1.brightbox.com,Uid:459c05ce4d751e27798a6b5386a0f2aa,Namespace:kube-system,Attempt:0,} returns sandbox id \"f9049aa02fcbbab853c4a61bcd99eef359c1860d08fb1ce2cfde250800be88ae\"" Oct 31 02:43:53.018333 containerd[1609]: time="2025-10-31T02:43:53.018286035Z" level=info msg="CreateContainer within sandbox \"f9049aa02fcbbab853c4a61bcd99eef359c1860d08fb1ce2cfde250800be88ae\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 31 02:43:53.023201 containerd[1609]: time="2025-10-31T02:43:53.023166429Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-srv-4lokn.gb1.brightbox.com,Uid:3ae349f67b9e50e769a8f49fcbc622f1,Namespace:kube-system,Attempt:0,} returns sandbox id \"f2132af9d364a070460eeb661262c28124b4e51c500554f2d54b4f2cb8d2e111\"" Oct 31 02:43:53.026792 containerd[1609]: time="2025-10-31T02:43:53.026758722Z" level=info msg="CreateContainer within sandbox \"f2132af9d364a070460eeb661262c28124b4e51c500554f2d54b4f2cb8d2e111\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 31 02:43:53.030701 kubelet[2479]: W1031 02:43:53.030643 2479 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.230.25.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.230.25.230:6443: connect: connection refused Oct 31 02:43:53.030948 kubelet[2479]: E1031 02:43:53.030885 2479 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.230.25.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:53.033614 containerd[1609]: time="2025-10-31T02:43:53.033581914Z" level=info msg="CreateContainer within sandbox \"bec014eeab45122ab075e8fa68aa6bd1ad7fb0a2d11dd4db5586248626a287a4\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"49dd1720b359419d34ec2723b94e0aa1b56134ed1a3e6c4c6fce3c4c320d2e01\"" Oct 31 02:43:53.034358 containerd[1609]: time="2025-10-31T02:43:53.034313283Z" level=info msg="StartContainer for \"49dd1720b359419d34ec2723b94e0aa1b56134ed1a3e6c4c6fce3c4c320d2e01\"" Oct 31 02:43:53.037006 containerd[1609]: time="2025-10-31T02:43:53.036970369Z" level=info msg="CreateContainer within sandbox \"f9049aa02fcbbab853c4a61bcd99eef359c1860d08fb1ce2cfde250800be88ae\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"30a1b72244a7f2aa4957d3153c14b2de3b4fafe5401b7d2d19e8723a5d55c68f\"" Oct 31 02:43:53.038049 containerd[1609]: time="2025-10-31T02:43:53.037518847Z" level=info msg="StartContainer for \"30a1b72244a7f2aa4957d3153c14b2de3b4fafe5401b7d2d19e8723a5d55c68f\"" Oct 31 02:43:53.064591 kubelet[2479]: I1031 02:43:53.064552 2479 kubelet_node_status.go:75] "Attempting to register node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:53.066237 kubelet[2479]: E1031 02:43:53.065862 2479 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.230.25.230:6443/api/v1/nodes\": dial tcp 10.230.25.230:6443: connect: connection refused" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:53.068286 containerd[1609]: time="2025-10-31T02:43:53.068229578Z" level=info msg="CreateContainer within sandbox \"f2132af9d364a070460eeb661262c28124b4e51c500554f2d54b4f2cb8d2e111\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"bb68d693a9da8ae915784283c3532b31679ef90ce7be626312eec4b18957a790\"" Oct 31 02:43:53.069226 containerd[1609]: time="2025-10-31T02:43:53.069191223Z" level=info msg="StartContainer for \"bb68d693a9da8ae915784283c3532b31679ef90ce7be626312eec4b18957a790\"" Oct 31 02:43:53.212054 containerd[1609]: time="2025-10-31T02:43:53.211294496Z" level=info msg="StartContainer for \"30a1b72244a7f2aa4957d3153c14b2de3b4fafe5401b7d2d19e8723a5d55c68f\" returns successfully" Oct 31 02:43:53.212383 containerd[1609]: time="2025-10-31T02:43:53.212250177Z" level=info msg="StartContainer for \"49dd1720b359419d34ec2723b94e0aa1b56134ed1a3e6c4c6fce3c4c320d2e01\" returns successfully" Oct 31 02:43:53.253496 containerd[1609]: time="2025-10-31T02:43:53.253293458Z" level=info msg="StartContainer for \"bb68d693a9da8ae915784283c3532b31679ef90ce7be626312eec4b18957a790\" returns successfully" Oct 31 02:43:53.506260 kubelet[2479]: E1031 02:43:53.505291 2479 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.230.25.230:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.230.25.230:6443: connect: connection refused" logger="UnhandledError" Oct 31 02:43:53.515762 kubelet[2479]: E1031 02:43:53.514318 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:53.520723 kubelet[2479]: E1031 02:43:53.520456 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:53.520723 kubelet[2479]: E1031 02:43:53.520470 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:54.530075 kubelet[2479]: E1031 02:43:54.528971 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:54.533173 kubelet[2479]: E1031 02:43:54.532472 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:54.673063 kubelet[2479]: I1031 02:43:54.672385 2479 kubelet_node_status.go:75] "Attempting to register node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:55.533310 kubelet[2479]: E1031 02:43:55.533237 2479 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.379007 kubelet[2479]: E1031 02:43:56.378468 2479 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"srv-4lokn.gb1.brightbox.com\" not found" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.425397 kubelet[2479]: I1031 02:43:56.425007 2479 apiserver.go:52] "Watching apiserver" Oct 31 02:43:56.452485 kubelet[2479]: I1031 02:43:56.452404 2479 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 31 02:43:56.481469 kubelet[2479]: E1031 02:43:56.481243 2479 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{srv-4lokn.gb1.brightbox.com.18737347cbbd7c76 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:srv-4lokn.gb1.brightbox.com,UID:srv-4lokn.gb1.brightbox.com,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:srv-4lokn.gb1.brightbox.com,},FirstTimestamp:2025-10-31 02:43:51.430061174 +0000 UTC m=+0.999957967,LastTimestamp:2025-10-31 02:43:51.430061174 +0000 UTC m=+0.999957967,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:srv-4lokn.gb1.brightbox.com,}" Oct 31 02:43:56.548055 kubelet[2479]: I1031 02:43:56.545888 2479 kubelet_node_status.go:78] "Successfully registered node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.548055 kubelet[2479]: E1031 02:43:56.545957 2479 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"srv-4lokn.gb1.brightbox.com\": node \"srv-4lokn.gb1.brightbox.com\" not found" Oct 31 02:43:56.554892 kubelet[2479]: I1031 02:43:56.554853 2479 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.579169 kubelet[2479]: E1031 02:43:56.579120 2479 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.579169 kubelet[2479]: I1031 02:43:56.579162 2479 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.583818 kubelet[2479]: E1031 02:43:56.583781 2479 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-srv-4lokn.gb1.brightbox.com\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.585741 kubelet[2479]: I1031 02:43:56.585716 2479 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:56.591051 kubelet[2479]: E1031 02:43:56.589675 2479 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:57.575570 kubelet[2479]: I1031 02:43:57.575145 2479 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:57.588217 kubelet[2479]: W1031 02:43:57.587841 2479 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 31 02:43:58.439623 systemd[1]: Reloading requested from client PID 2751 ('systemctl') (unit session-11.scope)... Oct 31 02:43:58.439669 systemd[1]: Reloading... Oct 31 02:43:58.584203 zram_generator::config[2796]: No configuration found. Oct 31 02:43:58.784032 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 31 02:43:58.908620 systemd[1]: Reloading finished in 468 ms. Oct 31 02:43:58.963514 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:58.981000 systemd[1]: kubelet.service: Deactivated successfully. Oct 31 02:43:58.981881 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:58.990544 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 31 02:43:59.288422 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 31 02:43:59.304158 (kubelet)[2864]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 31 02:43:59.411076 kubelet[2864]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 31 02:43:59.411076 kubelet[2864]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 31 02:43:59.411076 kubelet[2864]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 31 02:43:59.411076 kubelet[2864]: I1031 02:43:59.410872 2864 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 31 02:43:59.426069 kubelet[2864]: I1031 02:43:59.425275 2864 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 31 02:43:59.426069 kubelet[2864]: I1031 02:43:59.425304 2864 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 31 02:43:59.426069 kubelet[2864]: I1031 02:43:59.425612 2864 server.go:954] "Client rotation is on, will bootstrap in background" Oct 31 02:43:59.432679 kubelet[2864]: I1031 02:43:59.432620 2864 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 31 02:43:59.441572 kubelet[2864]: I1031 02:43:59.440802 2864 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 31 02:43:59.459049 kubelet[2864]: E1031 02:43:59.458563 2864 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Oct 31 02:43:59.459049 kubelet[2864]: I1031 02:43:59.458606 2864 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Oct 31 02:43:59.471270 kubelet[2864]: I1031 02:43:59.470834 2864 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 31 02:43:59.471590 kubelet[2864]: I1031 02:43:59.471541 2864 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 31 02:43:59.472538 kubelet[2864]: I1031 02:43:59.471596 2864 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"srv-4lokn.gb1.brightbox.com","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Oct 31 02:43:59.472538 kubelet[2864]: I1031 02:43:59.471950 2864 topology_manager.go:138] "Creating topology manager with none policy" Oct 31 02:43:59.472538 kubelet[2864]: I1031 02:43:59.471969 2864 container_manager_linux.go:304] "Creating device plugin manager" Oct 31 02:43:59.473354 kubelet[2864]: I1031 02:43:59.472613 2864 state_mem.go:36] "Initialized new in-memory state store" Oct 31 02:43:59.473354 kubelet[2864]: I1031 02:43:59.472905 2864 kubelet.go:446] "Attempting to sync node with API server" Oct 31 02:43:59.474119 kubelet[2864]: I1031 02:43:59.474066 2864 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 31 02:43:59.474119 kubelet[2864]: I1031 02:43:59.474111 2864 kubelet.go:352] "Adding apiserver pod source" Oct 31 02:43:59.474250 kubelet[2864]: I1031 02:43:59.474128 2864 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 31 02:43:59.487094 kubelet[2864]: I1031 02:43:59.486384 2864 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Oct 31 02:43:59.488699 kubelet[2864]: I1031 02:43:59.487875 2864 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 31 02:43:59.490484 kubelet[2864]: I1031 02:43:59.489597 2864 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 31 02:43:59.491681 kubelet[2864]: I1031 02:43:59.490637 2864 server.go:1287] "Started kubelet" Oct 31 02:43:59.509331 kubelet[2864]: I1031 02:43:59.504481 2864 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 31 02:43:59.509331 kubelet[2864]: I1031 02:43:59.508669 2864 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 31 02:43:59.509331 kubelet[2864]: I1031 02:43:59.508763 2864 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 31 02:43:59.514206 kubelet[2864]: I1031 02:43:59.514141 2864 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 31 02:43:59.517832 kubelet[2864]: I1031 02:43:59.517801 2864 server.go:479] "Adding debug handlers to kubelet server" Oct 31 02:43:59.524079 kubelet[2864]: I1031 02:43:59.521936 2864 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 31 02:43:59.525532 kubelet[2864]: I1031 02:43:59.524678 2864 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 31 02:43:59.526568 kubelet[2864]: I1031 02:43:59.526544 2864 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 31 02:43:59.526931 kubelet[2864]: I1031 02:43:59.526911 2864 reconciler.go:26] "Reconciler: start to sync state" Oct 31 02:43:59.533504 kubelet[2864]: I1031 02:43:59.533472 2864 factory.go:221] Registration of the systemd container factory successfully Oct 31 02:43:59.534608 kubelet[2864]: I1031 02:43:59.534173 2864 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 31 02:43:59.541848 kubelet[2864]: E1031 02:43:59.539223 2864 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 31 02:43:59.542525 kubelet[2864]: I1031 02:43:59.542491 2864 factory.go:221] Registration of the containerd container factory successfully Oct 31 02:43:59.555333 kubelet[2864]: I1031 02:43:59.555275 2864 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 31 02:43:59.573820 kubelet[2864]: I1031 02:43:59.570949 2864 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 31 02:43:59.573820 kubelet[2864]: I1031 02:43:59.571017 2864 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 31 02:43:59.573820 kubelet[2864]: I1031 02:43:59.571079 2864 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 31 02:43:59.573820 kubelet[2864]: I1031 02:43:59.571096 2864 kubelet.go:2382] "Starting kubelet main sync loop" Oct 31 02:43:59.573820 kubelet[2864]: E1031 02:43:59.572737 2864 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 31 02:43:59.665554 kubelet[2864]: I1031 02:43:59.665181 2864 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 31 02:43:59.665554 kubelet[2864]: I1031 02:43:59.665214 2864 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 31 02:43:59.665554 kubelet[2864]: I1031 02:43:59.665252 2864 state_mem.go:36] "Initialized new in-memory state store" Oct 31 02:43:59.665554 kubelet[2864]: I1031 02:43:59.665529 2864 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 31 02:43:59.665554 kubelet[2864]: I1031 02:43:59.665547 2864 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 31 02:43:59.665554 kubelet[2864]: I1031 02:43:59.665592 2864 policy_none.go:49] "None policy: Start" Oct 31 02:43:59.666323 kubelet[2864]: I1031 02:43:59.665620 2864 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 31 02:43:59.666323 kubelet[2864]: I1031 02:43:59.665650 2864 state_mem.go:35] "Initializing new in-memory state store" Oct 31 02:43:59.666323 kubelet[2864]: I1031 02:43:59.665845 2864 state_mem.go:75] "Updated machine memory state" Oct 31 02:43:59.668628 kubelet[2864]: I1031 02:43:59.668587 2864 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 31 02:43:59.670279 kubelet[2864]: I1031 02:43:59.669408 2864 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 31 02:43:59.670279 kubelet[2864]: I1031 02:43:59.669433 2864 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 31 02:43:59.676161 kubelet[2864]: I1031 02:43:59.674529 2864 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 31 02:43:59.676161 kubelet[2864]: I1031 02:43:59.675414 2864 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.676161 kubelet[2864]: I1031 02:43:59.675945 2864 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.677733 kubelet[2864]: I1031 02:43:59.677700 2864 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.686984 kubelet[2864]: E1031 02:43:59.686952 2864 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 31 02:43:59.702762 kubelet[2864]: W1031 02:43:59.702636 2864 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 31 02:43:59.708472 kubelet[2864]: W1031 02:43:59.708085 2864 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 31 02:43:59.708472 kubelet[2864]: W1031 02:43:59.708346 2864 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 31 02:43:59.708472 kubelet[2864]: E1031 02:43:59.708397 2864 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" already exists" pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.813738 kubelet[2864]: I1031 02:43:59.813421 2864 kubelet_node_status.go:75] "Attempting to register node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.828597 kubelet[2864]: I1031 02:43:59.827519 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3e2368e63ed376cc7b2900b2238f34c7-k8s-certs\") pod \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" (UID: \"3e2368e63ed376cc7b2900b2238f34c7\") " pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.828597 kubelet[2864]: I1031 02:43:59.827612 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3e2368e63ed376cc7b2900b2238f34c7-usr-share-ca-certificates\") pod \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" (UID: \"3e2368e63ed376cc7b2900b2238f34c7\") " pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.828597 kubelet[2864]: I1031 02:43:59.827650 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-ca-certs\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.828597 kubelet[2864]: I1031 02:43:59.827680 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-flexvolume-dir\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.828597 kubelet[2864]: I1031 02:43:59.827712 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-usr-share-ca-certificates\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.829047 kubelet[2864]: I1031 02:43:59.827741 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3ae349f67b9e50e769a8f49fcbc622f1-kubeconfig\") pod \"kube-scheduler-srv-4lokn.gb1.brightbox.com\" (UID: \"3ae349f67b9e50e769a8f49fcbc622f1\") " pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.829047 kubelet[2864]: I1031 02:43:59.827766 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3e2368e63ed376cc7b2900b2238f34c7-ca-certs\") pod \"kube-apiserver-srv-4lokn.gb1.brightbox.com\" (UID: \"3e2368e63ed376cc7b2900b2238f34c7\") " pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.829047 kubelet[2864]: I1031 02:43:59.827810 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-k8s-certs\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.829047 kubelet[2864]: I1031 02:43:59.827844 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/459c05ce4d751e27798a6b5386a0f2aa-kubeconfig\") pod \"kube-controller-manager-srv-4lokn.gb1.brightbox.com\" (UID: \"459c05ce4d751e27798a6b5386a0f2aa\") " pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.830042 kubelet[2864]: I1031 02:43:59.829985 2864 kubelet_node_status.go:124] "Node was previously registered" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:43:59.830177 kubelet[2864]: I1031 02:43:59.830143 2864 kubelet_node_status.go:78] "Successfully registered node" node="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:00.489627 kubelet[2864]: I1031 02:44:00.489579 2864 apiserver.go:52] "Watching apiserver" Oct 31 02:44:00.527014 kubelet[2864]: I1031 02:44:00.526914 2864 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 31 02:44:00.626622 kubelet[2864]: I1031 02:44:00.626584 2864 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:44:00.640508 kubelet[2864]: W1031 02:44:00.640464 2864 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 31 02:44:00.640685 kubelet[2864]: E1031 02:44:00.640568 2864 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-srv-4lokn.gb1.brightbox.com\" already exists" pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" Oct 31 02:44:00.680180 kubelet[2864]: I1031 02:44:00.680059 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-srv-4lokn.gb1.brightbox.com" podStartSLOduration=1.679976027 podStartE2EDuration="1.679976027s" podCreationTimestamp="2025-10-31 02:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-31 02:44:00.679682622 +0000 UTC m=+1.348171305" watchObservedRunningTime="2025-10-31 02:44:00.679976027 +0000 UTC m=+1.348464691" Oct 31 02:44:00.680613 kubelet[2864]: I1031 02:44:00.680292 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-srv-4lokn.gb1.brightbox.com" podStartSLOduration=1.680281845 podStartE2EDuration="1.680281845s" podCreationTimestamp="2025-10-31 02:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-31 02:44:00.65964703 +0000 UTC m=+1.328135704" watchObservedRunningTime="2025-10-31 02:44:00.680281845 +0000 UTC m=+1.348770508" Oct 31 02:44:00.701340 kubelet[2864]: I1031 02:44:00.701264 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-srv-4lokn.gb1.brightbox.com" podStartSLOduration=3.701241981 podStartE2EDuration="3.701241981s" podCreationTimestamp="2025-10-31 02:43:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-31 02:44:00.696437486 +0000 UTC m=+1.364926183" watchObservedRunningTime="2025-10-31 02:44:00.701241981 +0000 UTC m=+1.369730649" Oct 31 02:44:05.119869 kubelet[2864]: I1031 02:44:05.119495 2864 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 31 02:44:05.122486 containerd[1609]: time="2025-10-31T02:44:05.121241535Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 31 02:44:05.123088 kubelet[2864]: I1031 02:44:05.121631 2864 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 31 02:44:05.874622 kubelet[2864]: I1031 02:44:05.874543 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c11af102-f2e3-4ac9-bc13-2a8b220e2a9d-lib-modules\") pod \"kube-proxy-t56vv\" (UID: \"c11af102-f2e3-4ac9-bc13-2a8b220e2a9d\") " pod="kube-system/kube-proxy-t56vv" Oct 31 02:44:05.874622 kubelet[2864]: I1031 02:44:05.874623 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/c11af102-f2e3-4ac9-bc13-2a8b220e2a9d-kube-proxy\") pod \"kube-proxy-t56vv\" (UID: \"c11af102-f2e3-4ac9-bc13-2a8b220e2a9d\") " pod="kube-system/kube-proxy-t56vv" Oct 31 02:44:05.875108 kubelet[2864]: I1031 02:44:05.874664 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c11af102-f2e3-4ac9-bc13-2a8b220e2a9d-xtables-lock\") pod \"kube-proxy-t56vv\" (UID: \"c11af102-f2e3-4ac9-bc13-2a8b220e2a9d\") " pod="kube-system/kube-proxy-t56vv" Oct 31 02:44:05.875108 kubelet[2864]: I1031 02:44:05.874694 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts5mh\" (UniqueName: \"kubernetes.io/projected/c11af102-f2e3-4ac9-bc13-2a8b220e2a9d-kube-api-access-ts5mh\") pod \"kube-proxy-t56vv\" (UID: \"c11af102-f2e3-4ac9-bc13-2a8b220e2a9d\") " pod="kube-system/kube-proxy-t56vv" Oct 31 02:44:06.085147 containerd[1609]: time="2025-10-31T02:44:06.084793839Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-t56vv,Uid:c11af102-f2e3-4ac9-bc13-2a8b220e2a9d,Namespace:kube-system,Attempt:0,}" Oct 31 02:44:06.192166 containerd[1609]: time="2025-10-31T02:44:06.189825545Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:06.196000 containerd[1609]: time="2025-10-31T02:44:06.189976939Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:06.196000 containerd[1609]: time="2025-10-31T02:44:06.193826866Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:06.196000 containerd[1609]: time="2025-10-31T02:44:06.194454516Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:06.280468 kubelet[2864]: I1031 02:44:06.280417 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/c27e40a6-3211-45d5-8883-074f0d40906a-var-lib-calico\") pod \"tigera-operator-7dcd859c48-f68js\" (UID: \"c27e40a6-3211-45d5-8883-074f0d40906a\") " pod="tigera-operator/tigera-operator-7dcd859c48-f68js" Oct 31 02:44:06.281809 kubelet[2864]: I1031 02:44:06.280483 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2smc\" (UniqueName: \"kubernetes.io/projected/c27e40a6-3211-45d5-8883-074f0d40906a-kube-api-access-b2smc\") pod \"tigera-operator-7dcd859c48-f68js\" (UID: \"c27e40a6-3211-45d5-8883-074f0d40906a\") " pod="tigera-operator/tigera-operator-7dcd859c48-f68js" Oct 31 02:44:06.321569 containerd[1609]: time="2025-10-31T02:44:06.321505877Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-t56vv,Uid:c11af102-f2e3-4ac9-bc13-2a8b220e2a9d,Namespace:kube-system,Attempt:0,} returns sandbox id \"d9308e4ce77c73bbd6d1f099dc9b525ba963bf7627e33fe545c9e250a4e226d5\"" Oct 31 02:44:06.327632 containerd[1609]: time="2025-10-31T02:44:06.327334881Z" level=info msg="CreateContainer within sandbox \"d9308e4ce77c73bbd6d1f099dc9b525ba963bf7627e33fe545c9e250a4e226d5\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 31 02:44:06.365582 containerd[1609]: time="2025-10-31T02:44:06.365515022Z" level=info msg="CreateContainer within sandbox \"d9308e4ce77c73bbd6d1f099dc9b525ba963bf7627e33fe545c9e250a4e226d5\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"868830e91e460092030fa7fed87433eba863d16664f401858b2b4bc75b0bd60c\"" Oct 31 02:44:06.367113 containerd[1609]: time="2025-10-31T02:44:06.367082542Z" level=info msg="StartContainer for \"868830e91e460092030fa7fed87433eba863d16664f401858b2b4bc75b0bd60c\"" Oct 31 02:44:06.460115 containerd[1609]: time="2025-10-31T02:44:06.458063836Z" level=info msg="StartContainer for \"868830e91e460092030fa7fed87433eba863d16664f401858b2b4bc75b0bd60c\" returns successfully" Oct 31 02:44:06.575947 containerd[1609]: time="2025-10-31T02:44:06.574866830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-f68js,Uid:c27e40a6-3211-45d5-8883-074f0d40906a,Namespace:tigera-operator,Attempt:0,}" Oct 31 02:44:06.658698 containerd[1609]: time="2025-10-31T02:44:06.654203927Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:06.658698 containerd[1609]: time="2025-10-31T02:44:06.658408903Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:06.658698 containerd[1609]: time="2025-10-31T02:44:06.658497511Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:06.658998 containerd[1609]: time="2025-10-31T02:44:06.658836732Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:06.689390 kubelet[2864]: I1031 02:44:06.687514 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-t56vv" podStartSLOduration=1.687480402 podStartE2EDuration="1.687480402s" podCreationTimestamp="2025-10-31 02:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-31 02:44:06.670139426 +0000 UTC m=+7.338628134" watchObservedRunningTime="2025-10-31 02:44:06.687480402 +0000 UTC m=+7.355969070" Oct 31 02:44:06.789400 containerd[1609]: time="2025-10-31T02:44:06.789342585Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-f68js,Uid:c27e40a6-3211-45d5-8883-074f0d40906a,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"782d4a5b077b1003eaef53622e0acc0dcf9157154b6b330f4e0f037a9be10c5e\"" Oct 31 02:44:06.793348 containerd[1609]: time="2025-10-31T02:44:06.793283706Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Oct 31 02:44:07.002684 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1962140550.mount: Deactivated successfully. Oct 31 02:44:09.402513 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount690901110.mount: Deactivated successfully. Oct 31 02:44:10.526985 containerd[1609]: time="2025-10-31T02:44:10.526836532Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:10.529798 containerd[1609]: time="2025-10-31T02:44:10.529531477Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Oct 31 02:44:10.531562 containerd[1609]: time="2025-10-31T02:44:10.531523090Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:10.535693 containerd[1609]: time="2025-10-31T02:44:10.535654871Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:10.536903 containerd[1609]: time="2025-10-31T02:44:10.536745331Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 3.743405067s" Oct 31 02:44:10.536991 containerd[1609]: time="2025-10-31T02:44:10.536899902Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Oct 31 02:44:10.542757 containerd[1609]: time="2025-10-31T02:44:10.542703177Z" level=info msg="CreateContainer within sandbox \"782d4a5b077b1003eaef53622e0acc0dcf9157154b6b330f4e0f037a9be10c5e\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Oct 31 02:44:10.559203 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1094048907.mount: Deactivated successfully. Oct 31 02:44:10.562550 containerd[1609]: time="2025-10-31T02:44:10.562502724Z" level=info msg="CreateContainer within sandbox \"782d4a5b077b1003eaef53622e0acc0dcf9157154b6b330f4e0f037a9be10c5e\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"4491d1f24371393fd13b2cdc697c80bb05c6c42b44d560061a9a270c09aef5ed\"" Oct 31 02:44:10.563628 containerd[1609]: time="2025-10-31T02:44:10.563550069Z" level=info msg="StartContainer for \"4491d1f24371393fd13b2cdc697c80bb05c6c42b44d560061a9a270c09aef5ed\"" Oct 31 02:44:10.652045 containerd[1609]: time="2025-10-31T02:44:10.651863072Z" level=info msg="StartContainer for \"4491d1f24371393fd13b2cdc697c80bb05c6c42b44d560061a9a270c09aef5ed\" returns successfully" Oct 31 02:44:11.350086 kubelet[2864]: I1031 02:44:11.349890 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-f68js" podStartSLOduration=1.603432907 podStartE2EDuration="5.349779978s" podCreationTimestamp="2025-10-31 02:44:06 +0000 UTC" firstStartedPulling="2025-10-31 02:44:06.792168853 +0000 UTC m=+7.460657520" lastFinishedPulling="2025-10-31 02:44:10.538515922 +0000 UTC m=+11.207004591" observedRunningTime="2025-10-31 02:44:10.679920566 +0000 UTC m=+11.348409244" watchObservedRunningTime="2025-10-31 02:44:11.349779978 +0000 UTC m=+12.018268653" Oct 31 02:44:18.385719 sudo[1912]: pam_unix(sudo:session): session closed for user root Oct 31 02:44:18.542122 sshd[1908]: pam_unix(sshd:session): session closed for user core Oct 31 02:44:18.552169 systemd-logind[1591]: Session 11 logged out. Waiting for processes to exit. Oct 31 02:44:18.555290 systemd[1]: sshd@8-10.230.25.230:22-147.75.109.163:41778.service: Deactivated successfully. Oct 31 02:44:18.563598 systemd[1]: session-11.scope: Deactivated successfully. Oct 31 02:44:18.567004 systemd-logind[1591]: Removed session 11. Oct 31 02:44:25.637294 kubelet[2864]: I1031 02:44:25.636962 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/041a807a-aa5a-433d-a490-0c2e51864c3c-typha-certs\") pod \"calico-typha-58c6bb7b96-94hs8\" (UID: \"041a807a-aa5a-433d-a490-0c2e51864c3c\") " pod="calico-system/calico-typha-58c6bb7b96-94hs8" Oct 31 02:44:25.637294 kubelet[2864]: I1031 02:44:25.637112 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc7w4\" (UniqueName: \"kubernetes.io/projected/041a807a-aa5a-433d-a490-0c2e51864c3c-kube-api-access-nc7w4\") pod \"calico-typha-58c6bb7b96-94hs8\" (UID: \"041a807a-aa5a-433d-a490-0c2e51864c3c\") " pod="calico-system/calico-typha-58c6bb7b96-94hs8" Oct 31 02:44:25.637294 kubelet[2864]: I1031 02:44:25.637171 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/041a807a-aa5a-433d-a490-0c2e51864c3c-tigera-ca-bundle\") pod \"calico-typha-58c6bb7b96-94hs8\" (UID: \"041a807a-aa5a-433d-a490-0c2e51864c3c\") " pod="calico-system/calico-typha-58c6bb7b96-94hs8" Oct 31 02:44:25.840095 kubelet[2864]: I1031 02:44:25.839096 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-cni-net-dir\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.840095 kubelet[2864]: I1031 02:44:25.839215 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-var-lib-calico\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.840095 kubelet[2864]: I1031 02:44:25.839246 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-xtables-lock\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.840095 kubelet[2864]: I1031 02:44:25.839293 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cl4w\" (UniqueName: \"kubernetes.io/projected/86f173e2-0cc6-42d9-b460-1cd9d98ae136-kube-api-access-9cl4w\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.840095 kubelet[2864]: I1031 02:44:25.839341 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-var-run-calico\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844510 kubelet[2864]: I1031 02:44:25.839376 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-cni-bin-dir\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844510 kubelet[2864]: I1031 02:44:25.839408 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-cni-log-dir\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844510 kubelet[2864]: I1031 02:44:25.839436 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/86f173e2-0cc6-42d9-b460-1cd9d98ae136-node-certs\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844510 kubelet[2864]: I1031 02:44:25.839461 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86f173e2-0cc6-42d9-b460-1cd9d98ae136-tigera-ca-bundle\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844510 kubelet[2864]: I1031 02:44:25.839498 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-flexvol-driver-host\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844879 kubelet[2864]: I1031 02:44:25.839534 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-policysync\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.844879 kubelet[2864]: I1031 02:44:25.839571 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86f173e2-0cc6-42d9-b460-1cd9d98ae136-lib-modules\") pod \"calico-node-qfjs8\" (UID: \"86f173e2-0cc6-42d9-b460-1cd9d98ae136\") " pod="calico-system/calico-node-qfjs8" Oct 31 02:44:25.906753 containerd[1609]: time="2025-10-31T02:44:25.906156677Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-58c6bb7b96-94hs8,Uid:041a807a-aa5a-433d-a490-0c2e51864c3c,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:25.981439 kubelet[2864]: E1031 02:44:25.978099 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:25.981439 kubelet[2864]: W1031 02:44:25.978196 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:25.981439 kubelet[2864]: E1031 02:44:25.978939 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:25.992863 kubelet[2864]: E1031 02:44:25.990358 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:25.992863 kubelet[2864]: W1031 02:44:25.990384 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:25.992863 kubelet[2864]: E1031 02:44:25.990407 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:25.997540 kubelet[2864]: E1031 02:44:25.997320 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:25.997540 kubelet[2864]: W1031 02:44:25.997359 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:25.997540 kubelet[2864]: E1031 02:44:25.997399 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.041921 containerd[1609]: time="2025-10-31T02:44:26.039971543Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:26.042422 containerd[1609]: time="2025-10-31T02:44:26.042179400Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:26.042422 containerd[1609]: time="2025-10-31T02:44:26.042273106Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:26.043276 containerd[1609]: time="2025-10-31T02:44:26.043071824Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:26.096160 kubelet[2864]: E1031 02:44:26.094998 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:26.102908 kubelet[2864]: E1031 02:44:26.102698 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.102908 kubelet[2864]: W1031 02:44:26.102727 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.102908 kubelet[2864]: E1031 02:44:26.102753 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.104056 kubelet[2864]: E1031 02:44:26.103674 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.104056 kubelet[2864]: W1031 02:44:26.103693 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.104056 kubelet[2864]: E1031 02:44:26.103729 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.104511 kubelet[2864]: E1031 02:44:26.104298 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.104511 kubelet[2864]: W1031 02:44:26.104329 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.104511 kubelet[2864]: E1031 02:44:26.104367 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.105142 kubelet[2864]: E1031 02:44:26.104967 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.105142 kubelet[2864]: W1031 02:44:26.105003 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.105142 kubelet[2864]: E1031 02:44:26.105066 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.105771 kubelet[2864]: E1031 02:44:26.105603 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.105771 kubelet[2864]: W1031 02:44:26.105621 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.105771 kubelet[2864]: E1031 02:44:26.105657 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.106363 kubelet[2864]: E1031 02:44:26.106161 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.106363 kubelet[2864]: W1031 02:44:26.106179 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.106363 kubelet[2864]: E1031 02:44:26.106216 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.107287 kubelet[2864]: E1031 02:44:26.106946 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.107287 kubelet[2864]: W1031 02:44:26.106964 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.107287 kubelet[2864]: E1031 02:44:26.106999 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.107821 kubelet[2864]: E1031 02:44:26.107613 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.107821 kubelet[2864]: W1031 02:44:26.107631 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.107821 kubelet[2864]: E1031 02:44:26.107666 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.108413 kubelet[2864]: E1031 02:44:26.108215 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.108413 kubelet[2864]: W1031 02:44:26.108233 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.108413 kubelet[2864]: E1031 02:44:26.108269 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.108892 kubelet[2864]: E1031 02:44:26.108771 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.108892 kubelet[2864]: W1031 02:44:26.108789 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.108892 kubelet[2864]: E1031 02:44:26.108827 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.109508 kubelet[2864]: E1031 02:44:26.109332 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.109508 kubelet[2864]: W1031 02:44:26.109351 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.109508 kubelet[2864]: E1031 02:44:26.109387 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.110101 kubelet[2864]: E1031 02:44:26.109916 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.110101 kubelet[2864]: W1031 02:44:26.109943 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.110101 kubelet[2864]: E1031 02:44:26.109961 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.110569 kubelet[2864]: E1031 02:44:26.110470 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.110569 kubelet[2864]: W1031 02:44:26.110488 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.110569 kubelet[2864]: E1031 02:44:26.110503 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.112533 kubelet[2864]: E1031 02:44:26.112358 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.112533 kubelet[2864]: W1031 02:44:26.112404 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.112533 kubelet[2864]: E1031 02:44:26.112422 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.113229 kubelet[2864]: E1031 02:44:26.112945 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.113229 kubelet[2864]: W1031 02:44:26.112964 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.113229 kubelet[2864]: E1031 02:44:26.112991 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.113750 kubelet[2864]: E1031 02:44:26.113365 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.113750 kubelet[2864]: W1031 02:44:26.113571 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.113750 kubelet[2864]: E1031 02:44:26.113591 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.114444 kubelet[2864]: E1031 02:44:26.114179 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.114444 kubelet[2864]: W1031 02:44:26.114197 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.114444 kubelet[2864]: E1031 02:44:26.114213 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.115298 kubelet[2864]: E1031 02:44:26.114970 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.115298 kubelet[2864]: W1031 02:44:26.114987 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.115298 kubelet[2864]: E1031 02:44:26.115003 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.116505 kubelet[2864]: E1031 02:44:26.116338 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.116685 kubelet[2864]: W1031 02:44:26.116352 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.116685 kubelet[2864]: E1031 02:44:26.116574 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.117538 kubelet[2864]: E1031 02:44:26.117309 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.117538 kubelet[2864]: W1031 02:44:26.117328 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.117538 kubelet[2864]: E1031 02:44:26.117344 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.122925 containerd[1609]: time="2025-10-31T02:44:26.122714115Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-qfjs8,Uid:86f173e2-0cc6-42d9-b460-1cd9d98ae136,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:26.149498 kubelet[2864]: E1031 02:44:26.149344 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.151479 kubelet[2864]: W1031 02:44:26.149714 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.151479 kubelet[2864]: E1031 02:44:26.149747 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.151479 kubelet[2864]: I1031 02:44:26.150533 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/7293ccde-9a5c-4930-845b-a0c0b00c9ac3-varrun\") pod \"csi-node-driver-d2znc\" (UID: \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\") " pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:26.153945 kubelet[2864]: E1031 02:44:26.153482 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.153945 kubelet[2864]: W1031 02:44:26.153510 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.153945 kubelet[2864]: E1031 02:44:26.153549 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.157485 kubelet[2864]: E1031 02:44:26.156590 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.157485 kubelet[2864]: W1031 02:44:26.156615 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.159930 kubelet[2864]: E1031 02:44:26.159430 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.160347 kubelet[2864]: E1031 02:44:26.160326 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.161633 kubelet[2864]: W1031 02:44:26.160985 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.161633 kubelet[2864]: E1031 02:44:26.161016 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.162313 kubelet[2864]: I1031 02:44:26.161952 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7293ccde-9a5c-4930-845b-a0c0b00c9ac3-kubelet-dir\") pod \"csi-node-driver-d2znc\" (UID: \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\") " pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:26.164590 kubelet[2864]: E1031 02:44:26.164255 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.164590 kubelet[2864]: W1031 02:44:26.164278 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.166292 kubelet[2864]: E1031 02:44:26.165102 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.166738 kubelet[2864]: I1031 02:44:26.166653 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8nr7\" (UniqueName: \"kubernetes.io/projected/7293ccde-9a5c-4930-845b-a0c0b00c9ac3-kube-api-access-g8nr7\") pod \"csi-node-driver-d2znc\" (UID: \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\") " pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:26.167238 kubelet[2864]: E1031 02:44:26.167094 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.167238 kubelet[2864]: W1031 02:44:26.167119 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.167918 kubelet[2864]: E1031 02:44:26.167539 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.169569 kubelet[2864]: E1031 02:44:26.168422 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.169569 kubelet[2864]: W1031 02:44:26.168442 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.169569 kubelet[2864]: E1031 02:44:26.169366 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.171007 kubelet[2864]: E1031 02:44:26.170894 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.171189 kubelet[2864]: W1031 02:44:26.171164 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.171479 kubelet[2864]: E1031 02:44:26.171420 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.172287 kubelet[2864]: I1031 02:44:26.172242 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7293ccde-9a5c-4930-845b-a0c0b00c9ac3-registration-dir\") pod \"csi-node-driver-d2znc\" (UID: \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\") " pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:26.173242 kubelet[2864]: E1031 02:44:26.173220 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.173737 kubelet[2864]: W1031 02:44:26.173394 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.173737 kubelet[2864]: E1031 02:44:26.173446 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.175232 kubelet[2864]: E1031 02:44:26.174819 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.175232 kubelet[2864]: W1031 02:44:26.174851 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.175700 kubelet[2864]: E1031 02:44:26.175368 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.176923 kubelet[2864]: E1031 02:44:26.176432 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.176923 kubelet[2864]: W1031 02:44:26.176492 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.177693 kubelet[2864]: E1031 02:44:26.177316 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.179955 kubelet[2864]: I1031 02:44:26.177354 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7293ccde-9a5c-4930-845b-a0c0b00c9ac3-socket-dir\") pod \"csi-node-driver-d2znc\" (UID: \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\") " pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:26.179955 kubelet[2864]: E1031 02:44:26.179392 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.179955 kubelet[2864]: W1031 02:44:26.179423 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.179955 kubelet[2864]: E1031 02:44:26.179485 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.180527 kubelet[2864]: E1031 02:44:26.180454 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.180527 kubelet[2864]: W1031 02:44:26.180473 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.180804 kubelet[2864]: E1031 02:44:26.180780 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.181587 kubelet[2864]: E1031 02:44:26.181563 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.182297 kubelet[2864]: W1031 02:44:26.182063 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.182297 kubelet[2864]: E1031 02:44:26.182091 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.184002 kubelet[2864]: E1031 02:44:26.183153 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.184002 kubelet[2864]: W1031 02:44:26.183173 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.184002 kubelet[2864]: E1031 02:44:26.183189 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.215281 containerd[1609]: time="2025-10-31T02:44:26.215121771Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:26.218916 containerd[1609]: time="2025-10-31T02:44:26.218212152Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:26.218916 containerd[1609]: time="2025-10-31T02:44:26.218280603Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:26.219997 containerd[1609]: time="2025-10-31T02:44:26.219715478Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:26.279807 kubelet[2864]: E1031 02:44:26.279613 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.279807 kubelet[2864]: W1031 02:44:26.279642 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.280408 kubelet[2864]: E1031 02:44:26.279880 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.282099 kubelet[2864]: E1031 02:44:26.281095 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.282099 kubelet[2864]: W1031 02:44:26.281114 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.282099 kubelet[2864]: E1031 02:44:26.281137 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.282897 kubelet[2864]: E1031 02:44:26.282694 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.282897 kubelet[2864]: W1031 02:44:26.282811 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.283396 kubelet[2864]: E1031 02:44:26.282870 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.284163 kubelet[2864]: E1031 02:44:26.284016 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.284163 kubelet[2864]: W1031 02:44:26.284121 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.284968 kubelet[2864]: E1031 02:44:26.284365 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.286600 kubelet[2864]: E1031 02:44:26.286183 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.286600 kubelet[2864]: W1031 02:44:26.286202 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.287573 kubelet[2864]: E1031 02:44:26.286963 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.287961 kubelet[2864]: E1031 02:44:26.287932 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.288233 kubelet[2864]: W1031 02:44:26.288141 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.289063 kubelet[2864]: E1031 02:44:26.288757 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.291084 kubelet[2864]: E1031 02:44:26.290444 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.291084 kubelet[2864]: W1031 02:44:26.290506 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.291084 kubelet[2864]: E1031 02:44:26.290668 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.291900 kubelet[2864]: E1031 02:44:26.291643 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.291900 kubelet[2864]: W1031 02:44:26.291663 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.292696 kubelet[2864]: E1031 02:44:26.292152 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.293728 kubelet[2864]: E1031 02:44:26.293417 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.293728 kubelet[2864]: W1031 02:44:26.293438 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.294344 kubelet[2864]: E1031 02:44:26.294058 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.295265 kubelet[2864]: E1031 02:44:26.295134 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.295707 kubelet[2864]: W1031 02:44:26.295349 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.296504 kubelet[2864]: E1031 02:44:26.296202 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.297057 kubelet[2864]: E1031 02:44:26.296974 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.297057 kubelet[2864]: W1031 02:44:26.296990 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.297567 kubelet[2864]: E1031 02:44:26.297254 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.298556 kubelet[2864]: E1031 02:44:26.298340 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.298556 kubelet[2864]: W1031 02:44:26.298358 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.299139 kubelet[2864]: E1031 02:44:26.298829 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.299625 kubelet[2864]: E1031 02:44:26.299370 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.299625 kubelet[2864]: W1031 02:44:26.299388 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.300486 kubelet[2864]: E1031 02:44:26.300194 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.300898 kubelet[2864]: E1031 02:44:26.300637 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.300898 kubelet[2864]: W1031 02:44:26.300654 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.300898 kubelet[2864]: E1031 02:44:26.300866 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.301775 kubelet[2864]: E1031 02:44:26.301636 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.301775 kubelet[2864]: W1031 02:44:26.301654 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.302132 kubelet[2864]: E1031 02:44:26.301956 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.304358 kubelet[2864]: E1031 02:44:26.303312 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.304358 kubelet[2864]: W1031 02:44:26.303340 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.305365 kubelet[2864]: E1031 02:44:26.304687 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.306765 kubelet[2864]: E1031 02:44:26.305992 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.306765 kubelet[2864]: W1031 02:44:26.306012 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.308334 kubelet[2864]: E1031 02:44:26.307243 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.309667 kubelet[2864]: E1031 02:44:26.309646 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.310054 kubelet[2864]: W1031 02:44:26.309968 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.311552 kubelet[2864]: E1031 02:44:26.310984 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.313785 kubelet[2864]: E1031 02:44:26.313636 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.313785 kubelet[2864]: W1031 02:44:26.313656 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.314881 kubelet[2864]: E1031 02:44:26.314728 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.316346 kubelet[2864]: E1031 02:44:26.316292 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.316640 kubelet[2864]: W1031 02:44:26.316455 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.317532 kubelet[2864]: E1031 02:44:26.317300 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.320468 kubelet[2864]: E1031 02:44:26.319978 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.320468 kubelet[2864]: W1031 02:44:26.319999 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.320468 kubelet[2864]: E1031 02:44:26.320110 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.323852 kubelet[2864]: E1031 02:44:26.323285 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.323852 kubelet[2864]: W1031 02:44:26.323307 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.326826 kubelet[2864]: E1031 02:44:26.326203 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.326826 kubelet[2864]: W1031 02:44:26.326225 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.328312 kubelet[2864]: E1031 02:44:26.327818 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.328312 kubelet[2864]: W1031 02:44:26.327849 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.328312 kubelet[2864]: E1031 02:44:26.327870 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.328312 kubelet[2864]: E1031 02:44:26.327900 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.328312 kubelet[2864]: E1031 02:44:26.327919 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.330871 kubelet[2864]: E1031 02:44:26.330756 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.332401 kubelet[2864]: W1031 02:44:26.332057 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.332882 kubelet[2864]: E1031 02:44:26.332588 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.347046 kubelet[2864]: E1031 02:44:26.345827 2864 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 31 02:44:26.347046 kubelet[2864]: W1031 02:44:26.346079 2864 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 31 02:44:26.347046 kubelet[2864]: E1031 02:44:26.346115 2864 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 31 02:44:26.392380 containerd[1609]: time="2025-10-31T02:44:26.392182964Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-qfjs8,Uid:86f173e2-0cc6-42d9-b460-1cd9d98ae136,Namespace:calico-system,Attempt:0,} returns sandbox id \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\"" Oct 31 02:44:26.396043 containerd[1609]: time="2025-10-31T02:44:26.395991848Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Oct 31 02:44:26.445377 containerd[1609]: time="2025-10-31T02:44:26.444746590Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-58c6bb7b96-94hs8,Uid:041a807a-aa5a-433d-a490-0c2e51864c3c,Namespace:calico-system,Attempt:0,} returns sandbox id \"d2de8b9f90a752262b0a20ab5c13b2d3f949a886ceef220fca5f957a104e5280\"" Oct 31 02:44:27.575641 kubelet[2864]: E1031 02:44:27.575551 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:28.121646 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1788938088.mount: Deactivated successfully. Oct 31 02:44:28.392421 containerd[1609]: time="2025-10-31T02:44:28.392220315Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:28.393438 containerd[1609]: time="2025-10-31T02:44:28.393401055Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=5941492" Oct 31 02:44:28.394150 containerd[1609]: time="2025-10-31T02:44:28.394095781Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:28.397530 containerd[1609]: time="2025-10-31T02:44:28.397489520Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:28.398813 containerd[1609]: time="2025-10-31T02:44:28.398650568Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 2.00258583s" Oct 31 02:44:28.398813 containerd[1609]: time="2025-10-31T02:44:28.398701669Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Oct 31 02:44:28.401309 containerd[1609]: time="2025-10-31T02:44:28.401271133Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Oct 31 02:44:28.405294 containerd[1609]: time="2025-10-31T02:44:28.404989886Z" level=info msg="CreateContainer within sandbox \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Oct 31 02:44:28.430676 containerd[1609]: time="2025-10-31T02:44:28.430633966Z" level=info msg="CreateContainer within sandbox \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"24a73a643bd7a0e4ddb641ad2f62df65202ad81ae67137a12a6c75c7a3e606b5\"" Oct 31 02:44:28.434841 containerd[1609]: time="2025-10-31T02:44:28.432936305Z" level=info msg="StartContainer for \"24a73a643bd7a0e4ddb641ad2f62df65202ad81ae67137a12a6c75c7a3e606b5\"" Oct 31 02:44:28.556319 containerd[1609]: time="2025-10-31T02:44:28.555971571Z" level=info msg="StartContainer for \"24a73a643bd7a0e4ddb641ad2f62df65202ad81ae67137a12a6c75c7a3e606b5\" returns successfully" Oct 31 02:44:28.645704 containerd[1609]: time="2025-10-31T02:44:28.628135927Z" level=info msg="shim disconnected" id=24a73a643bd7a0e4ddb641ad2f62df65202ad81ae67137a12a6c75c7a3e606b5 namespace=k8s.io Oct 31 02:44:28.646179 containerd[1609]: time="2025-10-31T02:44:28.645949961Z" level=warning msg="cleaning up after shim disconnected" id=24a73a643bd7a0e4ddb641ad2f62df65202ad81ae67137a12a6c75c7a3e606b5 namespace=k8s.io Oct 31 02:44:28.646333 containerd[1609]: time="2025-10-31T02:44:28.646305298Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 31 02:44:29.042110 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-24a73a643bd7a0e4ddb641ad2f62df65202ad81ae67137a12a6c75c7a3e606b5-rootfs.mount: Deactivated successfully. Oct 31 02:44:29.572433 kubelet[2864]: E1031 02:44:29.571724 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:31.574046 kubelet[2864]: E1031 02:44:31.572845 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:32.244559 containerd[1609]: time="2025-10-31T02:44:32.244404357Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:32.246877 containerd[1609]: time="2025-10-31T02:44:32.246767243Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33739890" Oct 31 02:44:32.247676 containerd[1609]: time="2025-10-31T02:44:32.247133207Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:32.251790 containerd[1609]: time="2025-10-31T02:44:32.251708138Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:32.253273 containerd[1609]: time="2025-10-31T02:44:32.252989399Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 3.851657039s" Oct 31 02:44:32.253273 containerd[1609]: time="2025-10-31T02:44:32.253070316Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Oct 31 02:44:32.256817 containerd[1609]: time="2025-10-31T02:44:32.256773199Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Oct 31 02:44:32.284540 containerd[1609]: time="2025-10-31T02:44:32.284204318Z" level=info msg="CreateContainer within sandbox \"d2de8b9f90a752262b0a20ab5c13b2d3f949a886ceef220fca5f957a104e5280\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Oct 31 02:44:32.442376 containerd[1609]: time="2025-10-31T02:44:32.442297397Z" level=info msg="CreateContainer within sandbox \"d2de8b9f90a752262b0a20ab5c13b2d3f949a886ceef220fca5f957a104e5280\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"b8c67ae3abaf58d39371ec247903acc730b7576f32db4f04b47b0bc45378496d\"" Oct 31 02:44:32.444157 containerd[1609]: time="2025-10-31T02:44:32.443624017Z" level=info msg="StartContainer for \"b8c67ae3abaf58d39371ec247903acc730b7576f32db4f04b47b0bc45378496d\"" Oct 31 02:44:32.604499 containerd[1609]: time="2025-10-31T02:44:32.604431483Z" level=info msg="StartContainer for \"b8c67ae3abaf58d39371ec247903acc730b7576f32db4f04b47b0bc45378496d\" returns successfully" Oct 31 02:44:33.573073 kubelet[2864]: E1031 02:44:33.572972 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:33.766073 kubelet[2864]: I1031 02:44:33.765183 2864 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 31 02:44:35.573229 kubelet[2864]: E1031 02:44:35.571914 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:37.574922 kubelet[2864]: E1031 02:44:37.572742 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:37.762624 containerd[1609]: time="2025-10-31T02:44:37.762459574Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:37.764468 containerd[1609]: time="2025-10-31T02:44:37.764389758Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Oct 31 02:44:37.765550 containerd[1609]: time="2025-10-31T02:44:37.765503842Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:37.769904 containerd[1609]: time="2025-10-31T02:44:37.769380624Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:37.770616 containerd[1609]: time="2025-10-31T02:44:37.770577645Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 5.513737713s" Oct 31 02:44:37.770722 containerd[1609]: time="2025-10-31T02:44:37.770631761Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Oct 31 02:44:37.776726 containerd[1609]: time="2025-10-31T02:44:37.776573348Z" level=info msg="CreateContainer within sandbox \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Oct 31 02:44:37.799770 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1673615094.mount: Deactivated successfully. Oct 31 02:44:37.805672 containerd[1609]: time="2025-10-31T02:44:37.805605772Z" level=info msg="CreateContainer within sandbox \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772\"" Oct 31 02:44:37.806429 containerd[1609]: time="2025-10-31T02:44:37.806395498Z" level=info msg="StartContainer for \"83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772\"" Oct 31 02:44:37.868857 systemd[1]: run-containerd-runc-k8s.io-83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772-runc.PFqeXw.mount: Deactivated successfully. Oct 31 02:44:37.933525 containerd[1609]: time="2025-10-31T02:44:37.933459483Z" level=info msg="StartContainer for \"83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772\" returns successfully" Oct 31 02:44:38.813718 kubelet[2864]: I1031 02:44:38.813343 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-58c6bb7b96-94hs8" podStartSLOduration=8.004736489 podStartE2EDuration="13.810729305s" podCreationTimestamp="2025-10-31 02:44:25 +0000 UTC" firstStartedPulling="2025-10-31 02:44:26.449221762 +0000 UTC m=+27.117710424" lastFinishedPulling="2025-10-31 02:44:32.255214563 +0000 UTC m=+32.923703240" observedRunningTime="2025-10-31 02:44:32.788317026 +0000 UTC m=+33.456805697" watchObservedRunningTime="2025-10-31 02:44:38.810729305 +0000 UTC m=+39.479217975" Oct 31 02:44:39.147324 kubelet[2864]: I1031 02:44:39.147210 2864 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Oct 31 02:44:39.176790 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772-rootfs.mount: Deactivated successfully. Oct 31 02:44:39.185246 containerd[1609]: time="2025-10-31T02:44:39.176227808Z" level=info msg="shim disconnected" id=83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772 namespace=k8s.io Oct 31 02:44:39.185786 containerd[1609]: time="2025-10-31T02:44:39.185746632Z" level=warning msg="cleaning up after shim disconnected" id=83595f902f867f7c345876d98d6745a7434eb70dc17284a0f42b56d501dd7772 namespace=k8s.io Oct 31 02:44:39.185911 containerd[1609]: time="2025-10-31T02:44:39.185874274Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 31 02:44:39.218905 containerd[1609]: time="2025-10-31T02:44:39.218717679Z" level=warning msg="cleanup warnings time=\"2025-10-31T02:44:39Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Oct 31 02:44:39.249129 kubelet[2864]: W1031 02:44:39.241237 2864 reflector.go:569] object-"calico-system"/"whisker-ca-bundle": failed to list *v1.ConfigMap: configmaps "whisker-ca-bundle" is forbidden: User "system:node:srv-4lokn.gb1.brightbox.com" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'srv-4lokn.gb1.brightbox.com' and this object Oct 31 02:44:39.259446 kubelet[2864]: E1031 02:44:39.259226 2864 reflector.go:166] "Unhandled Error" err="object-\"calico-system\"/\"whisker-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"whisker-ca-bundle\" is forbidden: User \"system:node:srv-4lokn.gb1.brightbox.com\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-system\": no relationship found between node 'srv-4lokn.gb1.brightbox.com' and this object" logger="UnhandledError" Oct 31 02:44:39.265396 kubelet[2864]: W1031 02:44:39.257707 2864 reflector.go:569] object-"calico-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:srv-4lokn.gb1.brightbox.com" cannot list resource "configmaps" in API group "" in the namespace "calico-apiserver": no relationship found between node 'srv-4lokn.gb1.brightbox.com' and this object Oct 31 02:44:39.265914 kubelet[2864]: E1031 02:44:39.265685 2864 reflector.go:166] "Unhandled Error" err="object-\"calico-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:srv-4lokn.gb1.brightbox.com\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'srv-4lokn.gb1.brightbox.com' and this object" logger="UnhandledError" Oct 31 02:44:39.279950 kubelet[2864]: W1031 02:44:39.257792 2864 reflector.go:569] object-"calico-apiserver"/"calico-apiserver-certs": failed to list *v1.Secret: secrets "calico-apiserver-certs" is forbidden: User "system:node:srv-4lokn.gb1.brightbox.com" cannot list resource "secrets" in API group "" in the namespace "calico-apiserver": no relationship found between node 'srv-4lokn.gb1.brightbox.com' and this object Oct 31 02:44:39.280232 kubelet[2864]: E1031 02:44:39.280198 2864 reflector.go:166] "Unhandled Error" err="object-\"calico-apiserver\"/\"calico-apiserver-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"calico-apiserver-certs\" is forbidden: User \"system:node:srv-4lokn.gb1.brightbox.com\" cannot list resource \"secrets\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'srv-4lokn.gb1.brightbox.com' and this object" logger="UnhandledError" Oct 31 02:44:39.407484 kubelet[2864]: I1031 02:44:39.406409 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/2c40f225-7c8b-464e-9dd2-615c17de403d-calico-apiserver-certs\") pod \"calico-apiserver-589687cff4-ml2bp\" (UID: \"2c40f225-7c8b-464e-9dd2-615c17de403d\") " pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" Oct 31 02:44:39.407798 kubelet[2864]: I1031 02:44:39.407773 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmgb8\" (UniqueName: \"kubernetes.io/projected/e00b0c65-1a04-42fa-84a6-f5f0d45c983a-kube-api-access-hmgb8\") pod \"coredns-668d6bf9bc-rp8wt\" (UID: \"e00b0c65-1a04-42fa-84a6-f5f0d45c983a\") " pod="kube-system/coredns-668d6bf9bc-rp8wt" Oct 31 02:44:39.408395 kubelet[2864]: I1031 02:44:39.408187 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e5f0f90-f710-4f9a-b757-2db9fe751f0e-tigera-ca-bundle\") pod \"calico-kube-controllers-74fbc74c8b-kb8vb\" (UID: \"1e5f0f90-f710-4f9a-b757-2db9fe751f0e\") " pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" Oct 31 02:44:39.408395 kubelet[2864]: I1031 02:44:39.408236 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-455vv\" (UniqueName: \"kubernetes.io/projected/1e5f0f90-f710-4f9a-b757-2db9fe751f0e-kube-api-access-455vv\") pod \"calico-kube-controllers-74fbc74c8b-kb8vb\" (UID: \"1e5f0f90-f710-4f9a-b757-2db9fe751f0e\") " pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" Oct 31 02:44:39.408395 kubelet[2864]: I1031 02:44:39.408272 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1d3e6a01-de45-45a9-bd90-9cad31c5290e-calico-apiserver-certs\") pod \"calico-apiserver-589687cff4-5xd2q\" (UID: \"1d3e6a01-de45-45a9-bd90-9cad31c5290e\") " pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" Oct 31 02:44:39.409318 kubelet[2864]: I1031 02:44:39.408636 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/868ae699-bd25-491c-a828-139c5285406d-whisker-ca-bundle\") pod \"whisker-59b4654fb4-jbkzr\" (UID: \"868ae699-bd25-491c-a828-139c5285406d\") " pod="calico-system/whisker-59b4654fb4-jbkzr" Oct 31 02:44:39.409318 kubelet[2864]: I1031 02:44:39.408682 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqpd8\" (UniqueName: \"kubernetes.io/projected/1d3e6a01-de45-45a9-bd90-9cad31c5290e-kube-api-access-xqpd8\") pod \"calico-apiserver-589687cff4-5xd2q\" (UID: \"1d3e6a01-de45-45a9-bd90-9cad31c5290e\") " pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" Oct 31 02:44:39.409318 kubelet[2864]: I1031 02:44:39.408724 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/51cf0923-33f8-4c41-8a1b-075035cdaba9-calico-apiserver-certs\") pod \"calico-apiserver-75444f6b5f-gp7s4\" (UID: \"51cf0923-33f8-4c41-8a1b-075035cdaba9\") " pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" Oct 31 02:44:39.409318 kubelet[2864]: I1031 02:44:39.408757 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx5f2\" (UniqueName: \"kubernetes.io/projected/c44deaed-49c1-414b-a77a-e9b3a43a004d-kube-api-access-rx5f2\") pod \"goldmane-666569f655-n7gvx\" (UID: \"c44deaed-49c1-414b-a77a-e9b3a43a004d\") " pod="calico-system/goldmane-666569f655-n7gvx" Oct 31 02:44:39.409318 kubelet[2864]: I1031 02:44:39.408786 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/c44deaed-49c1-414b-a77a-e9b3a43a004d-goldmane-key-pair\") pod \"goldmane-666569f655-n7gvx\" (UID: \"c44deaed-49c1-414b-a77a-e9b3a43a004d\") " pod="calico-system/goldmane-666569f655-n7gvx" Oct 31 02:44:39.409631 kubelet[2864]: I1031 02:44:39.408823 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krkfx\" (UniqueName: \"kubernetes.io/projected/51cf0923-33f8-4c41-8a1b-075035cdaba9-kube-api-access-krkfx\") pod \"calico-apiserver-75444f6b5f-gp7s4\" (UID: \"51cf0923-33f8-4c41-8a1b-075035cdaba9\") " pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" Oct 31 02:44:39.409631 kubelet[2864]: I1031 02:44:39.408850 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e00b0c65-1a04-42fa-84a6-f5f0d45c983a-config-volume\") pod \"coredns-668d6bf9bc-rp8wt\" (UID: \"e00b0c65-1a04-42fa-84a6-f5f0d45c983a\") " pod="kube-system/coredns-668d6bf9bc-rp8wt" Oct 31 02:44:39.409631 kubelet[2864]: I1031 02:44:39.408884 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/868ae699-bd25-491c-a828-139c5285406d-whisker-backend-key-pair\") pod \"whisker-59b4654fb4-jbkzr\" (UID: \"868ae699-bd25-491c-a828-139c5285406d\") " pod="calico-system/whisker-59b4654fb4-jbkzr" Oct 31 02:44:39.409631 kubelet[2864]: I1031 02:44:39.408931 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c44deaed-49c1-414b-a77a-e9b3a43a004d-config\") pod \"goldmane-666569f655-n7gvx\" (UID: \"c44deaed-49c1-414b-a77a-e9b3a43a004d\") " pod="calico-system/goldmane-666569f655-n7gvx" Oct 31 02:44:39.409631 kubelet[2864]: I1031 02:44:39.408957 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c44deaed-49c1-414b-a77a-e9b3a43a004d-goldmane-ca-bundle\") pod \"goldmane-666569f655-n7gvx\" (UID: \"c44deaed-49c1-414b-a77a-e9b3a43a004d\") " pod="calico-system/goldmane-666569f655-n7gvx" Oct 31 02:44:39.409896 kubelet[2864]: I1031 02:44:39.408982 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f40f31f8-0b31-408d-8016-7a7232bfce65-config-volume\") pod \"coredns-668d6bf9bc-fhw82\" (UID: \"f40f31f8-0b31-408d-8016-7a7232bfce65\") " pod="kube-system/coredns-668d6bf9bc-fhw82" Oct 31 02:44:39.409896 kubelet[2864]: I1031 02:44:39.409022 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l45g\" (UniqueName: \"kubernetes.io/projected/f40f31f8-0b31-408d-8016-7a7232bfce65-kube-api-access-6l45g\") pod \"coredns-668d6bf9bc-fhw82\" (UID: \"f40f31f8-0b31-408d-8016-7a7232bfce65\") " pod="kube-system/coredns-668d6bf9bc-fhw82" Oct 31 02:44:39.409896 kubelet[2864]: I1031 02:44:39.409069 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbkxq\" (UniqueName: \"kubernetes.io/projected/868ae699-bd25-491c-a828-139c5285406d-kube-api-access-bbkxq\") pod \"whisker-59b4654fb4-jbkzr\" (UID: \"868ae699-bd25-491c-a828-139c5285406d\") " pod="calico-system/whisker-59b4654fb4-jbkzr" Oct 31 02:44:39.409896 kubelet[2864]: I1031 02:44:39.409110 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tffjx\" (UniqueName: \"kubernetes.io/projected/2c40f225-7c8b-464e-9dd2-615c17de403d-kube-api-access-tffjx\") pod \"calico-apiserver-589687cff4-ml2bp\" (UID: \"2c40f225-7c8b-464e-9dd2-615c17de403d\") " pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" Oct 31 02:44:39.590558 containerd[1609]: time="2025-10-31T02:44:39.590310504Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fhw82,Uid:f40f31f8-0b31-408d-8016-7a7232bfce65,Namespace:kube-system,Attempt:0,}" Oct 31 02:44:39.591906 containerd[1609]: time="2025-10-31T02:44:39.591691412Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-d2znc,Uid:7293ccde-9a5c-4930-845b-a0c0b00c9ac3,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:39.592066 containerd[1609]: time="2025-10-31T02:44:39.591707919Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74fbc74c8b-kb8vb,Uid:1e5f0f90-f710-4f9a-b757-2db9fe751f0e,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:39.602856 containerd[1609]: time="2025-10-31T02:44:39.602817416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-n7gvx,Uid:c44deaed-49c1-414b-a77a-e9b3a43a004d,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:39.792445 containerd[1609]: time="2025-10-31T02:44:39.792391582Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Oct 31 02:44:39.842967 containerd[1609]: time="2025-10-31T02:44:39.842814009Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rp8wt,Uid:e00b0c65-1a04-42fa-84a6-f5f0d45c983a,Namespace:kube-system,Attempt:0,}" Oct 31 02:44:40.009680 containerd[1609]: time="2025-10-31T02:44:40.007472232Z" level=error msg="Failed to destroy network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.028155 containerd[1609]: time="2025-10-31T02:44:40.026782769Z" level=error msg="encountered an error cleaning up failed sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.038628 containerd[1609]: time="2025-10-31T02:44:40.038548034Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fhw82,Uid:f40f31f8-0b31-408d-8016-7a7232bfce65,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.051255 containerd[1609]: time="2025-10-31T02:44:40.051109071Z" level=error msg="Failed to destroy network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.051867 containerd[1609]: time="2025-10-31T02:44:40.051829820Z" level=error msg="encountered an error cleaning up failed sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.052709 containerd[1609]: time="2025-10-31T02:44:40.052670084Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-n7gvx,Uid:c44deaed-49c1-414b-a77a-e9b3a43a004d,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.054705 containerd[1609]: time="2025-10-31T02:44:40.054639339Z" level=error msg="Failed to destroy network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.055254 containerd[1609]: time="2025-10-31T02:44:40.055204490Z" level=error msg="encountered an error cleaning up failed sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.055332 containerd[1609]: time="2025-10-31T02:44:40.055264541Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74fbc74c8b-kb8vb,Uid:1e5f0f90-f710-4f9a-b757-2db9fe751f0e,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.055416 containerd[1609]: time="2025-10-31T02:44:40.055368127Z" level=error msg="Failed to destroy network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.055815 containerd[1609]: time="2025-10-31T02:44:40.055775360Z" level=error msg="encountered an error cleaning up failed sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.055883 containerd[1609]: time="2025-10-31T02:44:40.055826314Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-d2znc,Uid:7293ccde-9a5c-4930-845b-a0c0b00c9ac3,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.056943 kubelet[2864]: E1031 02:44:40.056233 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.056943 kubelet[2864]: E1031 02:44:40.056293 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.056943 kubelet[2864]: E1031 02:44:40.056350 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:40.056943 kubelet[2864]: E1031 02:44:40.056394 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.058631 kubelet[2864]: E1031 02:44:40.056401 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-d2znc" Oct 31 02:44:40.058631 kubelet[2864]: E1031 02:44:40.056424 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" Oct 31 02:44:40.058631 kubelet[2864]: E1031 02:44:40.056459 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" Oct 31 02:44:40.059841 kubelet[2864]: E1031 02:44:40.056522 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:40.059841 kubelet[2864]: E1031 02:44:40.056350 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-n7gvx" Oct 31 02:44:40.059841 kubelet[2864]: E1031 02:44:40.056524 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-74fbc74c8b-kb8vb_calico-system(1e5f0f90-f710-4f9a-b757-2db9fe751f0e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-74fbc74c8b-kb8vb_calico-system(1e5f0f90-f710-4f9a-b757-2db9fe751f0e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:44:40.060308 kubelet[2864]: E1031 02:44:40.056580 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-n7gvx" Oct 31 02:44:40.060308 kubelet[2864]: E1031 02:44:40.056633 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-n7gvx_calico-system(c44deaed-49c1-414b-a77a-e9b3a43a004d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-n7gvx_calico-system(c44deaed-49c1-414b-a77a-e9b3a43a004d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:44:40.060308 kubelet[2864]: E1031 02:44:40.056233 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.060588 kubelet[2864]: E1031 02:44:40.056682 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-fhw82" Oct 31 02:44:40.060588 kubelet[2864]: E1031 02:44:40.056708 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-fhw82" Oct 31 02:44:40.060588 kubelet[2864]: E1031 02:44:40.056741 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-fhw82_kube-system(f40f31f8-0b31-408d-8016-7a7232bfce65)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-fhw82_kube-system(f40f31f8-0b31-408d-8016-7a7232bfce65)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-fhw82" podUID="f40f31f8-0b31-408d-8016-7a7232bfce65" Oct 31 02:44:40.125479 containerd[1609]: time="2025-10-31T02:44:40.125132533Z" level=error msg="Failed to destroy network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.126322 containerd[1609]: time="2025-10-31T02:44:40.126102761Z" level=error msg="encountered an error cleaning up failed sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.126322 containerd[1609]: time="2025-10-31T02:44:40.126211855Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rp8wt,Uid:e00b0c65-1a04-42fa-84a6-f5f0d45c983a,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.126777 kubelet[2864]: E1031 02:44:40.126699 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.126910 kubelet[2864]: E1031 02:44:40.126789 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-rp8wt" Oct 31 02:44:40.126910 kubelet[2864]: E1031 02:44:40.126821 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-rp8wt" Oct 31 02:44:40.127178 kubelet[2864]: E1031 02:44:40.126894 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-rp8wt_kube-system(e00b0c65-1a04-42fa-84a6-f5f0d45c983a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-rp8wt_kube-system(e00b0c65-1a04-42fa-84a6-f5f0d45c983a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-rp8wt" podUID="e00b0c65-1a04-42fa-84a6-f5f0d45c983a" Oct 31 02:44:40.188141 containerd[1609]: time="2025-10-31T02:44:40.187419032Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-5xd2q,Uid:1d3e6a01-de45-45a9-bd90-9cad31c5290e,Namespace:calico-apiserver,Attempt:0,}" Oct 31 02:44:40.215582 containerd[1609]: time="2025-10-31T02:44:40.215535302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-75444f6b5f-gp7s4,Uid:51cf0923-33f8-4c41-8a1b-075035cdaba9,Namespace:calico-apiserver,Attempt:0,}" Oct 31 02:44:40.328929 containerd[1609]: time="2025-10-31T02:44:40.328585626Z" level=error msg="Failed to destroy network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.329764 containerd[1609]: time="2025-10-31T02:44:40.329626369Z" level=error msg="encountered an error cleaning up failed sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.329764 containerd[1609]: time="2025-10-31T02:44:40.329717212Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-5xd2q,Uid:1d3e6a01-de45-45a9-bd90-9cad31c5290e,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.331796 kubelet[2864]: E1031 02:44:40.330365 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.331796 kubelet[2864]: E1031 02:44:40.330475 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" Oct 31 02:44:40.331796 kubelet[2864]: E1031 02:44:40.330508 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" Oct 31 02:44:40.332084 kubelet[2864]: E1031 02:44:40.330562 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-589687cff4-5xd2q_calico-apiserver(1d3e6a01-de45-45a9-bd90-9cad31c5290e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-589687cff4-5xd2q_calico-apiserver(1d3e6a01-de45-45a9-bd90-9cad31c5290e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:44:40.347314 containerd[1609]: time="2025-10-31T02:44:40.347261269Z" level=error msg="Failed to destroy network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.347734 containerd[1609]: time="2025-10-31T02:44:40.347697369Z" level=error msg="encountered an error cleaning up failed sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.347854 containerd[1609]: time="2025-10-31T02:44:40.347761520Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-75444f6b5f-gp7s4,Uid:51cf0923-33f8-4c41-8a1b-075035cdaba9,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.348137 kubelet[2864]: E1031 02:44:40.348016 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.348137 kubelet[2864]: E1031 02:44:40.348112 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" Oct 31 02:44:40.348372 kubelet[2864]: E1031 02:44:40.348144 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" Oct 31 02:44:40.348372 kubelet[2864]: E1031 02:44:40.348214 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-75444f6b5f-gp7s4_calico-apiserver(51cf0923-33f8-4c41-8a1b-075035cdaba9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-75444f6b5f-gp7s4_calico-apiserver(51cf0923-33f8-4c41-8a1b-075035cdaba9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:44:40.451194 containerd[1609]: time="2025-10-31T02:44:40.451135882Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-59b4654fb4-jbkzr,Uid:868ae699-bd25-491c-a828-139c5285406d,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:40.478523 containerd[1609]: time="2025-10-31T02:44:40.478054832Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-ml2bp,Uid:2c40f225-7c8b-464e-9dd2-615c17de403d,Namespace:calico-apiserver,Attempt:0,}" Oct 31 02:44:40.559694 containerd[1609]: time="2025-10-31T02:44:40.559630758Z" level=error msg="Failed to destroy network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.560527 containerd[1609]: time="2025-10-31T02:44:40.560152144Z" level=error msg="encountered an error cleaning up failed sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.560527 containerd[1609]: time="2025-10-31T02:44:40.560228622Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-59b4654fb4-jbkzr,Uid:868ae699-bd25-491c-a828-139c5285406d,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.561827 kubelet[2864]: E1031 02:44:40.561687 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.562257 kubelet[2864]: E1031 02:44:40.561929 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-59b4654fb4-jbkzr" Oct 31 02:44:40.562257 kubelet[2864]: E1031 02:44:40.562182 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-59b4654fb4-jbkzr" Oct 31 02:44:40.562656 kubelet[2864]: E1031 02:44:40.562398 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-59b4654fb4-jbkzr_calico-system(868ae699-bd25-491c-a828-139c5285406d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-59b4654fb4-jbkzr_calico-system(868ae699-bd25-491c-a828-139c5285406d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-59b4654fb4-jbkzr" podUID="868ae699-bd25-491c-a828-139c5285406d" Oct 31 02:44:40.577733 containerd[1609]: time="2025-10-31T02:44:40.577664179Z" level=error msg="Failed to destroy network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.578161 containerd[1609]: time="2025-10-31T02:44:40.578122622Z" level=error msg="encountered an error cleaning up failed sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.578300 containerd[1609]: time="2025-10-31T02:44:40.578189326Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-ml2bp,Uid:2c40f225-7c8b-464e-9dd2-615c17de403d,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.578865 kubelet[2864]: E1031 02:44:40.578587 2864 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.578865 kubelet[2864]: E1031 02:44:40.578708 2864 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" Oct 31 02:44:40.579255 kubelet[2864]: E1031 02:44:40.578822 2864 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" Oct 31 02:44:40.580114 kubelet[2864]: E1031 02:44:40.579167 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-589687cff4-ml2bp_calico-apiserver(2c40f225-7c8b-464e-9dd2-615c17de403d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-589687cff4-ml2bp_calico-apiserver(2c40f225-7c8b-464e-9dd2-615c17de403d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:44:40.794075 kubelet[2864]: I1031 02:44:40.793919 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:44:40.802270 kubelet[2864]: I1031 02:44:40.801527 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:44:40.802420 containerd[1609]: time="2025-10-31T02:44:40.802140756Z" level=info msg="StopPodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\"" Oct 31 02:44:40.804529 containerd[1609]: time="2025-10-31T02:44:40.803629597Z" level=info msg="Ensure that sandbox d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f in task-service has been cleanup successfully" Oct 31 02:44:40.805467 containerd[1609]: time="2025-10-31T02:44:40.804914362Z" level=info msg="StopPodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\"" Oct 31 02:44:40.805467 containerd[1609]: time="2025-10-31T02:44:40.805136766Z" level=info msg="Ensure that sandbox ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad in task-service has been cleanup successfully" Oct 31 02:44:40.806679 kubelet[2864]: I1031 02:44:40.806646 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:44:40.808831 containerd[1609]: time="2025-10-31T02:44:40.808211311Z" level=info msg="StopPodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\"" Oct 31 02:44:40.808831 containerd[1609]: time="2025-10-31T02:44:40.808537731Z" level=info msg="Ensure that sandbox 3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6 in task-service has been cleanup successfully" Oct 31 02:44:40.813336 kubelet[2864]: I1031 02:44:40.812130 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:44:40.817112 containerd[1609]: time="2025-10-31T02:44:40.816538701Z" level=info msg="StopPodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\"" Oct 31 02:44:40.817621 containerd[1609]: time="2025-10-31T02:44:40.817459748Z" level=info msg="Ensure that sandbox 59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f in task-service has been cleanup successfully" Oct 31 02:44:40.822439 kubelet[2864]: I1031 02:44:40.822284 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:44:40.835587 containerd[1609]: time="2025-10-31T02:44:40.835366539Z" level=info msg="StopPodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\"" Oct 31 02:44:40.837122 containerd[1609]: time="2025-10-31T02:44:40.835588107Z" level=info msg="Ensure that sandbox af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf in task-service has been cleanup successfully" Oct 31 02:44:40.840236 kubelet[2864]: I1031 02:44:40.840052 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:44:40.854775 containerd[1609]: time="2025-10-31T02:44:40.854295266Z" level=info msg="StopPodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\"" Oct 31 02:44:40.854775 containerd[1609]: time="2025-10-31T02:44:40.854537307Z" level=info msg="Ensure that sandbox 22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04 in task-service has been cleanup successfully" Oct 31 02:44:40.866684 kubelet[2864]: I1031 02:44:40.866644 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:40.870400 containerd[1609]: time="2025-10-31T02:44:40.869420319Z" level=info msg="StopPodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\"" Oct 31 02:44:40.871378 containerd[1609]: time="2025-10-31T02:44:40.871195790Z" level=info msg="Ensure that sandbox 08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba in task-service has been cleanup successfully" Oct 31 02:44:40.888787 kubelet[2864]: I1031 02:44:40.888732 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:44:40.902615 containerd[1609]: time="2025-10-31T02:44:40.902411479Z" level=info msg="StopPodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\"" Oct 31 02:44:40.902783 containerd[1609]: time="2025-10-31T02:44:40.902646963Z" level=info msg="Ensure that sandbox bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646 in task-service has been cleanup successfully" Oct 31 02:44:40.908509 kubelet[2864]: I1031 02:44:40.907785 2864 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:40.911362 containerd[1609]: time="2025-10-31T02:44:40.910361899Z" level=info msg="StopPodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\"" Oct 31 02:44:40.911362 containerd[1609]: time="2025-10-31T02:44:40.910566581Z" level=info msg="Ensure that sandbox 924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149 in task-service has been cleanup successfully" Oct 31 02:44:40.987899 containerd[1609]: time="2025-10-31T02:44:40.987833430Z" level=error msg="StopPodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" failed" error="failed to destroy network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:40.988718 kubelet[2864]: E1031 02:44:40.988658 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:44:41.008958 kubelet[2864]: E1031 02:44:40.988755 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6"} Oct 31 02:44:41.008958 kubelet[2864]: E1031 02:44:41.008646 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"e00b0c65-1a04-42fa-84a6-f5f0d45c983a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.008958 kubelet[2864]: E1031 02:44:41.008688 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"e00b0c65-1a04-42fa-84a6-f5f0d45c983a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-rp8wt" podUID="e00b0c65-1a04-42fa-84a6-f5f0d45c983a" Oct 31 02:44:41.052867 containerd[1609]: time="2025-10-31T02:44:41.052142646Z" level=error msg="StopPodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" failed" error="failed to destroy network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.057164 kubelet[2864]: E1031 02:44:41.056130 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:44:41.057164 kubelet[2864]: E1031 02:44:41.056285 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf"} Oct 31 02:44:41.057164 kubelet[2864]: E1031 02:44:41.056338 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"868ae699-bd25-491c-a828-139c5285406d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.057164 kubelet[2864]: E1031 02:44:41.056392 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"868ae699-bd25-491c-a828-139c5285406d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-59b4654fb4-jbkzr" podUID="868ae699-bd25-491c-a828-139c5285406d" Oct 31 02:44:41.061386 containerd[1609]: time="2025-10-31T02:44:41.059778566Z" level=error msg="StopPodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" failed" error="failed to destroy network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.061500 kubelet[2864]: E1031 02:44:41.061198 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:44:41.061500 kubelet[2864]: E1031 02:44:41.061240 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f"} Oct 31 02:44:41.061500 kubelet[2864]: E1031 02:44:41.061288 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"51cf0923-33f8-4c41-8a1b-075035cdaba9\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.061500 kubelet[2864]: E1031 02:44:41.061321 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"51cf0923-33f8-4c41-8a1b-075035cdaba9\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:44:41.069821 containerd[1609]: time="2025-10-31T02:44:41.069246928Z" level=error msg="StopPodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" failed" error="failed to destroy network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.069821 containerd[1609]: time="2025-10-31T02:44:41.069749850Z" level=error msg="StopPodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" failed" error="failed to destroy network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.070255 kubelet[2864]: E1031 02:44:41.069532 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:44:41.070255 kubelet[2864]: E1031 02:44:41.069592 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04"} Oct 31 02:44:41.070255 kubelet[2864]: E1031 02:44:41.069646 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c44deaed-49c1-414b-a77a-e9b3a43a004d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.070255 kubelet[2864]: E1031 02:44:41.069678 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c44deaed-49c1-414b-a77a-e9b3a43a004d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:44:41.071172 kubelet[2864]: E1031 02:44:41.069980 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:44:41.071172 kubelet[2864]: E1031 02:44:41.070047 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad"} Oct 31 02:44:41.071172 kubelet[2864]: E1031 02:44:41.070100 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2c40f225-7c8b-464e-9dd2-615c17de403d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.071172 kubelet[2864]: E1031 02:44:41.070143 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2c40f225-7c8b-464e-9dd2-615c17de403d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:44:41.084379 containerd[1609]: time="2025-10-31T02:44:41.084326860Z" level=error msg="StopPodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" failed" error="failed to destroy network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.084954 kubelet[2864]: E1031 02:44:41.084739 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:41.084954 kubelet[2864]: E1031 02:44:41.084802 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149"} Oct 31 02:44:41.084954 kubelet[2864]: E1031 02:44:41.084846 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1d3e6a01-de45-45a9-bd90-9cad31c5290e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.084954 kubelet[2864]: E1031 02:44:41.084877 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1d3e6a01-de45-45a9-bd90-9cad31c5290e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:44:41.091102 containerd[1609]: time="2025-10-31T02:44:41.089339410Z" level=error msg="StopPodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" failed" error="failed to destroy network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.091206 kubelet[2864]: E1031 02:44:41.089543 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:44:41.091206 kubelet[2864]: E1031 02:44:41.089628 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646"} Oct 31 02:44:41.091206 kubelet[2864]: E1031 02:44:41.089675 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.091206 kubelet[2864]: E1031 02:44:41.089705 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"7293ccde-9a5c-4930-845b-a0c0b00c9ac3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:41.098339 containerd[1609]: time="2025-10-31T02:44:41.098168400Z" level=error msg="StopPodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" failed" error="failed to destroy network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.098786 kubelet[2864]: E1031 02:44:41.098456 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:44:41.098786 kubelet[2864]: E1031 02:44:41.098528 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f"} Oct 31 02:44:41.098786 kubelet[2864]: E1031 02:44:41.098582 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f40f31f8-0b31-408d-8016-7a7232bfce65\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.098786 kubelet[2864]: E1031 02:44:41.098632 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f40f31f8-0b31-408d-8016-7a7232bfce65\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-fhw82" podUID="f40f31f8-0b31-408d-8016-7a7232bfce65" Oct 31 02:44:41.100868 containerd[1609]: time="2025-10-31T02:44:41.100828870Z" level=error msg="StopPodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" failed" error="failed to destroy network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 31 02:44:41.101237 kubelet[2864]: E1031 02:44:41.101050 2864 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:41.101237 kubelet[2864]: E1031 02:44:41.101097 2864 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba"} Oct 31 02:44:41.101237 kubelet[2864]: E1031 02:44:41.101133 2864 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1e5f0f90-f710-4f9a-b757-2db9fe751f0e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Oct 31 02:44:41.101237 kubelet[2864]: E1031 02:44:41.101180 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1e5f0f90-f710-4f9a-b757-2db9fe751f0e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:44:41.171425 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f-shm.mount: Deactivated successfully. Oct 31 02:44:41.171773 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149-shm.mount: Deactivated successfully. Oct 31 02:44:50.102639 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1556888031.mount: Deactivated successfully. Oct 31 02:44:50.240437 containerd[1609]: time="2025-10-31T02:44:50.232212696Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Oct 31 02:44:50.245350 containerd[1609]: time="2025-10-31T02:44:50.245282562Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:50.245746 containerd[1609]: time="2025-10-31T02:44:50.245322550Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 10.452567348s" Oct 31 02:44:50.245746 containerd[1609]: time="2025-10-31T02:44:50.245613217Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Oct 31 02:44:50.292706 containerd[1609]: time="2025-10-31T02:44:50.292544426Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:50.294803 containerd[1609]: time="2025-10-31T02:44:50.294758891Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 31 02:44:50.362957 containerd[1609]: time="2025-10-31T02:44:50.362121794Z" level=info msg="CreateContainer within sandbox \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Oct 31 02:44:50.430210 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount347454770.mount: Deactivated successfully. Oct 31 02:44:50.437903 containerd[1609]: time="2025-10-31T02:44:50.437840039Z" level=info msg="CreateContainer within sandbox \"fb5051496fa0e66bf417ea64e5636b24f8d18d0895a4087be3ba0899a4dee8d6\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"ac9c636af7438b38dd0192eb73f8fd3e81450384974fc37d6067ac5251e59d20\"" Oct 31 02:44:50.445493 containerd[1609]: time="2025-10-31T02:44:50.445448761Z" level=info msg="StartContainer for \"ac9c636af7438b38dd0192eb73f8fd3e81450384974fc37d6067ac5251e59d20\"" Oct 31 02:44:50.714837 containerd[1609]: time="2025-10-31T02:44:50.714495510Z" level=info msg="StartContainer for \"ac9c636af7438b38dd0192eb73f8fd3e81450384974fc37d6067ac5251e59d20\" returns successfully" Oct 31 02:44:50.976895 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Oct 31 02:44:50.978732 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Oct 31 02:44:51.020969 kubelet[2864]: I1031 02:44:51.017434 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-qfjs8" podStartSLOduration=2.118721901 podStartE2EDuration="26.017374317s" podCreationTimestamp="2025-10-31 02:44:25 +0000 UTC" firstStartedPulling="2025-10-31 02:44:26.394544247 +0000 UTC m=+27.063032909" lastFinishedPulling="2025-10-31 02:44:50.293196659 +0000 UTC m=+50.961685325" observedRunningTime="2025-10-31 02:44:51.013976789 +0000 UTC m=+51.682465494" watchObservedRunningTime="2025-10-31 02:44:51.017374317 +0000 UTC m=+51.685862986" Oct 31 02:44:51.235025 containerd[1609]: time="2025-10-31T02:44:51.234384465Z" level=info msg="StopPodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\"" Oct 31 02:44:51.580017 containerd[1609]: time="2025-10-31T02:44:51.579422798Z" level=info msg="StopPodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\"" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.389 [INFO][4070] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.389 [INFO][4070] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" iface="eth0" netns="/var/run/netns/cni-6f545d36-0140-0500-fb83-cf03765b813e" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.390 [INFO][4070] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" iface="eth0" netns="/var/run/netns/cni-6f545d36-0140-0500-fb83-cf03765b813e" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.392 [INFO][4070] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" iface="eth0" netns="/var/run/netns/cni-6f545d36-0140-0500-fb83-cf03765b813e" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.392 [INFO][4070] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.392 [INFO][4070] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.710 [INFO][4077] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.715 [INFO][4077] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.717 [INFO][4077] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.744 [WARNING][4077] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.744 [INFO][4077] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.749 [INFO][4077] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:51.769402 containerd[1609]: 2025-10-31 02:44:51.756 [INFO][4070] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:44:51.769402 containerd[1609]: time="2025-10-31T02:44:51.769140233Z" level=info msg="TearDown network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" successfully" Oct 31 02:44:51.769402 containerd[1609]: time="2025-10-31T02:44:51.769176252Z" level=info msg="StopPodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" returns successfully" Oct 31 02:44:51.778567 systemd[1]: run-netns-cni\x2d6f545d36\x2d0140\x2d0500\x2dfb83\x2dcf03765b813e.mount: Deactivated successfully. Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.786 [INFO][4091] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.787 [INFO][4091] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" iface="eth0" netns="/var/run/netns/cni-500b1676-a373-9396-6a34-e808f3d4efa2" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.788 [INFO][4091] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" iface="eth0" netns="/var/run/netns/cni-500b1676-a373-9396-6a34-e808f3d4efa2" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.789 [INFO][4091] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" iface="eth0" netns="/var/run/netns/cni-500b1676-a373-9396-6a34-e808f3d4efa2" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.789 [INFO][4091] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.789 [INFO][4091] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.853 [INFO][4101] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.853 [INFO][4101] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.853 [INFO][4101] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.865 [WARNING][4101] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.865 [INFO][4101] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.867 [INFO][4101] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:51.874579 containerd[1609]: 2025-10-31 02:44:51.871 [INFO][4091] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:44:51.880227 containerd[1609]: time="2025-10-31T02:44:51.875614607Z" level=info msg="TearDown network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" successfully" Oct 31 02:44:51.880227 containerd[1609]: time="2025-10-31T02:44:51.875650868Z" level=info msg="StopPodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" returns successfully" Oct 31 02:44:51.880227 containerd[1609]: time="2025-10-31T02:44:51.878331182Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-ml2bp,Uid:2c40f225-7c8b-464e-9dd2-615c17de403d,Namespace:calico-apiserver,Attempt:1,}" Oct 31 02:44:51.881629 systemd[1]: run-netns-cni\x2d500b1676\x2da373\x2d9396\x2d6a34\x2de808f3d4efa2.mount: Deactivated successfully. Oct 31 02:44:51.945123 kubelet[2864]: I1031 02:44:51.943320 2864 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/868ae699-bd25-491c-a828-139c5285406d-whisker-ca-bundle\") pod \"868ae699-bd25-491c-a828-139c5285406d\" (UID: \"868ae699-bd25-491c-a828-139c5285406d\") " Oct 31 02:44:51.945123 kubelet[2864]: I1031 02:44:51.943870 2864 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/868ae699-bd25-491c-a828-139c5285406d-whisker-backend-key-pair\") pod \"868ae699-bd25-491c-a828-139c5285406d\" (UID: \"868ae699-bd25-491c-a828-139c5285406d\") " Oct 31 02:44:51.945123 kubelet[2864]: I1031 02:44:51.943936 2864 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbkxq\" (UniqueName: \"kubernetes.io/projected/868ae699-bd25-491c-a828-139c5285406d-kube-api-access-bbkxq\") pod \"868ae699-bd25-491c-a828-139c5285406d\" (UID: \"868ae699-bd25-491c-a828-139c5285406d\") " Oct 31 02:44:51.972604 kubelet[2864]: I1031 02:44:51.971432 2864 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/868ae699-bd25-491c-a828-139c5285406d-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "868ae699-bd25-491c-a828-139c5285406d" (UID: "868ae699-bd25-491c-a828-139c5285406d"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Oct 31 02:44:51.974233 kubelet[2864]: I1031 02:44:51.974176 2864 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/868ae699-bd25-491c-a828-139c5285406d-kube-api-access-bbkxq" (OuterVolumeSpecName: "kube-api-access-bbkxq") pod "868ae699-bd25-491c-a828-139c5285406d" (UID: "868ae699-bd25-491c-a828-139c5285406d"). InnerVolumeSpecName "kube-api-access-bbkxq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Oct 31 02:44:51.974418 kubelet[2864]: I1031 02:44:51.974346 2864 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/868ae699-bd25-491c-a828-139c5285406d-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "868ae699-bd25-491c-a828-139c5285406d" (UID: "868ae699-bd25-491c-a828-139c5285406d"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Oct 31 02:44:52.005052 kubelet[2864]: I1031 02:44:52.004278 2864 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 31 02:44:52.050215 kubelet[2864]: I1031 02:44:52.050065 2864 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bbkxq\" (UniqueName: \"kubernetes.io/projected/868ae699-bd25-491c-a828-139c5285406d-kube-api-access-bbkxq\") on node \"srv-4lokn.gb1.brightbox.com\" DevicePath \"\"" Oct 31 02:44:52.054547 kubelet[2864]: I1031 02:44:52.052834 2864 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/868ae699-bd25-491c-a828-139c5285406d-whisker-ca-bundle\") on node \"srv-4lokn.gb1.brightbox.com\" DevicePath \"\"" Oct 31 02:44:52.054547 kubelet[2864]: I1031 02:44:52.052866 2864 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/868ae699-bd25-491c-a828-139c5285406d-whisker-backend-key-pair\") on node \"srv-4lokn.gb1.brightbox.com\" DevicePath \"\"" Oct 31 02:44:52.111652 systemd[1]: var-lib-kubelet-pods-868ae699\x2dbd25\x2d491c\x2da828\x2d139c5285406d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbbkxq.mount: Deactivated successfully. Oct 31 02:44:52.112012 systemd[1]: var-lib-kubelet-pods-868ae699\x2dbd25\x2d491c\x2da828\x2d139c5285406d-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Oct 31 02:44:52.257909 kubelet[2864]: I1031 02:44:52.254282 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec0f842a-be78-41a7-a220-2e129c052a42-whisker-ca-bundle\") pod \"whisker-5d659c4594-7qhbb\" (UID: \"ec0f842a-be78-41a7-a220-2e129c052a42\") " pod="calico-system/whisker-5d659c4594-7qhbb" Oct 31 02:44:52.257909 kubelet[2864]: I1031 02:44:52.254355 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6djh\" (UniqueName: \"kubernetes.io/projected/ec0f842a-be78-41a7-a220-2e129c052a42-kube-api-access-s6djh\") pod \"whisker-5d659c4594-7qhbb\" (UID: \"ec0f842a-be78-41a7-a220-2e129c052a42\") " pod="calico-system/whisker-5d659c4594-7qhbb" Oct 31 02:44:52.257909 kubelet[2864]: I1031 02:44:52.254400 2864 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ec0f842a-be78-41a7-a220-2e129c052a42-whisker-backend-key-pair\") pod \"whisker-5d659c4594-7qhbb\" (UID: \"ec0f842a-be78-41a7-a220-2e129c052a42\") " pod="calico-system/whisker-5d659c4594-7qhbb" Oct 31 02:44:52.324200 systemd-networkd[1257]: calia0a200d3e49: Link UP Oct 31 02:44:52.325705 systemd-networkd[1257]: calia0a200d3e49: Gained carrier Oct 31 02:44:52.378174 systemd-journald[1184]: Under memory pressure, flushing caches. Oct 31 02:44:52.367791 systemd-resolved[1505]: Under memory pressure, flushing caches. Oct 31 02:44:52.367858 systemd-resolved[1505]: Flushed all caches. Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.010 [INFO][4119] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.050 [INFO][4119] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0 calico-apiserver-589687cff4- calico-apiserver 2c40f225-7c8b-464e-9dd2-615c17de403d 938 0 2025-10-31 02:44:19 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:589687cff4 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com calico-apiserver-589687cff4-ml2bp eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calia0a200d3e49 [] [] }} ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.051 [INFO][4119] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.200 [INFO][4133] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" HandleID="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.202 [INFO][4133] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" HandleID="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00043c050), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"srv-4lokn.gb1.brightbox.com", "pod":"calico-apiserver-589687cff4-ml2bp", "timestamp":"2025-10-31 02:44:52.200096575 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.202 [INFO][4133] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.202 [INFO][4133] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.202 [INFO][4133] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.234 [INFO][4133] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.259 [INFO][4133] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.269 [INFO][4133] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.272 [INFO][4133] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.276 [INFO][4133] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.276 [INFO][4133] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.279 [INFO][4133] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.287 [INFO][4133] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.298 [INFO][4133] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.1/26] block=192.168.103.0/26 handle="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.298 [INFO][4133] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.1/26] handle="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.298 [INFO][4133] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:52.387057 containerd[1609]: 2025-10-31 02:44:52.298 [INFO][4133] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.1/26] IPv6=[] ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" HandleID="k8s-pod-network.cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.391682 containerd[1609]: 2025-10-31 02:44:52.303 [INFO][4119] cni-plugin/k8s.go 418: Populated endpoint ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"2c40f225-7c8b-464e-9dd2-615c17de403d", ResourceVersion:"938", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"calico-apiserver-589687cff4-ml2bp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia0a200d3e49", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:52.391682 containerd[1609]: 2025-10-31 02:44:52.305 [INFO][4119] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.1/32] ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.391682 containerd[1609]: 2025-10-31 02:44:52.305 [INFO][4119] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia0a200d3e49 ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.391682 containerd[1609]: 2025-10-31 02:44:52.325 [INFO][4119] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.391682 containerd[1609]: 2025-10-31 02:44:52.326 [INFO][4119] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"2c40f225-7c8b-464e-9dd2-615c17de403d", ResourceVersion:"938", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb", Pod:"calico-apiserver-589687cff4-ml2bp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia0a200d3e49", MAC:"7a:f8:d5:49:ae:a3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:52.391682 containerd[1609]: 2025-10-31 02:44:52.373 [INFO][4119] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-ml2bp" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:44:52.528869 containerd[1609]: time="2025-10-31T02:44:52.528632327Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:52.530050 containerd[1609]: time="2025-10-31T02:44:52.529086332Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:52.530050 containerd[1609]: time="2025-10-31T02:44:52.529316086Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:52.537045 containerd[1609]: time="2025-10-31T02:44:52.535430638Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:52.543168 containerd[1609]: time="2025-10-31T02:44:52.542823500Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d659c4594-7qhbb,Uid:ec0f842a-be78-41a7-a220-2e129c052a42,Namespace:calico-system,Attempt:0,}" Oct 31 02:44:52.575941 containerd[1609]: time="2025-10-31T02:44:52.574223018Z" level=info msg="StopPodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\"" Oct 31 02:44:52.577498 containerd[1609]: time="2025-10-31T02:44:52.576439391Z" level=info msg="StopPodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\"" Oct 31 02:44:52.825279 containerd[1609]: time="2025-10-31T02:44:52.824988208Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-ml2bp,Uid:2c40f225-7c8b-464e-9dd2-615c17de403d,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb\"" Oct 31 02:44:52.837071 containerd[1609]: time="2025-10-31T02:44:52.836196572Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.762 [INFO][4204] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.764 [INFO][4204] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" iface="eth0" netns="/var/run/netns/cni-41e3c612-5899-49e0-8fdd-666be4f1d730" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.765 [INFO][4204] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" iface="eth0" netns="/var/run/netns/cni-41e3c612-5899-49e0-8fdd-666be4f1d730" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.771 [INFO][4204] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" iface="eth0" netns="/var/run/netns/cni-41e3c612-5899-49e0-8fdd-666be4f1d730" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.771 [INFO][4204] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.771 [INFO][4204] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.842 [INFO][4223] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.844 [INFO][4223] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.844 [INFO][4223] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.861 [WARNING][4223] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.862 [INFO][4223] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.865 [INFO][4223] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:52.871886 containerd[1609]: 2025-10-31 02:44:52.868 [INFO][4204] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:44:52.874159 containerd[1609]: time="2025-10-31T02:44:52.872080428Z" level=info msg="TearDown network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" successfully" Oct 31 02:44:52.874159 containerd[1609]: time="2025-10-31T02:44:52.872145111Z" level=info msg="StopPodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" returns successfully" Oct 31 02:44:52.874280 containerd[1609]: time="2025-10-31T02:44:52.874243219Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-75444f6b5f-gp7s4,Uid:51cf0923-33f8-4c41-8a1b-075035cdaba9,Namespace:calico-apiserver,Attempt:1,}" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.829 [INFO][4205] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.829 [INFO][4205] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" iface="eth0" netns="/var/run/netns/cni-9cd9d574-452a-1104-eea9-152f86d08a02" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.831 [INFO][4205] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" iface="eth0" netns="/var/run/netns/cni-9cd9d574-452a-1104-eea9-152f86d08a02" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.834 [INFO][4205] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" iface="eth0" netns="/var/run/netns/cni-9cd9d574-452a-1104-eea9-152f86d08a02" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.834 [INFO][4205] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.834 [INFO][4205] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.894 [INFO][4242] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.894 [INFO][4242] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.894 [INFO][4242] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.907 [WARNING][4242] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.907 [INFO][4242] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.911 [INFO][4242] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:52.954064 containerd[1609]: 2025-10-31 02:44:52.936 [INFO][4205] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:52.954064 containerd[1609]: time="2025-10-31T02:44:52.952714780Z" level=info msg="TearDown network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" successfully" Oct 31 02:44:52.954064 containerd[1609]: time="2025-10-31T02:44:52.952773927Z" level=info msg="StopPodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" returns successfully" Oct 31 02:44:52.954064 containerd[1609]: time="2025-10-31T02:44:52.953811445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-5xd2q,Uid:1d3e6a01-de45-45a9-bd90-9cad31c5290e,Namespace:calico-apiserver,Attempt:1,}" Oct 31 02:44:53.088697 systemd-networkd[1257]: calicdc4ef52b6e: Link UP Oct 31 02:44:53.090139 systemd-networkd[1257]: calicdc4ef52b6e: Gained carrier Oct 31 02:44:53.114814 systemd[1]: run-netns-cni\x2d41e3c612\x2d5899\x2d49e0\x2d8fdd\x2d666be4f1d730.mount: Deactivated successfully. Oct 31 02:44:53.115244 systemd[1]: run-netns-cni\x2d9cd9d574\x2d452a\x2d1104\x2deea9\x2d152f86d08a02.mount: Deactivated successfully. Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.721 [INFO][4193] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.770 [INFO][4193] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0 whisker-5d659c4594- calico-system ec0f842a-be78-41a7-a220-2e129c052a42 955 0 2025-10-31 02:44:52 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5d659c4594 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com whisker-5d659c4594-7qhbb eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calicdc4ef52b6e [] [] }} ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.770 [INFO][4193] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.912 [INFO][4236] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" HandleID="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.912 [INFO][4236] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" HandleID="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00033db20), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-4lokn.gb1.brightbox.com", "pod":"whisker-5d659c4594-7qhbb", "timestamp":"2025-10-31 02:44:52.912120664 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.912 [INFO][4236] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.912 [INFO][4236] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.912 [INFO][4236] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.957 [INFO][4236] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.971 [INFO][4236] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:52.983 [INFO][4236] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.009 [INFO][4236] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.021 [INFO][4236] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.022 [INFO][4236] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.029 [INFO][4236] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092 Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.039 [INFO][4236] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.050 [INFO][4236] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.2/26] block=192.168.103.0/26 handle="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.050 [INFO][4236] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.2/26] handle="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.050 [INFO][4236] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:53.143319 containerd[1609]: 2025-10-31 02:44:53.050 [INFO][4236] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.2/26] IPv6=[] ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" HandleID="k8s-pod-network.e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.144444 containerd[1609]: 2025-10-31 02:44:53.071 [INFO][4193] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0", GenerateName:"whisker-5d659c4594-", Namespace:"calico-system", SelfLink:"", UID:"ec0f842a-be78-41a7-a220-2e129c052a42", ResourceVersion:"955", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5d659c4594", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"whisker-5d659c4594-7qhbb", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.103.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calicdc4ef52b6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:53.144444 containerd[1609]: 2025-10-31 02:44:53.071 [INFO][4193] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.2/32] ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.144444 containerd[1609]: 2025-10-31 02:44:53.071 [INFO][4193] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicdc4ef52b6e ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.144444 containerd[1609]: 2025-10-31 02:44:53.097 [INFO][4193] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.144444 containerd[1609]: 2025-10-31 02:44:53.107 [INFO][4193] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0", GenerateName:"whisker-5d659c4594-", Namespace:"calico-system", SelfLink:"", UID:"ec0f842a-be78-41a7-a220-2e129c052a42", ResourceVersion:"955", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5d659c4594", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092", Pod:"whisker-5d659c4594-7qhbb", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.103.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calicdc4ef52b6e", MAC:"fa:fa:96:b8:09:20", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:53.144444 containerd[1609]: 2025-10-31 02:44:53.126 [INFO][4193] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092" Namespace="calico-system" Pod="whisker-5d659c4594-7qhbb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--5d659c4594--7qhbb-eth0" Oct 31 02:44:53.169059 containerd[1609]: time="2025-10-31T02:44:53.167539123Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:53.236228 kubelet[2864]: I1031 02:44:53.233973 2864 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 31 02:44:53.270054 containerd[1609]: time="2025-10-31T02:44:53.168929604Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:44:53.270054 containerd[1609]: time="2025-10-31T02:44:53.169677715Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:44:53.270325 kubelet[2864]: E1031 02:44:53.269797 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:44:53.270325 kubelet[2864]: E1031 02:44:53.269887 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:44:53.294939 containerd[1609]: time="2025-10-31T02:44:53.292126207Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:53.294939 containerd[1609]: time="2025-10-31T02:44:53.292238817Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:53.294939 containerd[1609]: time="2025-10-31T02:44:53.292257893Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:53.294939 containerd[1609]: time="2025-10-31T02:44:53.292414819Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:53.295629 kubelet[2864]: E1031 02:44:53.290361 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tffjx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-ml2bp_calico-apiserver(2c40f225-7c8b-464e-9dd2-615c17de403d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:53.295629 kubelet[2864]: E1031 02:44:53.291811 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:44:53.502621 systemd-networkd[1257]: calia0a200d3e49: Gained IPv6LL Oct 31 02:44:53.533670 systemd-networkd[1257]: cali673c8bf8054: Link UP Oct 31 02:44:53.535934 systemd-networkd[1257]: cali673c8bf8054: Gained carrier Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.036 [INFO][4249] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.066 [INFO][4249] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0 calico-apiserver-75444f6b5f- calico-apiserver 51cf0923-33f8-4c41-8a1b-075035cdaba9 962 0 2025-10-31 02:44:20 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:75444f6b5f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com calico-apiserver-75444f6b5f-gp7s4 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali673c8bf8054 [] [] }} ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.066 [INFO][4249] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.360 [INFO][4309] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" HandleID="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.369 [INFO][4309] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" HandleID="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000369d40), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"srv-4lokn.gb1.brightbox.com", "pod":"calico-apiserver-75444f6b5f-gp7s4", "timestamp":"2025-10-31 02:44:53.360007903 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.369 [INFO][4309] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.369 [INFO][4309] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.369 [INFO][4309] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.411 [INFO][4309] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.425 [INFO][4309] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.445 [INFO][4309] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.450 [INFO][4309] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.457 [INFO][4309] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.457 [INFO][4309] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.465 [INFO][4309] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.491 [INFO][4309] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.509 [INFO][4309] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.3/26] block=192.168.103.0/26 handle="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.509 [INFO][4309] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.3/26] handle="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.509 [INFO][4309] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:53.564790 containerd[1609]: 2025-10-31 02:44:53.509 [INFO][4309] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.3/26] IPv6=[] ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" HandleID="k8s-pod-network.bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.572434 containerd[1609]: 2025-10-31 02:44:53.520 [INFO][4249] cni-plugin/k8s.go 418: Populated endpoint ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0", GenerateName:"calico-apiserver-75444f6b5f-", Namespace:"calico-apiserver", SelfLink:"", UID:"51cf0923-33f8-4c41-8a1b-075035cdaba9", ResourceVersion:"962", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"75444f6b5f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"calico-apiserver-75444f6b5f-gp7s4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali673c8bf8054", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:53.572434 containerd[1609]: 2025-10-31 02:44:53.525 [INFO][4249] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.3/32] ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.572434 containerd[1609]: 2025-10-31 02:44:53.526 [INFO][4249] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali673c8bf8054 ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.572434 containerd[1609]: 2025-10-31 02:44:53.535 [INFO][4249] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.572434 containerd[1609]: 2025-10-31 02:44:53.536 [INFO][4249] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0", GenerateName:"calico-apiserver-75444f6b5f-", Namespace:"calico-apiserver", SelfLink:"", UID:"51cf0923-33f8-4c41-8a1b-075035cdaba9", ResourceVersion:"962", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"75444f6b5f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f", Pod:"calico-apiserver-75444f6b5f-gp7s4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali673c8bf8054", MAC:"f2:ce:51:7b:dc:0d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:53.572434 containerd[1609]: 2025-10-31 02:44:53.556 [INFO][4249] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f" Namespace="calico-apiserver" Pod="calico-apiserver-75444f6b5f-gp7s4" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:44:53.601378 kubelet[2864]: I1031 02:44:53.601301 2864 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="868ae699-bd25-491c-a828-139c5285406d" path="/var/lib/kubelet/pods/868ae699-bd25-491c-a828-139c5285406d/volumes" Oct 31 02:44:53.616826 containerd[1609]: time="2025-10-31T02:44:53.616504269Z" level=info msg="StopPodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\"" Oct 31 02:44:53.620283 containerd[1609]: time="2025-10-31T02:44:53.620236347Z" level=info msg="StopPodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\"" Oct 31 02:44:53.630156 containerd[1609]: time="2025-10-31T02:44:53.629311001Z" level=info msg="StopPodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\"" Oct 31 02:44:53.733597 systemd-networkd[1257]: cali29865f6d492: Link UP Oct 31 02:44:53.749429 systemd-networkd[1257]: cali29865f6d492: Gained carrier Oct 31 02:44:53.861886 containerd[1609]: time="2025-10-31T02:44:53.861424719Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5d659c4594-7qhbb,Uid:ec0f842a-be78-41a7-a220-2e129c052a42,Namespace:calico-system,Attempt:0,} returns sandbox id \"e238f2e44fabdad1861eadc035f12b3edbb1567841d2d02357baeba6f99b0092\"" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.127 [INFO][4291] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.164 [INFO][4291] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0 calico-apiserver-589687cff4- calico-apiserver 1d3e6a01-de45-45a9-bd90-9cad31c5290e 963 0 2025-10-31 02:44:19 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:589687cff4 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com calico-apiserver-589687cff4-5xd2q eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali29865f6d492 [] [] }} ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.164 [INFO][4291] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.456 [INFO][4330] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" HandleID="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.460 [INFO][4330] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" HandleID="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5b50), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"srv-4lokn.gb1.brightbox.com", "pod":"calico-apiserver-589687cff4-5xd2q", "timestamp":"2025-10-31 02:44:53.448013255 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.469 [INFO][4330] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.510 [INFO][4330] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.511 [INFO][4330] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.540 [INFO][4330] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.556 [INFO][4330] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.582 [INFO][4330] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.590 [INFO][4330] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.601 [INFO][4330] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.603 [INFO][4330] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.620 [INFO][4330] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5 Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.640 [INFO][4330] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.672 [INFO][4330] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.4/26] block=192.168.103.0/26 handle="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.672 [INFO][4330] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.4/26] handle="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.672 [INFO][4330] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:53.871651 containerd[1609]: 2025-10-31 02:44:53.673 [INFO][4330] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.4/26] IPv6=[] ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" HandleID="k8s-pod-network.cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.873941 containerd[1609]: 2025-10-31 02:44:53.709 [INFO][4291] cni-plugin/k8s.go 418: Populated endpoint ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1d3e6a01-de45-45a9-bd90-9cad31c5290e", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"calico-apiserver-589687cff4-5xd2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali29865f6d492", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:53.873941 containerd[1609]: 2025-10-31 02:44:53.709 [INFO][4291] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.4/32] ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.873941 containerd[1609]: 2025-10-31 02:44:53.709 [INFO][4291] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali29865f6d492 ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.873941 containerd[1609]: 2025-10-31 02:44:53.772 [INFO][4291] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.873941 containerd[1609]: 2025-10-31 02:44:53.786 [INFO][4291] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1d3e6a01-de45-45a9-bd90-9cad31c5290e", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5", Pod:"calico-apiserver-589687cff4-5xd2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali29865f6d492", MAC:"d6:98:5d:75:a7:44", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:53.873941 containerd[1609]: 2025-10-31 02:44:53.824 [INFO][4291] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5" Namespace="calico-apiserver" Pod="calico-apiserver-589687cff4-5xd2q" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:53.884244 containerd[1609]: time="2025-10-31T02:44:53.883244856Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 31 02:44:53.896713 containerd[1609]: time="2025-10-31T02:44:53.894267533Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:53.896713 containerd[1609]: time="2025-10-31T02:44:53.894528843Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:53.896713 containerd[1609]: time="2025-10-31T02:44:53.894557028Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:53.896713 containerd[1609]: time="2025-10-31T02:44:53.894844785Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:54.067831 kubelet[2864]: E1031 02:44:54.067778 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:44:54.079403 containerd[1609]: time="2025-10-31T02:44:54.075485473Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:54.079403 containerd[1609]: time="2025-10-31T02:44:54.075588865Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:54.079403 containerd[1609]: time="2025-10-31T02:44:54.075607715Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:54.079403 containerd[1609]: time="2025-10-31T02:44:54.075776822Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:54.213180 systemd[1]: run-containerd-runc-k8s.io-bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f-runc.7hBVXN.mount: Deactivated successfully. Oct 31 02:44:54.395813 containerd[1609]: time="2025-10-31T02:44:54.395560748Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:54.397042 containerd[1609]: time="2025-10-31T02:44:54.396985827Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 31 02:44:54.397728 containerd[1609]: time="2025-10-31T02:44:54.397084566Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 31 02:44:54.398303 kubelet[2864]: E1031 02:44:54.398131 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:44:54.401092 kubelet[2864]: E1031 02:44:54.399839 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:44:54.401092 kubelet[2864]: E1031 02:44:54.400148 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:bd677a2a2fa44c8a81a4028fc03bd754,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:53.983 [INFO][4439] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:53.983 [INFO][4439] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" iface="eth0" netns="/var/run/netns/cni-08da79d5-ef11-b679-f315-f3601d6f6efd" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:53.986 [INFO][4439] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" iface="eth0" netns="/var/run/netns/cni-08da79d5-ef11-b679-f315-f3601d6f6efd" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.001 [INFO][4439] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" iface="eth0" netns="/var/run/netns/cni-08da79d5-ef11-b679-f315-f3601d6f6efd" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.001 [INFO][4439] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.001 [INFO][4439] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.303 [INFO][4511] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.303 [INFO][4511] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.303 [INFO][4511] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.342 [WARNING][4511] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.342 [INFO][4511] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.356 [INFO][4511] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:54.409493 containerd[1609]: 2025-10-31 02:44:54.381 [INFO][4439] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:44:54.416045 containerd[1609]: time="2025-10-31T02:44:54.413438991Z" level=info msg="TearDown network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" successfully" Oct 31 02:44:54.416045 containerd[1609]: time="2025-10-31T02:44:54.413476941Z" level=info msg="StopPodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" returns successfully" Oct 31 02:44:54.416045 containerd[1609]: time="2025-10-31T02:44:54.415621134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rp8wt,Uid:e00b0c65-1a04-42fa-84a6-f5f0d45c983a,Namespace:kube-system,Attempt:1,}" Oct 31 02:44:54.419801 containerd[1609]: time="2025-10-31T02:44:54.418999099Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 31 02:44:54.420405 systemd[1]: run-netns-cni\x2d08da79d5\x2def11\x2db679\x2df315\x2df3601d6f6efd.mount: Deactivated successfully. Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.285 [INFO][4465] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.286 [INFO][4465] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" iface="eth0" netns="/var/run/netns/cni-68b45efb-af14-cbd2-0f92-e2d201cdd5bd" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.287 [INFO][4465] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" iface="eth0" netns="/var/run/netns/cni-68b45efb-af14-cbd2-0f92-e2d201cdd5bd" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.287 [INFO][4465] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" iface="eth0" netns="/var/run/netns/cni-68b45efb-af14-cbd2-0f92-e2d201cdd5bd" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.287 [INFO][4465] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.288 [INFO][4465] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.494 [INFO][4558] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.508 [INFO][4558] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.508 [INFO][4558] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.547 [WARNING][4558] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.548 [INFO][4558] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.551 [INFO][4558] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:54.589748 containerd[1609]: 2025-10-31 02:44:54.565 [INFO][4465] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:44:54.596916 containerd[1609]: time="2025-10-31T02:44:54.591802420Z" level=info msg="TearDown network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" successfully" Oct 31 02:44:54.596916 containerd[1609]: time="2025-10-31T02:44:54.592056073Z" level=info msg="StopPodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" returns successfully" Oct 31 02:44:54.596916 containerd[1609]: time="2025-10-31T02:44:54.596017374Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fhw82,Uid:f40f31f8-0b31-408d-8016-7a7232bfce65,Namespace:kube-system,Attempt:1,}" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.265 [INFO][4469] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.266 [INFO][4469] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" iface="eth0" netns="/var/run/netns/cni-693a208a-77cc-69b7-adec-eb732f1076e1" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.266 [INFO][4469] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" iface="eth0" netns="/var/run/netns/cni-693a208a-77cc-69b7-adec-eb732f1076e1" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.267 [INFO][4469] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" iface="eth0" netns="/var/run/netns/cni-693a208a-77cc-69b7-adec-eb732f1076e1" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.267 [INFO][4469] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.267 [INFO][4469] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.574 [INFO][4550] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.575 [INFO][4550] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.575 [INFO][4550] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.586 [WARNING][4550] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.586 [INFO][4550] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.591 [INFO][4550] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:54.622496 containerd[1609]: 2025-10-31 02:44:54.606 [INFO][4469] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:44:54.644313 containerd[1609]: time="2025-10-31T02:44:54.642935491Z" level=info msg="TearDown network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" successfully" Oct 31 02:44:54.644313 containerd[1609]: time="2025-10-31T02:44:54.642990265Z" level=info msg="StopPodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" returns successfully" Oct 31 02:44:54.649215 containerd[1609]: time="2025-10-31T02:44:54.646318752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-n7gvx,Uid:c44deaed-49c1-414b-a77a-e9b3a43a004d,Namespace:calico-system,Attempt:1,}" Oct 31 02:44:54.825855 containerd[1609]: time="2025-10-31T02:44:54.824916362Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-75444f6b5f-gp7s4,Uid:51cf0923-33f8-4c41-8a1b-075035cdaba9,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f\"" Oct 31 02:44:54.845265 containerd[1609]: time="2025-10-31T02:44:54.843764277Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-589687cff4-5xd2q,Uid:1d3e6a01-de45-45a9-bd90-9cad31c5290e,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5\"" Oct 31 02:44:54.847330 systemd-networkd[1257]: cali673c8bf8054: Gained IPv6LL Oct 31 02:44:54.921071 containerd[1609]: time="2025-10-31T02:44:54.913217577Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:54.986325 containerd[1609]: time="2025-10-31T02:44:54.985045066Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 31 02:44:54.994749 containerd[1609]: time="2025-10-31T02:44:54.993859759Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 31 02:44:54.995979 kubelet[2864]: E1031 02:44:54.994752 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:44:54.995979 kubelet[2864]: E1031 02:44:54.994930 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:44:55.000043 containerd[1609]: time="2025-10-31T02:44:54.999606779Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:44:55.000957 kubelet[2864]: E1031 02:44:55.000645 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:55.035360 kubelet[2864]: E1031 02:44:55.002363 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:44:55.037264 systemd-networkd[1257]: cali29865f6d492: Gained IPv6LL Oct 31 02:44:55.037789 systemd-networkd[1257]: calicdc4ef52b6e: Gained IPv6LL Oct 31 02:44:55.134781 systemd[1]: run-netns-cni\x2d693a208a\x2d77cc\x2d69b7\x2dadec\x2deb732f1076e1.mount: Deactivated successfully. Oct 31 02:44:55.136250 systemd[1]: run-netns-cni\x2d68b45efb\x2daf14\x2dcbd2\x2d0f92\x2de2d201cdd5bd.mount: Deactivated successfully. Oct 31 02:44:55.159377 kubelet[2864]: E1031 02:44:55.159006 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:44:55.303931 kernel: bpftool[4682]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Oct 31 02:44:55.423312 systemd-networkd[1257]: cali7dcbd8df3e4: Link UP Oct 31 02:44:55.425105 systemd-networkd[1257]: cali7dcbd8df3e4: Gained carrier Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:54.868 [INFO][4578] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.044 [INFO][4578] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0 coredns-668d6bf9bc- kube-system e00b0c65-1a04-42fa-84a6-f5f0d45c983a 987 0 2025-10-31 02:44:06 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com coredns-668d6bf9bc-rp8wt eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali7dcbd8df3e4 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.046 [INFO][4578] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.291 [INFO][4655] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" HandleID="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.304 [INFO][4655] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" HandleID="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000267ce0), Attrs:map[string]string{"namespace":"kube-system", "node":"srv-4lokn.gb1.brightbox.com", "pod":"coredns-668d6bf9bc-rp8wt", "timestamp":"2025-10-31 02:44:55.291842572 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.306 [INFO][4655] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.306 [INFO][4655] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.306 [INFO][4655] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.332 [INFO][4655] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.350 [INFO][4655] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.367 [INFO][4655] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.372 [INFO][4655] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.380 [INFO][4655] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.380 [INFO][4655] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.383 [INFO][4655] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636 Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.390 [INFO][4655] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.399 [INFO][4655] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.5/26] block=192.168.103.0/26 handle="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.399 [INFO][4655] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.5/26] handle="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.399 [INFO][4655] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:55.462696 containerd[1609]: 2025-10-31 02:44:55.400 [INFO][4655] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.5/26] IPv6=[] ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" HandleID="k8s-pod-network.4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.465774 containerd[1609]: 2025-10-31 02:44:55.409 [INFO][4578] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"e00b0c65-1a04-42fa-84a6-f5f0d45c983a", ResourceVersion:"987", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"coredns-668d6bf9bc-rp8wt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7dcbd8df3e4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:55.465774 containerd[1609]: 2025-10-31 02:44:55.409 [INFO][4578] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.5/32] ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.465774 containerd[1609]: 2025-10-31 02:44:55.409 [INFO][4578] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7dcbd8df3e4 ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.465774 containerd[1609]: 2025-10-31 02:44:55.425 [INFO][4578] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.465774 containerd[1609]: 2025-10-31 02:44:55.426 [INFO][4578] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"e00b0c65-1a04-42fa-84a6-f5f0d45c983a", ResourceVersion:"987", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636", Pod:"coredns-668d6bf9bc-rp8wt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7dcbd8df3e4", MAC:"52:bc:57:a7:8f:34", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:55.465774 containerd[1609]: 2025-10-31 02:44:55.445 [INFO][4578] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636" Namespace="kube-system" Pod="coredns-668d6bf9bc-rp8wt" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:44:55.468549 kubelet[2864]: I1031 02:44:55.467755 2864 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 31 02:44:55.558612 containerd[1609]: time="2025-10-31T02:44:55.556683366Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:55.558612 containerd[1609]: time="2025-10-31T02:44:55.556803340Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:55.558612 containerd[1609]: time="2025-10-31T02:44:55.556822365Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:55.558612 containerd[1609]: time="2025-10-31T02:44:55.557005837Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:55.592450 containerd[1609]: time="2025-10-31T02:44:55.591628903Z" level=info msg="StopPodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\"" Oct 31 02:44:55.619017 systemd-networkd[1257]: calie764cb6e456: Link UP Oct 31 02:44:55.621256 systemd-networkd[1257]: calie764cb6e456: Gained carrier Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:54.990 [INFO][4620] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.079 [INFO][4620] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0 goldmane-666569f655- calico-system c44deaed-49c1-414b-a77a-e9b3a43a004d 994 0 2025-10-31 02:44:23 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com goldmane-666569f655-n7gvx eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calie764cb6e456 [] [] }} ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.079 [INFO][4620] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.306 [INFO][4661] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" HandleID="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.306 [INFO][4661] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" HandleID="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000123d00), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-4lokn.gb1.brightbox.com", "pod":"goldmane-666569f655-n7gvx", "timestamp":"2025-10-31 02:44:55.306293869 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.308 [INFO][4661] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.400 [INFO][4661] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.400 [INFO][4661] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.435 [INFO][4661] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.457 [INFO][4661] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.480 [INFO][4661] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.486 [INFO][4661] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.492 [INFO][4661] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.492 [INFO][4661] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.497 [INFO][4661] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.511 [INFO][4661] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.528 [INFO][4661] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.6/26] block=192.168.103.0/26 handle="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.529 [INFO][4661] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.6/26] handle="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.530 [INFO][4661] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:55.687176 containerd[1609]: 2025-10-31 02:44:55.534 [INFO][4661] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.6/26] IPv6=[] ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" HandleID="k8s-pod-network.c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.692288 containerd[1609]: 2025-10-31 02:44:55.551 [INFO][4620] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"c44deaed-49c1-414b-a77a-e9b3a43a004d", ResourceVersion:"994", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"goldmane-666569f655-n7gvx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calie764cb6e456", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:55.692288 containerd[1609]: 2025-10-31 02:44:55.555 [INFO][4620] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.6/32] ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.692288 containerd[1609]: 2025-10-31 02:44:55.555 [INFO][4620] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie764cb6e456 ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.692288 containerd[1609]: 2025-10-31 02:44:55.623 [INFO][4620] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.692288 containerd[1609]: 2025-10-31 02:44:55.635 [INFO][4620] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"c44deaed-49c1-414b-a77a-e9b3a43a004d", ResourceVersion:"994", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c", Pod:"goldmane-666569f655-n7gvx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calie764cb6e456", MAC:"66:c0:57:26:09:b1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:55.692288 containerd[1609]: 2025-10-31 02:44:55.671 [INFO][4620] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c" Namespace="calico-system" Pod="goldmane-666569f655-n7gvx" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:44:55.821392 containerd[1609]: time="2025-10-31T02:44:55.820806902Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:55.821392 containerd[1609]: time="2025-10-31T02:44:55.820936580Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:55.821392 containerd[1609]: time="2025-10-31T02:44:55.820963401Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:55.821392 containerd[1609]: time="2025-10-31T02:44:55.821152138Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:55.830638 systemd-networkd[1257]: cali8b9dec24775: Link UP Oct 31 02:44:55.843671 systemd-networkd[1257]: cali8b9dec24775: Gained carrier Oct 31 02:44:55.870122 containerd[1609]: time="2025-10-31T02:44:55.869583530Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:55.876438 containerd[1609]: time="2025-10-31T02:44:55.875086222Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:44:55.876438 containerd[1609]: time="2025-10-31T02:44:55.875381784Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:44:55.877735 kubelet[2864]: E1031 02:44:55.875563 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:44:55.877735 kubelet[2864]: E1031 02:44:55.876826 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:44:55.877735 kubelet[2864]: E1031 02:44:55.877190 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krkfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-75444f6b5f-gp7s4_calico-apiserver(51cf0923-33f8-4c41-8a1b-075035cdaba9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:55.880808 containerd[1609]: time="2025-10-31T02:44:55.879912601Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:44:55.882269 kubelet[2864]: E1031 02:44:55.882097 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:54.966 [INFO][4590] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.081 [INFO][4590] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0 coredns-668d6bf9bc- kube-system f40f31f8-0b31-408d-8016-7a7232bfce65 995 0 2025-10-31 02:44:06 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com coredns-668d6bf9bc-fhw82 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali8b9dec24775 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.083 [INFO][4590] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.359 [INFO][4664] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" HandleID="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.359 [INFO][4664] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" HandleID="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000123280), Attrs:map[string]string{"namespace":"kube-system", "node":"srv-4lokn.gb1.brightbox.com", "pod":"coredns-668d6bf9bc-fhw82", "timestamp":"2025-10-31 02:44:55.359548022 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.359 [INFO][4664] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.530 [INFO][4664] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.534 [INFO][4664] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.558 [INFO][4664] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.609 [INFO][4664] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.642 [INFO][4664] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.649 [INFO][4664] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.671 [INFO][4664] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.672 [INFO][4664] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.681 [INFO][4664] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466 Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.694 [INFO][4664] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.712 [INFO][4664] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.7/26] block=192.168.103.0/26 handle="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.714 [INFO][4664] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.7/26] handle="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.716 [INFO][4664] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:55.916309 containerd[1609]: 2025-10-31 02:44:55.716 [INFO][4664] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.7/26] IPv6=[] ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" HandleID="k8s-pod-network.c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:55.924544 containerd[1609]: 2025-10-31 02:44:55.767 [INFO][4590] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"f40f31f8-0b31-408d-8016-7a7232bfce65", ResourceVersion:"995", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"coredns-668d6bf9bc-fhw82", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali8b9dec24775", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:55.924544 containerd[1609]: 2025-10-31 02:44:55.771 [INFO][4590] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.7/32] ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:55.924544 containerd[1609]: 2025-10-31 02:44:55.771 [INFO][4590] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8b9dec24775 ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:55.924544 containerd[1609]: 2025-10-31 02:44:55.857 [INFO][4590] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:55.924544 containerd[1609]: 2025-10-31 02:44:55.878 [INFO][4590] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"f40f31f8-0b31-408d-8016-7a7232bfce65", ResourceVersion:"995", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466", Pod:"coredns-668d6bf9bc-fhw82", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali8b9dec24775", MAC:"da:ae:02:75:9f:e9", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:55.924544 containerd[1609]: 2025-10-31 02:44:55.910 [INFO][4590] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466" Namespace="kube-system" Pod="coredns-668d6bf9bc-fhw82" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:44:56.132408 containerd[1609]: time="2025-10-31T02:44:56.132094577Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rp8wt,Uid:e00b0c65-1a04-42fa-84a6-f5f0d45c983a,Namespace:kube-system,Attempt:1,} returns sandbox id \"4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636\"" Oct 31 02:44:56.158128 containerd[1609]: time="2025-10-31T02:44:56.155473817Z" level=info msg="CreateContainer within sandbox \"4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 31 02:44:56.160807 containerd[1609]: time="2025-10-31T02:44:56.152756613Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:56.160807 containerd[1609]: time="2025-10-31T02:44:56.153322512Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:56.160807 containerd[1609]: time="2025-10-31T02:44:56.153352052Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:56.160807 containerd[1609]: time="2025-10-31T02:44:56.154147486Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:56.196062 kubelet[2864]: E1031 02:44:56.192422 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.124 [INFO][4742] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.127 [INFO][4742] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" iface="eth0" netns="/var/run/netns/cni-6fdbcec3-ddff-89b6-1d1d-d2eb4d4321f0" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.128 [INFO][4742] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" iface="eth0" netns="/var/run/netns/cni-6fdbcec3-ddff-89b6-1d1d-d2eb4d4321f0" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.130 [INFO][4742] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" iface="eth0" netns="/var/run/netns/cni-6fdbcec3-ddff-89b6-1d1d-d2eb4d4321f0" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.130 [INFO][4742] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.130 [INFO][4742] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.317 [INFO][4833] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.317 [INFO][4833] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.317 [INFO][4833] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.349 [WARNING][4833] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.349 [INFO][4833] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.353 [INFO][4833] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:56.380327 containerd[1609]: 2025-10-31 02:44:56.367 [INFO][4742] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:56.385590 containerd[1609]: time="2025-10-31T02:44:56.384159194Z" level=info msg="TearDown network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" successfully" Oct 31 02:44:56.385590 containerd[1609]: time="2025-10-31T02:44:56.384345581Z" level=info msg="StopPodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" returns successfully" Oct 31 02:44:56.387233 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2959792755.mount: Deactivated successfully. Oct 31 02:44:56.392326 containerd[1609]: time="2025-10-31T02:44:56.390388958Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74fbc74c8b-kb8vb,Uid:1e5f0f90-f710-4f9a-b757-2db9fe751f0e,Namespace:calico-system,Attempt:1,}" Oct 31 02:44:56.401354 containerd[1609]: time="2025-10-31T02:44:56.401153117Z" level=info msg="CreateContainer within sandbox \"4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"847cc6a7e3313928ba17f6424ad70b7221000342aa2494412fa441cd002ec178\"" Oct 31 02:44:56.405614 containerd[1609]: time="2025-10-31T02:44:56.404562303Z" level=info msg="StartContainer for \"847cc6a7e3313928ba17f6424ad70b7221000342aa2494412fa441cd002ec178\"" Oct 31 02:44:56.406980 containerd[1609]: time="2025-10-31T02:44:56.404673108Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:56.412530 containerd[1609]: time="2025-10-31T02:44:56.412099451Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:44:56.413244 containerd[1609]: time="2025-10-31T02:44:56.412202229Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:44:56.414636 kubelet[2864]: E1031 02:44:56.414272 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:44:56.414636 kubelet[2864]: E1031 02:44:56.414343 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:44:56.414636 kubelet[2864]: E1031 02:44:56.414562 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xqpd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-5xd2q_calico-apiserver(1d3e6a01-de45-45a9-bd90-9cad31c5290e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:56.417617 kubelet[2864]: E1031 02:44:56.416346 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:44:56.509180 systemd-networkd[1257]: cali7dcbd8df3e4: Gained IPv6LL Oct 31 02:44:56.564816 containerd[1609]: time="2025-10-31T02:44:56.564107580Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fhw82,Uid:f40f31f8-0b31-408d-8016-7a7232bfce65,Namespace:kube-system,Attempt:1,} returns sandbox id \"c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466\"" Oct 31 02:44:56.577047 containerd[1609]: time="2025-10-31T02:44:56.576625828Z" level=info msg="StartContainer for \"847cc6a7e3313928ba17f6424ad70b7221000342aa2494412fa441cd002ec178\" returns successfully" Oct 31 02:44:56.580162 containerd[1609]: time="2025-10-31T02:44:56.580131637Z" level=info msg="StopPodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\"" Oct 31 02:44:56.590345 containerd[1609]: time="2025-10-31T02:44:56.590082292Z" level=info msg="CreateContainer within sandbox \"c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 31 02:44:56.686408 containerd[1609]: time="2025-10-31T02:44:56.685887330Z" level=info msg="CreateContainer within sandbox \"c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"1398782229302a646396ebcc43cf1d77e6136dacb8e4a2b13afecfd18fd23d0b\"" Oct 31 02:44:56.695330 containerd[1609]: time="2025-10-31T02:44:56.692856176Z" level=info msg="StartContainer for \"1398782229302a646396ebcc43cf1d77e6136dacb8e4a2b13afecfd18fd23d0b\"" Oct 31 02:44:56.746980 containerd[1609]: time="2025-10-31T02:44:56.746218009Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-n7gvx,Uid:c44deaed-49c1-414b-a77a-e9b3a43a004d,Namespace:calico-system,Attempt:1,} returns sandbox id \"c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c\"" Oct 31 02:44:56.758406 containerd[1609]: time="2025-10-31T02:44:56.758268680Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 31 02:44:56.829522 systemd-networkd[1257]: calie764cb6e456: Gained IPv6LL Oct 31 02:44:57.065118 containerd[1609]: time="2025-10-31T02:44:57.064313327Z" level=info msg="StartContainer for \"1398782229302a646396ebcc43cf1d77e6136dacb8e4a2b13afecfd18fd23d0b\" returns successfully" Oct 31 02:44:57.089129 systemd-networkd[1257]: cali8b9dec24775: Gained IPv6LL Oct 31 02:44:57.133747 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3216763603.mount: Deactivated successfully. Oct 31 02:44:57.134042 systemd[1]: run-netns-cni\x2d6fdbcec3\x2dddff\x2d89b6\x2d1d1d\x2dd2eb4d4321f0.mount: Deactivated successfully. Oct 31 02:44:57.219207 containerd[1609]: time="2025-10-31T02:44:57.218890484Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:57.228828 containerd[1609]: time="2025-10-31T02:44:57.228669122Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 31 02:44:57.228828 containerd[1609]: time="2025-10-31T02:44:57.228807301Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 31 02:44:57.231715 kubelet[2864]: E1031 02:44:57.230947 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:44:57.231715 kubelet[2864]: E1031 02:44:57.231049 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:44:57.231715 kubelet[2864]: E1031 02:44:57.231268 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rx5f2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-n7gvx_calico-system(c44deaed-49c1-414b-a77a-e9b3a43a004d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:57.233298 kubelet[2864]: E1031 02:44:57.232850 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:44:57.271204 kubelet[2864]: I1031 02:44:57.270900 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-rp8wt" podStartSLOduration=51.270839352 podStartE2EDuration="51.270839352s" podCreationTimestamp="2025-10-31 02:44:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-31 02:44:57.264416631 +0000 UTC m=+57.932905319" watchObservedRunningTime="2025-10-31 02:44:57.270839352 +0000 UTC m=+57.939328034" Oct 31 02:44:57.281743 kubelet[2864]: E1031 02:44:57.281693 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:44:57.341306 kubelet[2864]: I1031 02:44:57.340678 2864 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-fhw82" podStartSLOduration=51.340660203 podStartE2EDuration="51.340660203s" podCreationTimestamp="2025-10-31 02:44:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-31 02:44:57.335165024 +0000 UTC m=+58.003653707" watchObservedRunningTime="2025-10-31 02:44:57.340660203 +0000 UTC m=+58.009148871" Oct 31 02:44:57.457240 systemd-networkd[1257]: vxlan.calico: Link UP Oct 31 02:44:57.457254 systemd-networkd[1257]: vxlan.calico: Gained carrier Oct 31 02:44:57.484458 systemd-networkd[1257]: cali4063d8e1d41: Link UP Oct 31 02:44:57.490061 systemd-networkd[1257]: cali4063d8e1d41: Gained carrier Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.059 [INFO][4939] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.062 [INFO][4939] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" iface="eth0" netns="/var/run/netns/cni-282cd2c0-35f3-2ecc-ca05-3d2ab5cc26f5" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.062 [INFO][4939] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" iface="eth0" netns="/var/run/netns/cni-282cd2c0-35f3-2ecc-ca05-3d2ab5cc26f5" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.064 [INFO][4939] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" iface="eth0" netns="/var/run/netns/cni-282cd2c0-35f3-2ecc-ca05-3d2ab5cc26f5" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.064 [INFO][4939] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.064 [INFO][4939] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.284 [INFO][4996] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.285 [INFO][4996] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.437 [INFO][4996] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.473 [WARNING][4996] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.473 [INFO][4996] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.477 [INFO][4996] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:57.505188 containerd[1609]: 2025-10-31 02:44:57.498 [INFO][4939] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:44:57.518134 containerd[1609]: time="2025-10-31T02:44:57.515067746Z" level=info msg="TearDown network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" successfully" Oct 31 02:44:57.518134 containerd[1609]: time="2025-10-31T02:44:57.515142341Z" level=info msg="StopPodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" returns successfully" Oct 31 02:44:57.523241 systemd[1]: run-netns-cni\x2d282cd2c0\x2d35f3\x2d2ecc\x2dca05\x2d3d2ab5cc26f5.mount: Deactivated successfully. Oct 31 02:44:57.526411 containerd[1609]: time="2025-10-31T02:44:57.523866969Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-d2znc,Uid:7293ccde-9a5c-4930-845b-a0c0b00c9ac3,Namespace:calico-system,Attempt:1,}" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:56.937 [INFO][4882] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0 calico-kube-controllers-74fbc74c8b- calico-system 1e5f0f90-f710-4f9a-b757-2db9fe751f0e 1030 0 2025-10-31 02:44:26 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:74fbc74c8b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com calico-kube-controllers-74fbc74c8b-kb8vb eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali4063d8e1d41 [] [] }} ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:56.938 [INFO][4882] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.176 [INFO][4982] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" HandleID="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.184 [INFO][4982] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" HandleID="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002b9690), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-4lokn.gb1.brightbox.com", "pod":"calico-kube-controllers-74fbc74c8b-kb8vb", "timestamp":"2025-10-31 02:44:57.176350057 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.184 [INFO][4982] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.184 [INFO][4982] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.185 [INFO][4982] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.237 [INFO][4982] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.265 [INFO][4982] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.319 [INFO][4982] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.329 [INFO][4982] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.369 [INFO][4982] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.369 [INFO][4982] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.380 [INFO][4982] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96 Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.417 [INFO][4982] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.433 [INFO][4982] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.8/26] block=192.168.103.0/26 handle="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.434 [INFO][4982] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.8/26] handle="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.435 [INFO][4982] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:57.536404 containerd[1609]: 2025-10-31 02:44:57.435 [INFO][4982] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.8/26] IPv6=[] ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" HandleID="k8s-pod-network.97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.540399 containerd[1609]: 2025-10-31 02:44:57.470 [INFO][4882] cni-plugin/k8s.go 418: Populated endpoint ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0", GenerateName:"calico-kube-controllers-74fbc74c8b-", Namespace:"calico-system", SelfLink:"", UID:"1e5f0f90-f710-4f9a-b757-2db9fe751f0e", ResourceVersion:"1030", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"74fbc74c8b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"calico-kube-controllers-74fbc74c8b-kb8vb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali4063d8e1d41", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:57.540399 containerd[1609]: 2025-10-31 02:44:57.471 [INFO][4882] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.8/32] ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.540399 containerd[1609]: 2025-10-31 02:44:57.473 [INFO][4882] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4063d8e1d41 ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.540399 containerd[1609]: 2025-10-31 02:44:57.481 [INFO][4882] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.540399 containerd[1609]: 2025-10-31 02:44:57.483 [INFO][4882] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0", GenerateName:"calico-kube-controllers-74fbc74c8b-", Namespace:"calico-system", SelfLink:"", UID:"1e5f0f90-f710-4f9a-b757-2db9fe751f0e", ResourceVersion:"1030", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"74fbc74c8b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96", Pod:"calico-kube-controllers-74fbc74c8b-kb8vb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali4063d8e1d41", MAC:"2a:34:63:f0:f4:56", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:57.540399 containerd[1609]: 2025-10-31 02:44:57.518 [INFO][4882] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96" Namespace="calico-system" Pod="calico-kube-controllers-74fbc74c8b-kb8vb" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:57.720485 containerd[1609]: time="2025-10-31T02:44:57.718762027Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:57.720485 containerd[1609]: time="2025-10-31T02:44:57.718888974Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:57.720485 containerd[1609]: time="2025-10-31T02:44:57.718910059Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:57.725243 containerd[1609]: time="2025-10-31T02:44:57.720089698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:57.937158 containerd[1609]: time="2025-10-31T02:44:57.936809685Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74fbc74c8b-kb8vb,Uid:1e5f0f90-f710-4f9a-b757-2db9fe751f0e,Namespace:calico-system,Attempt:1,} returns sandbox id \"97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96\"" Oct 31 02:44:57.957471 containerd[1609]: time="2025-10-31T02:44:57.952977146Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 31 02:44:58.196924 systemd-networkd[1257]: caliee63302a000: Link UP Oct 31 02:44:58.211071 systemd-networkd[1257]: caliee63302a000: Gained carrier Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:57.849 [INFO][5018] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0 csi-node-driver- calico-system 7293ccde-9a5c-4930-845b-a0c0b00c9ac3 1048 0 2025-10-31 02:44:26 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s srv-4lokn.gb1.brightbox.com csi-node-driver-d2znc eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] caliee63302a000 [] [] }} ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:57.850 [INFO][5018] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.097 [INFO][5074] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" HandleID="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.101 [INFO][5074] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" HandleID="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00034ca60), Attrs:map[string]string{"namespace":"calico-system", "node":"srv-4lokn.gb1.brightbox.com", "pod":"csi-node-driver-d2znc", "timestamp":"2025-10-31 02:44:58.097905 +0000 UTC"}, Hostname:"srv-4lokn.gb1.brightbox.com", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.101 [INFO][5074] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.101 [INFO][5074] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.101 [INFO][5074] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'srv-4lokn.gb1.brightbox.com' Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.123 [INFO][5074] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.132 [INFO][5074] ipam/ipam.go 394: Looking up existing affinities for host host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.145 [INFO][5074] ipam/ipam.go 511: Trying affinity for 192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.152 [INFO][5074] ipam/ipam.go 158: Attempting to load block cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.160 [INFO][5074] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.103.0/26 host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.160 [INFO][5074] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.103.0/26 handle="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.163 [INFO][5074] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19 Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.171 [INFO][5074] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.103.0/26 handle="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.181 [INFO][5074] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.103.9/26] block=192.168.103.0/26 handle="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.181 [INFO][5074] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.103.9/26] handle="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" host="srv-4lokn.gb1.brightbox.com" Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.182 [INFO][5074] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:58.248838 containerd[1609]: 2025-10-31 02:44:58.182 [INFO][5074] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.103.9/26] IPv6=[] ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" HandleID="k8s-pod-network.7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.255822 containerd[1609]: 2025-10-31 02:44:58.187 [INFO][5018] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7293ccde-9a5c-4930-845b-a0c0b00c9ac3", ResourceVersion:"1048", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"", Pod:"csi-node-driver-d2znc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.9/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliee63302a000", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:58.255822 containerd[1609]: 2025-10-31 02:44:58.187 [INFO][5018] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.103.9/32] ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.255822 containerd[1609]: 2025-10-31 02:44:58.187 [INFO][5018] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliee63302a000 ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.255822 containerd[1609]: 2025-10-31 02:44:58.207 [INFO][5018] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.255822 containerd[1609]: 2025-10-31 02:44:58.212 [INFO][5018] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7293ccde-9a5c-4930-845b-a0c0b00c9ac3", ResourceVersion:"1048", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19", Pod:"csi-node-driver-d2znc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.9/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliee63302a000", MAC:"56:a1:0e:98:80:8d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:58.255822 containerd[1609]: 2025-10-31 02:44:58.240 [INFO][5018] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19" Namespace="calico-system" Pod="csi-node-driver-d2znc" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:44:58.309350 kubelet[2864]: E1031 02:44:58.308818 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:44:58.335790 containerd[1609]: time="2025-10-31T02:44:58.333349558Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:58.338289 containerd[1609]: time="2025-10-31T02:44:58.337244117Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 31 02:44:58.338289 containerd[1609]: time="2025-10-31T02:44:58.337931089Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 31 02:44:58.340460 kubelet[2864]: E1031 02:44:58.338563 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:44:58.340460 kubelet[2864]: E1031 02:44:58.339118 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:44:58.342125 kubelet[2864]: E1031 02:44:58.340588 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-455vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-74fbc74c8b-kb8vb_calico-system(1e5f0f90-f710-4f9a-b757-2db9fe751f0e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:58.343801 kubelet[2864]: E1031 02:44:58.343114 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:44:58.368133 containerd[1609]: time="2025-10-31T02:44:58.367571330Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 31 02:44:58.368133 containerd[1609]: time="2025-10-31T02:44:58.367658620Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 31 02:44:58.368133 containerd[1609]: time="2025-10-31T02:44:58.367676843Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:58.369963 containerd[1609]: time="2025-10-31T02:44:58.368690954Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 31 02:44:58.505856 containerd[1609]: time="2025-10-31T02:44:58.505804986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-d2znc,Uid:7293ccde-9a5c-4930-845b-a0c0b00c9ac3,Namespace:calico-system,Attempt:1,} returns sandbox id \"7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19\"" Oct 31 02:44:58.510969 containerd[1609]: time="2025-10-31T02:44:58.510903010Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 31 02:44:58.818264 containerd[1609]: time="2025-10-31T02:44:58.818050838Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:58.821136 containerd[1609]: time="2025-10-31T02:44:58.819796813Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 31 02:44:58.821136 containerd[1609]: time="2025-10-31T02:44:58.819912181Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 31 02:44:58.821354 kubelet[2864]: E1031 02:44:58.820345 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:44:58.821354 kubelet[2864]: E1031 02:44:58.820446 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:44:58.821354 kubelet[2864]: E1031 02:44:58.820668 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:58.824112 containerd[1609]: time="2025-10-31T02:44:58.824081934Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 31 02:44:59.111635 systemd[1]: run-containerd-runc-k8s.io-7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19-runc.xbZUif.mount: Deactivated successfully. Oct 31 02:44:59.141503 containerd[1609]: time="2025-10-31T02:44:59.141406888Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:44:59.142938 containerd[1609]: time="2025-10-31T02:44:59.142887189Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 31 02:44:59.143104 containerd[1609]: time="2025-10-31T02:44:59.143001269Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 31 02:44:59.143405 kubelet[2864]: E1031 02:44:59.143328 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:44:59.144170 kubelet[2864]: E1031 02:44:59.143398 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:44:59.144170 kubelet[2864]: E1031 02:44:59.143595 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 31 02:44:59.145505 kubelet[2864]: E1031 02:44:59.145198 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:59.261676 systemd-networkd[1257]: vxlan.calico: Gained IPv6LL Oct 31 02:44:59.304136 kubelet[2864]: E1031 02:44:59.304072 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:44:59.306726 kubelet[2864]: E1031 02:44:59.306646 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:44:59.453582 systemd-networkd[1257]: cali4063d8e1d41: Gained IPv6LL Oct 31 02:44:59.549672 containerd[1609]: time="2025-10-31T02:44:59.549430266Z" level=info msg="StopPodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\"" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.621 [WARNING][5239] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1d3e6a01-de45-45a9-bd90-9cad31c5290e", ResourceVersion:"1068", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5", Pod:"calico-apiserver-589687cff4-5xd2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali29865f6d492", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.621 [INFO][5239] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.621 [INFO][5239] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" iface="eth0" netns="" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.621 [INFO][5239] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.621 [INFO][5239] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.672 [INFO][5248] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.673 [INFO][5248] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.673 [INFO][5248] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.687 [WARNING][5248] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.687 [INFO][5248] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.689 [INFO][5248] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:59.698400 containerd[1609]: 2025-10-31 02:44:59.693 [INFO][5239] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.698400 containerd[1609]: time="2025-10-31T02:44:59.698221372Z" level=info msg="TearDown network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" successfully" Oct 31 02:44:59.698400 containerd[1609]: time="2025-10-31T02:44:59.698257184Z" level=info msg="StopPodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" returns successfully" Oct 31 02:44:59.702772 containerd[1609]: time="2025-10-31T02:44:59.702303108Z" level=info msg="RemovePodSandbox for \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\"" Oct 31 02:44:59.702772 containerd[1609]: time="2025-10-31T02:44:59.702377622Z" level=info msg="Forcibly stopping sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\"" Oct 31 02:44:59.773362 systemd-networkd[1257]: caliee63302a000: Gained IPv6LL Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.761 [WARNING][5262] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"1d3e6a01-de45-45a9-bd90-9cad31c5290e", ResourceVersion:"1068", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"cc022e0842d2d23bf3a2f7a992ac1918ec988b240849f8c0967e0e0e790fb6b5", Pod:"calico-apiserver-589687cff4-5xd2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali29865f6d492", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.762 [INFO][5262] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.762 [INFO][5262] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" iface="eth0" netns="" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.762 [INFO][5262] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.762 [INFO][5262] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.803 [INFO][5269] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.804 [INFO][5269] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.804 [INFO][5269] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.814 [WARNING][5269] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.814 [INFO][5269] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" HandleID="k8s-pod-network.924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--5xd2q-eth0" Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.817 [INFO][5269] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:59.823144 containerd[1609]: 2025-10-31 02:44:59.819 [INFO][5262] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149" Oct 31 02:44:59.823144 containerd[1609]: time="2025-10-31T02:44:59.821427073Z" level=info msg="TearDown network for sandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" successfully" Oct 31 02:44:59.833792 containerd[1609]: time="2025-10-31T02:44:59.833747657Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:44:59.834004 containerd[1609]: time="2025-10-31T02:44:59.833973033Z" level=info msg="RemovePodSandbox \"924202060b2680eab5c06d9151acf505521da8cfd39795cefb2905b724fcb149\" returns successfully" Oct 31 02:44:59.834840 containerd[1609]: time="2025-10-31T02:44:59.834809875Z" level=info msg="StopPodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\"" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.892 [WARNING][5284] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0", GenerateName:"calico-kube-controllers-74fbc74c8b-", Namespace:"calico-system", SelfLink:"", UID:"1e5f0f90-f710-4f9a-b757-2db9fe751f0e", ResourceVersion:"1105", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"74fbc74c8b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96", Pod:"calico-kube-controllers-74fbc74c8b-kb8vb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali4063d8e1d41", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.892 [INFO][5284] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.892 [INFO][5284] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" iface="eth0" netns="" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.892 [INFO][5284] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.892 [INFO][5284] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.932 [INFO][5291] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.932 [INFO][5291] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.932 [INFO][5291] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.944 [WARNING][5291] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.944 [INFO][5291] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.947 [INFO][5291] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:44:59.954496 containerd[1609]: 2025-10-31 02:44:59.952 [INFO][5284] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:44:59.955954 containerd[1609]: time="2025-10-31T02:44:59.955110520Z" level=info msg="TearDown network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" successfully" Oct 31 02:44:59.955954 containerd[1609]: time="2025-10-31T02:44:59.955262111Z" level=info msg="StopPodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" returns successfully" Oct 31 02:44:59.959801 containerd[1609]: time="2025-10-31T02:44:59.958378398Z" level=info msg="RemovePodSandbox for \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\"" Oct 31 02:44:59.959801 containerd[1609]: time="2025-10-31T02:44:59.958551851Z" level=info msg="Forcibly stopping sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\"" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.013 [WARNING][5305] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0", GenerateName:"calico-kube-controllers-74fbc74c8b-", Namespace:"calico-system", SelfLink:"", UID:"1e5f0f90-f710-4f9a-b757-2db9fe751f0e", ResourceVersion:"1105", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"74fbc74c8b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"97b32ca4fb943528fadaf7d64031023ad32449bd9568338dc869acd6a4962c96", Pod:"calico-kube-controllers-74fbc74c8b-kb8vb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.103.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali4063d8e1d41", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.014 [INFO][5305] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.014 [INFO][5305] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" iface="eth0" netns="" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.014 [INFO][5305] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.014 [INFO][5305] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.053 [INFO][5312] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.053 [INFO][5312] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.054 [INFO][5312] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.066 [WARNING][5312] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.066 [INFO][5312] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" HandleID="k8s-pod-network.08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--kube--controllers--74fbc74c8b--kb8vb-eth0" Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.069 [INFO][5312] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.077317 containerd[1609]: 2025-10-31 02:45:00.072 [INFO][5305] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba" Oct 31 02:45:00.079931 containerd[1609]: time="2025-10-31T02:45:00.078160569Z" level=info msg="TearDown network for sandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" successfully" Oct 31 02:45:00.083453 containerd[1609]: time="2025-10-31T02:45:00.083411130Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:00.083644 containerd[1609]: time="2025-10-31T02:45:00.083614725Z" level=info msg="RemovePodSandbox \"08a5ed3b34ae611fe9511f51bd79707451cad67f1484c6b9bdb877f83528b3ba\" returns successfully" Oct 31 02:45:00.084847 containerd[1609]: time="2025-10-31T02:45:00.084798926Z" level=info msg="StopPodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\"" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.141 [WARNING][5326] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"e00b0c65-1a04-42fa-84a6-f5f0d45c983a", ResourceVersion:"1056", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636", Pod:"coredns-668d6bf9bc-rp8wt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7dcbd8df3e4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.142 [INFO][5326] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.142 [INFO][5326] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" iface="eth0" netns="" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.142 [INFO][5326] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.142 [INFO][5326] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.177 [INFO][5333] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.178 [INFO][5333] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.178 [INFO][5333] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.187 [WARNING][5333] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.188 [INFO][5333] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.190 [INFO][5333] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.195187 containerd[1609]: 2025-10-31 02:45:00.192 [INFO][5326] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.197769 containerd[1609]: time="2025-10-31T02:45:00.195279864Z" level=info msg="TearDown network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" successfully" Oct 31 02:45:00.197769 containerd[1609]: time="2025-10-31T02:45:00.195334817Z" level=info msg="StopPodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" returns successfully" Oct 31 02:45:00.197769 containerd[1609]: time="2025-10-31T02:45:00.196834458Z" level=info msg="RemovePodSandbox for \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\"" Oct 31 02:45:00.197769 containerd[1609]: time="2025-10-31T02:45:00.196892888Z" level=info msg="Forcibly stopping sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\"" Oct 31 02:45:00.334200 kubelet[2864]: E1031 02:45:00.331672 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.306 [WARNING][5347] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"e00b0c65-1a04-42fa-84a6-f5f0d45c983a", ResourceVersion:"1056", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"4a9a91b2a0ad2c40e02d43a8cb734ae26fe639bc57eec0f48ce35d6339626636", Pod:"coredns-668d6bf9bc-rp8wt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7dcbd8df3e4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.306 [INFO][5347] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.306 [INFO][5347] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" iface="eth0" netns="" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.306 [INFO][5347] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.306 [INFO][5347] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.370 [INFO][5356] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.373 [INFO][5356] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.373 [INFO][5356] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.382 [WARNING][5356] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.382 [INFO][5356] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" HandleID="k8s-pod-network.3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--rp8wt-eth0" Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.384 [INFO][5356] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.391066 containerd[1609]: 2025-10-31 02:45:00.387 [INFO][5347] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6" Oct 31 02:45:00.391066 containerd[1609]: time="2025-10-31T02:45:00.389295870Z" level=info msg="TearDown network for sandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" successfully" Oct 31 02:45:00.393557 containerd[1609]: time="2025-10-31T02:45:00.393501535Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:00.393735 containerd[1609]: time="2025-10-31T02:45:00.393705861Z" level=info msg="RemovePodSandbox \"3e5d44bda68612a1f0a2532bffc32c9a5d941253b93318cc500580abd90205a6\" returns successfully" Oct 31 02:45:00.394501 containerd[1609]: time="2025-10-31T02:45:00.394457111Z" level=info msg="StopPodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\"" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.448 [WARNING][5371] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0", GenerateName:"calico-apiserver-75444f6b5f-", Namespace:"calico-apiserver", SelfLink:"", UID:"51cf0923-33f8-4c41-8a1b-075035cdaba9", ResourceVersion:"1034", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"75444f6b5f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f", Pod:"calico-apiserver-75444f6b5f-gp7s4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali673c8bf8054", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.448 [INFO][5371] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.448 [INFO][5371] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" iface="eth0" netns="" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.448 [INFO][5371] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.448 [INFO][5371] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.491 [INFO][5378] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.491 [INFO][5378] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.491 [INFO][5378] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.511 [WARNING][5378] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.511 [INFO][5378] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.515 [INFO][5378] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.521187 containerd[1609]: 2025-10-31 02:45:00.519 [INFO][5371] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.522216 containerd[1609]: time="2025-10-31T02:45:00.522176613Z" level=info msg="TearDown network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" successfully" Oct 31 02:45:00.522336 containerd[1609]: time="2025-10-31T02:45:00.522310475Z" level=info msg="StopPodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" returns successfully" Oct 31 02:45:00.523827 containerd[1609]: time="2025-10-31T02:45:00.523791964Z" level=info msg="RemovePodSandbox for \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\"" Oct 31 02:45:00.523965 containerd[1609]: time="2025-10-31T02:45:00.523936572Z" level=info msg="Forcibly stopping sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\"" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.587 [WARNING][5392] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0", GenerateName:"calico-apiserver-75444f6b5f-", Namespace:"calico-apiserver", SelfLink:"", UID:"51cf0923-33f8-4c41-8a1b-075035cdaba9", ResourceVersion:"1034", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"75444f6b5f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"bd89d99ce83faef626fd34962b5867685964460fddb32864a809e40428f9bd9f", Pod:"calico-apiserver-75444f6b5f-gp7s4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali673c8bf8054", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.587 [INFO][5392] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.587 [INFO][5392] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" iface="eth0" netns="" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.587 [INFO][5392] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.587 [INFO][5392] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.653 [INFO][5399] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.654 [INFO][5399] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.654 [INFO][5399] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.671 [WARNING][5399] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.671 [INFO][5399] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" HandleID="k8s-pod-network.d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--75444f6b5f--gp7s4-eth0" Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.673 [INFO][5399] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.678383 containerd[1609]: 2025-10-31 02:45:00.676 [INFO][5392] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f" Oct 31 02:45:00.680528 containerd[1609]: time="2025-10-31T02:45:00.680450451Z" level=info msg="TearDown network for sandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" successfully" Oct 31 02:45:00.684540 containerd[1609]: time="2025-10-31T02:45:00.684439023Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:00.684540 containerd[1609]: time="2025-10-31T02:45:00.684518547Z" level=info msg="RemovePodSandbox \"d3852e5879e74f470dbdbb7cab17543e352a067f3638410fa3fe3e67bc5d124f\" returns successfully" Oct 31 02:45:00.685474 containerd[1609]: time="2025-10-31T02:45:00.685427205Z" level=info msg="StopPodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\"" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.752 [WARNING][5413] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"2c40f225-7c8b-464e-9dd2-615c17de403d", ResourceVersion:"990", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb", Pod:"calico-apiserver-589687cff4-ml2bp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia0a200d3e49", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.753 [INFO][5413] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.753 [INFO][5413] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" iface="eth0" netns="" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.753 [INFO][5413] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.753 [INFO][5413] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.803 [INFO][5420] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.805 [INFO][5420] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.805 [INFO][5420] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.816 [WARNING][5420] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.817 [INFO][5420] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.819 [INFO][5420] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.824208 containerd[1609]: 2025-10-31 02:45:00.821 [INFO][5413] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.825813 containerd[1609]: time="2025-10-31T02:45:00.824404551Z" level=info msg="TearDown network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" successfully" Oct 31 02:45:00.825813 containerd[1609]: time="2025-10-31T02:45:00.824462006Z" level=info msg="StopPodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" returns successfully" Oct 31 02:45:00.827066 containerd[1609]: time="2025-10-31T02:45:00.826303217Z" level=info msg="RemovePodSandbox for \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\"" Oct 31 02:45:00.827066 containerd[1609]: time="2025-10-31T02:45:00.826344064Z" level=info msg="Forcibly stopping sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\"" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.899 [WARNING][5434] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0", GenerateName:"calico-apiserver-589687cff4-", Namespace:"calico-apiserver", SelfLink:"", UID:"2c40f225-7c8b-464e-9dd2-615c17de403d", ResourceVersion:"990", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"589687cff4", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"cd3e9df9dba1a9824f334a047fa8285e384c874bc1a80e1833f877f7c3e652eb", Pod:"calico-apiserver-589687cff4-ml2bp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.103.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia0a200d3e49", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.900 [INFO][5434] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.900 [INFO][5434] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" iface="eth0" netns="" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.900 [INFO][5434] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.900 [INFO][5434] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.950 [INFO][5441] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.951 [INFO][5441] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.951 [INFO][5441] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.961 [WARNING][5441] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.961 [INFO][5441] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" HandleID="k8s-pod-network.ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Workload="srv--4lokn.gb1.brightbox.com-k8s-calico--apiserver--589687cff4--ml2bp-eth0" Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.963 [INFO][5441] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:00.971168 containerd[1609]: 2025-10-31 02:45:00.968 [INFO][5434] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad" Oct 31 02:45:00.971168 containerd[1609]: time="2025-10-31T02:45:00.970448354Z" level=info msg="TearDown network for sandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" successfully" Oct 31 02:45:00.979051 containerd[1609]: time="2025-10-31T02:45:00.978576345Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:00.979051 containerd[1609]: time="2025-10-31T02:45:00.978648627Z" level=info msg="RemovePodSandbox \"ade5b1781eecf579ba1a9c202c36bdc3e2f72e614d378d8093b1ddfff843b4ad\" returns successfully" Oct 31 02:45:00.979549 containerd[1609]: time="2025-10-31T02:45:00.979501921Z" level=info msg="StopPodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\"" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.047 [WARNING][5456] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7293ccde-9a5c-4930-845b-a0c0b00c9ac3", ResourceVersion:"1109", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19", Pod:"csi-node-driver-d2znc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.9/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliee63302a000", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.048 [INFO][5456] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.048 [INFO][5456] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" iface="eth0" netns="" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.048 [INFO][5456] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.048 [INFO][5456] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.083 [INFO][5464] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.083 [INFO][5464] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.083 [INFO][5464] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.092 [WARNING][5464] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.092 [INFO][5464] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.096 [INFO][5464] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.104143 containerd[1609]: 2025-10-31 02:45:01.100 [INFO][5456] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.104143 containerd[1609]: time="2025-10-31T02:45:01.103845018Z" level=info msg="TearDown network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" successfully" Oct 31 02:45:01.104143 containerd[1609]: time="2025-10-31T02:45:01.103903562Z" level=info msg="StopPodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" returns successfully" Oct 31 02:45:01.106775 containerd[1609]: time="2025-10-31T02:45:01.106710170Z" level=info msg="RemovePodSandbox for \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\"" Oct 31 02:45:01.107182 containerd[1609]: time="2025-10-31T02:45:01.106898627Z" level=info msg="Forcibly stopping sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\"" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.186 [WARNING][5478] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7293ccde-9a5c-4930-845b-a0c0b00c9ac3", ResourceVersion:"1109", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"7a8ec0570b72225b343ab41c8c74f3d6e992fcf5827b683d2c42e1703948cb19", Pod:"csi-node-driver-d2znc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.103.9/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliee63302a000", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.186 [INFO][5478] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.186 [INFO][5478] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" iface="eth0" netns="" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.186 [INFO][5478] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.186 [INFO][5478] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.231 [INFO][5485] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.231 [INFO][5485] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.231 [INFO][5485] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.240 [WARNING][5485] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.240 [INFO][5485] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" HandleID="k8s-pod-network.bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Workload="srv--4lokn.gb1.brightbox.com-k8s-csi--node--driver--d2znc-eth0" Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.242 [INFO][5485] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.249260 containerd[1609]: 2025-10-31 02:45:01.244 [INFO][5478] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646" Oct 31 02:45:01.252315 containerd[1609]: time="2025-10-31T02:45:01.250148658Z" level=info msg="TearDown network for sandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" successfully" Oct 31 02:45:01.254925 containerd[1609]: time="2025-10-31T02:45:01.254693480Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:01.254925 containerd[1609]: time="2025-10-31T02:45:01.254787370Z" level=info msg="RemovePodSandbox \"bf6d49055cb790a42c2fa425ad8abc3a4e9ed1a6c7e327115dca3d7ccb692646\" returns successfully" Oct 31 02:45:01.255568 containerd[1609]: time="2025-10-31T02:45:01.255451164Z" level=info msg="StopPodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\"" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.307 [WARNING][5499] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"f40f31f8-0b31-408d-8016-7a7232bfce65", ResourceVersion:"1061", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466", Pod:"coredns-668d6bf9bc-fhw82", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali8b9dec24775", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.308 [INFO][5499] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.308 [INFO][5499] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" iface="eth0" netns="" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.308 [INFO][5499] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.308 [INFO][5499] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.347 [INFO][5507] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.347 [INFO][5507] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.347 [INFO][5507] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.367 [WARNING][5507] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.367 [INFO][5507] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.372 [INFO][5507] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.383681 containerd[1609]: 2025-10-31 02:45:01.378 [INFO][5499] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.385751 containerd[1609]: time="2025-10-31T02:45:01.384109954Z" level=info msg="TearDown network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" successfully" Oct 31 02:45:01.385751 containerd[1609]: time="2025-10-31T02:45:01.384146954Z" level=info msg="StopPodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" returns successfully" Oct 31 02:45:01.385751 containerd[1609]: time="2025-10-31T02:45:01.385195294Z" level=info msg="RemovePodSandbox for \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\"" Oct 31 02:45:01.385751 containerd[1609]: time="2025-10-31T02:45:01.385233920Z" level=info msg="Forcibly stopping sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\"" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.456 [WARNING][5521] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"f40f31f8-0b31-408d-8016-7a7232bfce65", ResourceVersion:"1061", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"c5c755043564ae2eb9b8e3de9c61b1bb2757919acaa62350b296998cec1ed466", Pod:"coredns-668d6bf9bc-fhw82", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.103.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali8b9dec24775", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.456 [INFO][5521] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.456 [INFO][5521] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" iface="eth0" netns="" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.456 [INFO][5521] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.456 [INFO][5521] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.496 [INFO][5529] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.496 [INFO][5529] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.496 [INFO][5529] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.507 [WARNING][5529] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.507 [INFO][5529] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" HandleID="k8s-pod-network.59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Workload="srv--4lokn.gb1.brightbox.com-k8s-coredns--668d6bf9bc--fhw82-eth0" Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.509 [INFO][5529] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.513494 containerd[1609]: 2025-10-31 02:45:01.511 [INFO][5521] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f" Oct 31 02:45:01.513494 containerd[1609]: time="2025-10-31T02:45:01.513459283Z" level=info msg="TearDown network for sandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" successfully" Oct 31 02:45:01.519128 containerd[1609]: time="2025-10-31T02:45:01.519088707Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:01.519242 containerd[1609]: time="2025-10-31T02:45:01.519157414Z" level=info msg="RemovePodSandbox \"59d71ea30b433aa860d6796886e6679a45a0631afec3643b9c16eba74c85ca5f\" returns successfully" Oct 31 02:45:01.519868 containerd[1609]: time="2025-10-31T02:45:01.519832235Z" level=info msg="StopPodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\"" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.580 [WARNING][5543] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.581 [INFO][5543] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.581 [INFO][5543] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" iface="eth0" netns="" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.581 [INFO][5543] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.581 [INFO][5543] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.622 [INFO][5550] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.623 [INFO][5550] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.623 [INFO][5550] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.635 [WARNING][5550] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.635 [INFO][5550] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.636 [INFO][5550] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.641169 containerd[1609]: 2025-10-31 02:45:01.639 [INFO][5543] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.642368 containerd[1609]: time="2025-10-31T02:45:01.641249191Z" level=info msg="TearDown network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" successfully" Oct 31 02:45:01.642368 containerd[1609]: time="2025-10-31T02:45:01.641287393Z" level=info msg="StopPodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" returns successfully" Oct 31 02:45:01.642368 containerd[1609]: time="2025-10-31T02:45:01.642142700Z" level=info msg="RemovePodSandbox for \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\"" Oct 31 02:45:01.642368 containerd[1609]: time="2025-10-31T02:45:01.642178559Z" level=info msg="Forcibly stopping sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\"" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.705 [WARNING][5565] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" WorkloadEndpoint="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.705 [INFO][5565] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.705 [INFO][5565] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" iface="eth0" netns="" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.705 [INFO][5565] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.705 [INFO][5565] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.735 [INFO][5573] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.735 [INFO][5573] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.735 [INFO][5573] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.744 [WARNING][5573] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.744 [INFO][5573] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" HandleID="k8s-pod-network.af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Workload="srv--4lokn.gb1.brightbox.com-k8s-whisker--59b4654fb4--jbkzr-eth0" Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.746 [INFO][5573] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.750332 containerd[1609]: 2025-10-31 02:45:01.748 [INFO][5565] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf" Oct 31 02:45:01.753018 containerd[1609]: time="2025-10-31T02:45:01.750389932Z" level=info msg="TearDown network for sandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" successfully" Oct 31 02:45:01.754529 containerd[1609]: time="2025-10-31T02:45:01.754457301Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:01.754617 containerd[1609]: time="2025-10-31T02:45:01.754593451Z" level=info msg="RemovePodSandbox \"af4c589752f318adb9c6bd052fd945b55dabd2dcedae2d2d7c0c0005b64227cf\" returns successfully" Oct 31 02:45:01.755393 containerd[1609]: time="2025-10-31T02:45:01.755344456Z" level=info msg="StopPodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\"" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.810 [WARNING][5588] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"c44deaed-49c1-414b-a77a-e9b3a43a004d", ResourceVersion:"1083", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c", Pod:"goldmane-666569f655-n7gvx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calie764cb6e456", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.811 [INFO][5588] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.811 [INFO][5588] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" iface="eth0" netns="" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.811 [INFO][5588] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.811 [INFO][5588] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.841 [INFO][5595] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.841 [INFO][5595] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.841 [INFO][5595] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.857 [WARNING][5595] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.858 [INFO][5595] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.862 [INFO][5595] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:01.867012 containerd[1609]: 2025-10-31 02:45:01.864 [INFO][5588] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:01.868992 containerd[1609]: time="2025-10-31T02:45:01.867087389Z" level=info msg="TearDown network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" successfully" Oct 31 02:45:01.868992 containerd[1609]: time="2025-10-31T02:45:01.867125343Z" level=info msg="StopPodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" returns successfully" Oct 31 02:45:01.868992 containerd[1609]: time="2025-10-31T02:45:01.867872831Z" level=info msg="RemovePodSandbox for \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\"" Oct 31 02:45:01.868992 containerd[1609]: time="2025-10-31T02:45:01.867928828Z" level=info msg="Forcibly stopping sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\"" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.934 [WARNING][5609] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"c44deaed-49c1-414b-a77a-e9b3a43a004d", ResourceVersion:"1083", Generation:0, CreationTimestamp:time.Date(2025, time.October, 31, 2, 44, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"srv-4lokn.gb1.brightbox.com", ContainerID:"c4c72e94290bcf328662419ba1d54839918e00009c1af581c32a7dcb9154468c", Pod:"goldmane-666569f655-n7gvx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.103.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calie764cb6e456", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.934 [INFO][5609] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.934 [INFO][5609] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" iface="eth0" netns="" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.934 [INFO][5609] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.934 [INFO][5609] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.980 [INFO][5617] ipam/ipam_plugin.go 436: Releasing address using handleID ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.981 [INFO][5617] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.981 [INFO][5617] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.990 [WARNING][5617] ipam/ipam_plugin.go 453: Asked to release address but it doesn't exist. Ignoring ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.990 [INFO][5617] ipam/ipam_plugin.go 464: Releasing address using workloadID ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" HandleID="k8s-pod-network.22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Workload="srv--4lokn.gb1.brightbox.com-k8s-goldmane--666569f655--n7gvx-eth0" Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.992 [INFO][5617] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 31 02:45:02.000146 containerd[1609]: 2025-10-31 02:45:01.995 [INFO][5609] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04" Oct 31 02:45:02.002061 containerd[1609]: time="2025-10-31T02:45:02.000188384Z" level=info msg="TearDown network for sandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" successfully" Oct 31 02:45:02.003716 containerd[1609]: time="2025-10-31T02:45:02.003672750Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Oct 31 02:45:02.003784 containerd[1609]: time="2025-10-31T02:45:02.003741180Z" level=info msg="RemovePodSandbox \"22c5a803cb5855d6dbca846b4c881f54fee2c54ce386942ea279ca7ec8545b04\" returns successfully" Oct 31 02:45:05.577443 containerd[1609]: time="2025-10-31T02:45:05.577054037Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:45:05.908997 containerd[1609]: time="2025-10-31T02:45:05.908306902Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:05.910733 containerd[1609]: time="2025-10-31T02:45:05.910495823Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:45:05.910733 containerd[1609]: time="2025-10-31T02:45:05.910512784Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:05.911137 kubelet[2864]: E1031 02:45:05.911059 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:05.911708 kubelet[2864]: E1031 02:45:05.911174 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:05.911708 kubelet[2864]: E1031 02:45:05.911470 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tffjx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-ml2bp_calico-apiserver(2c40f225-7c8b-464e-9dd2-615c17de403d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:05.912780 kubelet[2864]: E1031 02:45:05.912723 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:45:08.575574 containerd[1609]: time="2025-10-31T02:45:08.574328099Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:45:08.902654 containerd[1609]: time="2025-10-31T02:45:08.902218778Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:08.903868 containerd[1609]: time="2025-10-31T02:45:08.903798356Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:45:08.904120 containerd[1609]: time="2025-10-31T02:45:08.903910270Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:08.904223 kubelet[2864]: E1031 02:45:08.904155 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:08.905372 kubelet[2864]: E1031 02:45:08.904239 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:08.905372 kubelet[2864]: E1031 02:45:08.904460 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krkfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-75444f6b5f-gp7s4_calico-apiserver(51cf0923-33f8-4c41-8a1b-075035cdaba9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:08.905907 kubelet[2864]: E1031 02:45:08.905616 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:45:09.579389 containerd[1609]: time="2025-10-31T02:45:09.579059369Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 31 02:45:09.888866 containerd[1609]: time="2025-10-31T02:45:09.888493067Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:09.890089 containerd[1609]: time="2025-10-31T02:45:09.889934639Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 31 02:45:09.890089 containerd[1609]: time="2025-10-31T02:45:09.890037181Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 31 02:45:09.890321 kubelet[2864]: E1031 02:45:09.890238 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:45:09.890321 kubelet[2864]: E1031 02:45:09.890311 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:45:09.890525 kubelet[2864]: E1031 02:45:09.890473 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:bd677a2a2fa44c8a81a4028fc03bd754,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:09.893444 containerd[1609]: time="2025-10-31T02:45:09.893345938Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 31 02:45:10.204213 containerd[1609]: time="2025-10-31T02:45:10.203975544Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:10.205393 containerd[1609]: time="2025-10-31T02:45:10.205227945Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 31 02:45:10.205393 containerd[1609]: time="2025-10-31T02:45:10.205292158Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 31 02:45:10.205684 kubelet[2864]: E1031 02:45:10.205599 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:45:10.206318 kubelet[2864]: E1031 02:45:10.205701 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:45:10.206318 kubelet[2864]: E1031 02:45:10.205926 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:10.207300 kubelet[2864]: E1031 02:45:10.207217 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:45:11.580164 containerd[1609]: time="2025-10-31T02:45:11.578902927Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:45:11.891862 containerd[1609]: time="2025-10-31T02:45:11.891656862Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:11.893181 containerd[1609]: time="2025-10-31T02:45:11.893111859Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:45:11.893376 containerd[1609]: time="2025-10-31T02:45:11.893144677Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:11.893518 kubelet[2864]: E1031 02:45:11.893456 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:11.893518 kubelet[2864]: E1031 02:45:11.893534 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:11.894277 kubelet[2864]: E1031 02:45:11.893906 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xqpd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-5xd2q_calico-apiserver(1d3e6a01-de45-45a9-bd90-9cad31c5290e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:11.895111 containerd[1609]: time="2025-10-31T02:45:11.894863964Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 31 02:45:11.895935 kubelet[2864]: E1031 02:45:11.895118 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:45:12.226482 containerd[1609]: time="2025-10-31T02:45:12.226285385Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:12.227941 containerd[1609]: time="2025-10-31T02:45:12.227855915Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 31 02:45:12.228064 containerd[1609]: time="2025-10-31T02:45:12.227979018Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 31 02:45:12.228413 kubelet[2864]: E1031 02:45:12.228330 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:45:12.228562 kubelet[2864]: E1031 02:45:12.228412 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:45:12.228679 kubelet[2864]: E1031 02:45:12.228621 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:12.231263 containerd[1609]: time="2025-10-31T02:45:12.231224284Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 31 02:45:12.539584 containerd[1609]: time="2025-10-31T02:45:12.539498900Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:12.540920 containerd[1609]: time="2025-10-31T02:45:12.540865795Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 31 02:45:12.541076 containerd[1609]: time="2025-10-31T02:45:12.540985679Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 31 02:45:12.541948 kubelet[2864]: E1031 02:45:12.541330 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:45:12.541948 kubelet[2864]: E1031 02:45:12.541483 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:45:12.541948 kubelet[2864]: E1031 02:45:12.541846 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:12.543615 kubelet[2864]: E1031 02:45:12.543516 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:45:12.575576 containerd[1609]: time="2025-10-31T02:45:12.575510744Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 31 02:45:12.899633 containerd[1609]: time="2025-10-31T02:45:12.899400028Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:12.901324 containerd[1609]: time="2025-10-31T02:45:12.901253627Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 31 02:45:12.901438 containerd[1609]: time="2025-10-31T02:45:12.901377531Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:12.902046 kubelet[2864]: E1031 02:45:12.901638 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:45:12.902046 kubelet[2864]: E1031 02:45:12.901709 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:45:12.902046 kubelet[2864]: E1031 02:45:12.901908 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rx5f2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-n7gvx_calico-system(c44deaed-49c1-414b-a77a-e9b3a43a004d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:12.903876 kubelet[2864]: E1031 02:45:12.903042 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:45:13.574852 containerd[1609]: time="2025-10-31T02:45:13.574480549Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 31 02:45:13.885320 containerd[1609]: time="2025-10-31T02:45:13.885130704Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:13.886486 containerd[1609]: time="2025-10-31T02:45:13.886430392Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 31 02:45:13.887337 containerd[1609]: time="2025-10-31T02:45:13.886461661Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 31 02:45:13.887428 kubelet[2864]: E1031 02:45:13.886942 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:45:13.887428 kubelet[2864]: E1031 02:45:13.887057 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:45:13.887428 kubelet[2864]: E1031 02:45:13.887245 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-455vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-74fbc74c8b-kb8vb_calico-system(1e5f0f90-f710-4f9a-b757-2db9fe751f0e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:13.888582 kubelet[2864]: E1031 02:45:13.888506 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:45:19.576392 kubelet[2864]: E1031 02:45:19.576280 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:45:20.574078 kubelet[2864]: E1031 02:45:20.573876 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:45:22.574111 kubelet[2864]: E1031 02:45:22.574053 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:45:24.574402 kubelet[2864]: E1031 02:45:24.574213 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:45:25.579869 kubelet[2864]: E1031 02:45:25.579775 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:45:25.582668 kubelet[2864]: E1031 02:45:25.580337 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:45:27.605604 kubelet[2864]: E1031 02:45:27.605462 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:45:31.581279 containerd[1609]: time="2025-10-31T02:45:31.579534739Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:45:31.923642 containerd[1609]: time="2025-10-31T02:45:31.923328226Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:31.924945 containerd[1609]: time="2025-10-31T02:45:31.924690887Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:45:31.924945 containerd[1609]: time="2025-10-31T02:45:31.924765417Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:31.925143 kubelet[2864]: E1031 02:45:31.925085 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:31.925725 kubelet[2864]: E1031 02:45:31.925179 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:31.925725 kubelet[2864]: E1031 02:45:31.925463 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tffjx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-ml2bp_calico-apiserver(2c40f225-7c8b-464e-9dd2-615c17de403d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:31.927160 kubelet[2864]: E1031 02:45:31.927097 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:45:34.147511 systemd[1]: Started sshd@9-10.230.25.230:22-147.75.109.163:53476.service - OpenSSH per-connection server daemon (147.75.109.163:53476). Oct 31 02:45:34.578384 containerd[1609]: time="2025-10-31T02:45:34.578122070Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:45:34.903932 containerd[1609]: time="2025-10-31T02:45:34.902281012Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:34.904176 containerd[1609]: time="2025-10-31T02:45:34.903979957Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:45:34.904305 containerd[1609]: time="2025-10-31T02:45:34.904184472Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:34.907044 kubelet[2864]: E1031 02:45:34.905337 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:34.907044 kubelet[2864]: E1031 02:45:34.905503 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:34.909343 containerd[1609]: time="2025-10-31T02:45:34.908476723Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:45:34.910343 kubelet[2864]: E1031 02:45:34.909855 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krkfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-75444f6b5f-gp7s4_calico-apiserver(51cf0923-33f8-4c41-8a1b-075035cdaba9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:34.915041 kubelet[2864]: E1031 02:45:34.913244 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:45:35.158451 sshd[5697]: Accepted publickey for core from 147.75.109.163 port 53476 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:45:35.163768 sshd[5697]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:45:35.197873 systemd-logind[1591]: New session 12 of user core. Oct 31 02:45:35.203507 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 31 02:45:35.243917 containerd[1609]: time="2025-10-31T02:45:35.243015068Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:35.246544 containerd[1609]: time="2025-10-31T02:45:35.246482201Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:45:35.246724 containerd[1609]: time="2025-10-31T02:45:35.246607178Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:35.248203 kubelet[2864]: E1031 02:45:35.247686 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:35.248203 kubelet[2864]: E1031 02:45:35.247768 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:45:35.249163 kubelet[2864]: E1031 02:45:35.247970 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xqpd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-5xd2q_calico-apiserver(1d3e6a01-de45-45a9-bd90-9cad31c5290e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:35.253218 kubelet[2864]: E1031 02:45:35.252093 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:45:35.578553 containerd[1609]: time="2025-10-31T02:45:35.578441452Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 31 02:45:35.901255 containerd[1609]: time="2025-10-31T02:45:35.899795467Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:35.905739 containerd[1609]: time="2025-10-31T02:45:35.901730493Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 31 02:45:35.905739 containerd[1609]: time="2025-10-31T02:45:35.901868410Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 31 02:45:35.905883 kubelet[2864]: E1031 02:45:35.903231 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:45:35.905883 kubelet[2864]: E1031 02:45:35.903312 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:45:35.905883 kubelet[2864]: E1031 02:45:35.903484 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:bd677a2a2fa44c8a81a4028fc03bd754,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:35.911587 containerd[1609]: time="2025-10-31T02:45:35.911348672Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 31 02:45:36.240890 containerd[1609]: time="2025-10-31T02:45:36.240532751Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:36.245045 containerd[1609]: time="2025-10-31T02:45:36.243310711Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 31 02:45:36.245045 containerd[1609]: time="2025-10-31T02:45:36.243418037Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 31 02:45:36.245223 kubelet[2864]: E1031 02:45:36.243696 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:45:36.245223 kubelet[2864]: E1031 02:45:36.243800 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:45:36.249454 kubelet[2864]: E1031 02:45:36.248083 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:36.251359 kubelet[2864]: E1031 02:45:36.251138 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:45:36.575713 containerd[1609]: time="2025-10-31T02:45:36.574906137Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 31 02:45:36.580510 sshd[5697]: pam_unix(sshd:session): session closed for user core Oct 31 02:45:36.600972 systemd[1]: sshd@9-10.230.25.230:22-147.75.109.163:53476.service: Deactivated successfully. Oct 31 02:45:36.611846 systemd-logind[1591]: Session 12 logged out. Waiting for processes to exit. Oct 31 02:45:36.613742 systemd[1]: session-12.scope: Deactivated successfully. Oct 31 02:45:36.626618 systemd-logind[1591]: Removed session 12. Oct 31 02:45:36.921724 containerd[1609]: time="2025-10-31T02:45:36.921334055Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:36.925076 containerd[1609]: time="2025-10-31T02:45:36.923651473Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 31 02:45:36.925076 containerd[1609]: time="2025-10-31T02:45:36.923821282Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 31 02:45:36.925486 kubelet[2864]: E1031 02:45:36.925207 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:45:36.925823 kubelet[2864]: E1031 02:45:36.925413 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:45:36.926639 kubelet[2864]: E1031 02:45:36.926225 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rx5f2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-n7gvx_calico-system(c44deaed-49c1-414b-a77a-e9b3a43a004d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:36.928594 kubelet[2864]: E1031 02:45:36.927582 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:45:39.580695 containerd[1609]: time="2025-10-31T02:45:39.580326902Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 31 02:45:39.905327 containerd[1609]: time="2025-10-31T02:45:39.903636459Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:39.907414 containerd[1609]: time="2025-10-31T02:45:39.907153005Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 31 02:45:39.907632 containerd[1609]: time="2025-10-31T02:45:39.907175822Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 31 02:45:39.908293 kubelet[2864]: E1031 02:45:39.908180 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:45:39.909208 kubelet[2864]: E1031 02:45:39.908978 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:45:39.910410 kubelet[2864]: E1031 02:45:39.910336 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:39.917344 containerd[1609]: time="2025-10-31T02:45:39.917303844Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 31 02:45:40.234143 containerd[1609]: time="2025-10-31T02:45:40.233566165Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:40.237319 containerd[1609]: time="2025-10-31T02:45:40.237245875Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 31 02:45:40.237717 containerd[1609]: time="2025-10-31T02:45:40.237335669Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 31 02:45:40.243043 kubelet[2864]: E1031 02:45:40.238165 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:45:40.243043 kubelet[2864]: E1031 02:45:40.238430 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:45:40.243043 kubelet[2864]: E1031 02:45:40.238979 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:40.243689 kubelet[2864]: E1031 02:45:40.243618 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:45:40.580574 containerd[1609]: time="2025-10-31T02:45:40.579613495Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 31 02:45:40.919894 containerd[1609]: time="2025-10-31T02:45:40.918126326Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:45:40.920791 containerd[1609]: time="2025-10-31T02:45:40.920185133Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 31 02:45:40.920791 containerd[1609]: time="2025-10-31T02:45:40.920296375Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 31 02:45:40.922432 kubelet[2864]: E1031 02:45:40.921426 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:45:40.922432 kubelet[2864]: E1031 02:45:40.921526 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:45:40.922432 kubelet[2864]: E1031 02:45:40.921864 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-455vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-74fbc74c8b-kb8vb_calico-system(1e5f0f90-f710-4f9a-b757-2db9fe751f0e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 31 02:45:40.924497 kubelet[2864]: E1031 02:45:40.923539 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:45:41.738418 systemd[1]: Started sshd@10-10.230.25.230:22-147.75.109.163:35982.service - OpenSSH per-connection server daemon (147.75.109.163:35982). Oct 31 02:45:42.704256 sshd[5731]: Accepted publickey for core from 147.75.109.163 port 35982 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:45:42.708819 sshd[5731]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:45:42.721708 systemd-logind[1591]: New session 13 of user core. Oct 31 02:45:42.731620 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 31 02:45:43.527565 sshd[5731]: pam_unix(sshd:session): session closed for user core Oct 31 02:45:43.537935 systemd[1]: sshd@10-10.230.25.230:22-147.75.109.163:35982.service: Deactivated successfully. Oct 31 02:45:43.547834 systemd[1]: session-13.scope: Deactivated successfully. Oct 31 02:45:43.550593 systemd-logind[1591]: Session 13 logged out. Waiting for processes to exit. Oct 31 02:45:43.553171 systemd-logind[1591]: Removed session 13. Oct 31 02:45:45.584118 kubelet[2864]: E1031 02:45:45.582792 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:45:47.581403 kubelet[2864]: E1031 02:45:47.580935 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:45:47.587151 kubelet[2864]: E1031 02:45:47.585647 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:45:48.696114 systemd[1]: Started sshd@11-10.230.25.230:22-147.75.109.163:35984.service - OpenSSH per-connection server daemon (147.75.109.163:35984). Oct 31 02:45:49.584073 kubelet[2864]: E1031 02:45:49.584006 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:45:49.652059 sshd[5746]: Accepted publickey for core from 147.75.109.163 port 35984 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:45:49.662911 sshd[5746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:45:49.675556 systemd-logind[1591]: New session 14 of user core. Oct 31 02:45:49.687569 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 31 02:45:50.575785 kubelet[2864]: E1031 02:45:50.575720 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:45:50.598256 sshd[5746]: pam_unix(sshd:session): session closed for user core Oct 31 02:45:50.606692 systemd[1]: sshd@11-10.230.25.230:22-147.75.109.163:35984.service: Deactivated successfully. Oct 31 02:45:50.612689 systemd-logind[1591]: Session 14 logged out. Waiting for processes to exit. Oct 31 02:45:50.613932 systemd[1]: session-14.scope: Deactivated successfully. Oct 31 02:45:50.616947 systemd-logind[1591]: Removed session 14. Oct 31 02:45:50.756073 systemd[1]: Started sshd@12-10.230.25.230:22-147.75.109.163:37580.service - OpenSSH per-connection server daemon (147.75.109.163:37580). Oct 31 02:45:51.692314 sshd[5761]: Accepted publickey for core from 147.75.109.163 port 37580 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:45:51.699267 sshd[5761]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:45:51.707790 systemd-logind[1591]: New session 15 of user core. Oct 31 02:45:51.717824 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 31 02:45:52.586884 kubelet[2864]: E1031 02:45:52.586663 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:45:52.588420 sshd[5761]: pam_unix(sshd:session): session closed for user core Oct 31 02:45:52.604970 systemd[1]: sshd@12-10.230.25.230:22-147.75.109.163:37580.service: Deactivated successfully. Oct 31 02:45:52.626360 systemd[1]: session-15.scope: Deactivated successfully. Oct 31 02:45:52.628473 systemd-logind[1591]: Session 15 logged out. Waiting for processes to exit. Oct 31 02:45:52.630757 systemd-logind[1591]: Removed session 15. Oct 31 02:45:52.747094 systemd[1]: Started sshd@13-10.230.25.230:22-147.75.109.163:37586.service - OpenSSH per-connection server daemon (147.75.109.163:37586). Oct 31 02:45:53.706197 sshd[5774]: Accepted publickey for core from 147.75.109.163 port 37586 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:45:53.706937 sshd[5774]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:45:53.720686 systemd-logind[1591]: New session 16 of user core. Oct 31 02:45:53.731632 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 31 02:45:54.451152 sshd[5774]: pam_unix(sshd:session): session closed for user core Oct 31 02:45:54.458770 systemd[1]: sshd@13-10.230.25.230:22-147.75.109.163:37586.service: Deactivated successfully. Oct 31 02:45:54.463791 systemd[1]: session-16.scope: Deactivated successfully. Oct 31 02:45:54.465411 systemd-logind[1591]: Session 16 logged out. Waiting for processes to exit. Oct 31 02:45:54.466818 systemd-logind[1591]: Removed session 16. Oct 31 02:45:56.573779 kubelet[2864]: E1031 02:45:56.573665 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:45:59.577953 kubelet[2864]: E1031 02:45:59.576208 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:45:59.579833 kubelet[2864]: E1031 02:45:59.578502 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:45:59.610427 systemd[1]: Started sshd@14-10.230.25.230:22-147.75.109.163:37588.service - OpenSSH per-connection server daemon (147.75.109.163:37588). Oct 31 02:46:00.547101 sshd[5816]: Accepted publickey for core from 147.75.109.163 port 37588 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:00.553914 sshd[5816]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:00.571660 systemd-logind[1591]: New session 17 of user core. Oct 31 02:46:00.581231 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 31 02:46:01.291598 sshd[5816]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:01.298230 systemd[1]: sshd@14-10.230.25.230:22-147.75.109.163:37588.service: Deactivated successfully. Oct 31 02:46:01.302998 systemd[1]: session-17.scope: Deactivated successfully. Oct 31 02:46:01.305258 systemd-logind[1591]: Session 17 logged out. Waiting for processes to exit. Oct 31 02:46:01.307173 systemd-logind[1591]: Removed session 17. Oct 31 02:46:02.578412 kubelet[2864]: E1031 02:46:02.578291 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:46:03.589584 kubelet[2864]: E1031 02:46:03.589389 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:46:04.575155 kubelet[2864]: E1031 02:46:04.575074 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:46:04.576748 kubelet[2864]: E1031 02:46:04.576554 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:46:06.450500 systemd[1]: Started sshd@15-10.230.25.230:22-147.75.109.163:59282.service - OpenSSH per-connection server daemon (147.75.109.163:59282). Oct 31 02:46:07.448735 sshd[5831]: Accepted publickey for core from 147.75.109.163 port 59282 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:07.454114 sshd[5831]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:07.463324 systemd-logind[1591]: New session 18 of user core. Oct 31 02:46:07.470470 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 31 02:46:07.581272 kubelet[2864]: E1031 02:46:07.580678 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:46:08.293335 sshd[5831]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:08.299950 systemd[1]: sshd@15-10.230.25.230:22-147.75.109.163:59282.service: Deactivated successfully. Oct 31 02:46:08.312861 systemd[1]: session-18.scope: Deactivated successfully. Oct 31 02:46:08.316332 systemd-logind[1591]: Session 18 logged out. Waiting for processes to exit. Oct 31 02:46:08.319298 systemd-logind[1591]: Removed session 18. Oct 31 02:46:11.574356 kubelet[2864]: E1031 02:46:11.573907 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:46:11.576591 kubelet[2864]: E1031 02:46:11.574470 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:46:13.447509 systemd[1]: Started sshd@16-10.230.25.230:22-147.75.109.163:59394.service - OpenSSH per-connection server daemon (147.75.109.163:59394). Oct 31 02:46:14.370486 sshd[5847]: Accepted publickey for core from 147.75.109.163 port 59394 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:14.373151 sshd[5847]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:14.384432 systemd-logind[1591]: New session 19 of user core. Oct 31 02:46:14.392766 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 31 02:46:14.575301 kubelet[2864]: E1031 02:46:14.575196 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:46:14.579313 kubelet[2864]: E1031 02:46:14.578762 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:46:15.158431 sshd[5847]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:15.167407 systemd[1]: Started sshd@17-10.230.25.230:22-183.94.76.8:63516.service - OpenSSH per-connection server daemon (183.94.76.8:63516). Oct 31 02:46:15.177578 systemd[1]: sshd@16-10.230.25.230:22-147.75.109.163:59394.service: Deactivated successfully. Oct 31 02:46:15.189929 systemd[1]: session-19.scope: Deactivated successfully. Oct 31 02:46:15.192489 systemd-logind[1591]: Session 19 logged out. Waiting for processes to exit. Oct 31 02:46:15.195564 systemd-logind[1591]: Removed session 19. Oct 31 02:46:15.313474 systemd[1]: Started sshd@18-10.230.25.230:22-147.75.109.163:59408.service - OpenSSH per-connection server daemon (147.75.109.163:59408). Oct 31 02:46:16.241981 sshd[5863]: Accepted publickey for core from 147.75.109.163 port 59408 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:16.246766 sshd[5863]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:16.264277 systemd-logind[1591]: New session 20 of user core. Oct 31 02:46:16.273142 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 31 02:46:16.314629 sshd[5858]: Connection closed by authenticating user root 183.94.76.8 port 63516 [preauth] Oct 31 02:46:16.317640 systemd[1]: sshd@17-10.230.25.230:22-183.94.76.8:63516.service: Deactivated successfully. Oct 31 02:46:16.565214 systemd[1]: Started sshd@19-10.230.25.230:22-183.94.76.8:63763.service - OpenSSH per-connection server daemon (183.94.76.8:63763). Oct 31 02:46:17.453860 sshd[5863]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:17.471661 systemd[1]: sshd@18-10.230.25.230:22-147.75.109.163:59408.service: Deactivated successfully. Oct 31 02:46:17.478538 systemd[1]: session-20.scope: Deactivated successfully. Oct 31 02:46:17.482609 systemd-logind[1591]: Session 20 logged out. Waiting for processes to exit. Oct 31 02:46:17.485252 systemd-logind[1591]: Removed session 20. Oct 31 02:46:17.577657 containerd[1609]: time="2025-10-31T02:46:17.577217160Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 31 02:46:17.607910 systemd[1]: Started sshd@20-10.230.25.230:22-147.75.109.163:59416.service - OpenSSH per-connection server daemon (147.75.109.163:59416). Oct 31 02:46:17.765091 sshd[5870]: Connection closed by authenticating user root 183.94.76.8 port 63763 [preauth] Oct 31 02:46:17.770678 systemd[1]: sshd@19-10.230.25.230:22-183.94.76.8:63763.service: Deactivated successfully. Oct 31 02:46:17.900872 containerd[1609]: time="2025-10-31T02:46:17.900791786Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:17.902653 containerd[1609]: time="2025-10-31T02:46:17.902576433Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 31 02:46:17.903665 containerd[1609]: time="2025-10-31T02:46:17.902617557Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 31 02:46:17.903732 kubelet[2864]: E1031 02:46:17.903051 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:46:17.903732 kubelet[2864]: E1031 02:46:17.903184 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 31 02:46:17.905048 kubelet[2864]: E1031 02:46:17.903479 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:bd677a2a2fa44c8a81a4028fc03bd754,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:17.906993 containerd[1609]: time="2025-10-31T02:46:17.906928707Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 31 02:46:18.007530 systemd[1]: Started sshd@21-10.230.25.230:22-183.94.76.8:64078.service - OpenSSH per-connection server daemon (183.94.76.8:64078). Oct 31 02:46:18.230055 containerd[1609]: time="2025-10-31T02:46:18.228967828Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:18.230674 containerd[1609]: time="2025-10-31T02:46:18.230621585Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 31 02:46:18.230781 containerd[1609]: time="2025-10-31T02:46:18.230727971Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 31 02:46:18.232054 kubelet[2864]: E1031 02:46:18.230946 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:46:18.232054 kubelet[2864]: E1031 02:46:18.231011 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 31 02:46:18.232054 kubelet[2864]: E1031 02:46:18.231272 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6djh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5d659c4594-7qhbb_calico-system(ec0f842a-be78-41a7-a220-2e129c052a42): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:18.232687 kubelet[2864]: E1031 02:46:18.232625 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:46:18.551903 sshd[5881]: Accepted publickey for core from 147.75.109.163 port 59416 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:18.555910 sshd[5881]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:18.571983 systemd-logind[1591]: New session 21 of user core. Oct 31 02:46:18.577912 containerd[1609]: time="2025-10-31T02:46:18.577812930Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 31 02:46:18.581578 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 31 02:46:18.923786 containerd[1609]: time="2025-10-31T02:46:18.921353887Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:18.923786 containerd[1609]: time="2025-10-31T02:46:18.923285018Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 31 02:46:18.923786 containerd[1609]: time="2025-10-31T02:46:18.923485025Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 31 02:46:18.924167 kubelet[2864]: E1031 02:46:18.923907 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:46:18.924167 kubelet[2864]: E1031 02:46:18.924010 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 31 02:46:18.927263 kubelet[2864]: E1031 02:46:18.924274 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rx5f2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-n7gvx_calico-system(c44deaed-49c1-414b-a77a-e9b3a43a004d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:18.927263 kubelet[2864]: E1031 02:46:18.926187 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:46:19.234161 sshd[5886]: Connection closed by authenticating user root 183.94.76.8 port 64078 [preauth] Oct 31 02:46:19.236063 systemd[1]: sshd@21-10.230.25.230:22-183.94.76.8:64078.service: Deactivated successfully. Oct 31 02:46:19.457481 systemd[1]: Started sshd@22-10.230.25.230:22-183.94.76.8:64352.service - OpenSSH per-connection server daemon (183.94.76.8:64352). Oct 31 02:46:20.453990 sshd[5881]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:20.465853 systemd[1]: sshd@20-10.230.25.230:22-147.75.109.163:59416.service: Deactivated successfully. Oct 31 02:46:20.467479 systemd-logind[1591]: Session 21 logged out. Waiting for processes to exit. Oct 31 02:46:20.480062 systemd[1]: session-21.scope: Deactivated successfully. Oct 31 02:46:20.485837 systemd-logind[1591]: Removed session 21. Oct 31 02:46:20.601811 systemd[1]: Started sshd@23-10.230.25.230:22-147.75.109.163:56962.service - OpenSSH per-connection server daemon (147.75.109.163:56962). Oct 31 02:46:21.528177 sshd[5916]: Accepted publickey for core from 147.75.109.163 port 56962 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:21.532110 sshd[5916]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:21.550266 systemd-logind[1591]: New session 22 of user core. Oct 31 02:46:21.556551 systemd[1]: Started session-22.scope - Session 22 of User core. Oct 31 02:46:22.578629 containerd[1609]: time="2025-10-31T02:46:22.578271990Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:46:22.736860 sshd[5916]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:22.753256 systemd[1]: sshd@23-10.230.25.230:22-147.75.109.163:56962.service: Deactivated successfully. Oct 31 02:46:22.760909 systemd[1]: session-22.scope: Deactivated successfully. Oct 31 02:46:22.761269 systemd-logind[1591]: Session 22 logged out. Waiting for processes to exit. Oct 31 02:46:22.763980 systemd-logind[1591]: Removed session 22. Oct 31 02:46:22.887442 systemd[1]: Started sshd@24-10.230.25.230:22-147.75.109.163:56966.service - OpenSSH per-connection server daemon (147.75.109.163:56966). Oct 31 02:46:22.933341 containerd[1609]: time="2025-10-31T02:46:22.933060102Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:22.934823 containerd[1609]: time="2025-10-31T02:46:22.934684759Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:46:22.936406 containerd[1609]: time="2025-10-31T02:46:22.934955775Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:46:22.936500 kubelet[2864]: E1031 02:46:22.936108 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:46:22.936500 kubelet[2864]: E1031 02:46:22.936224 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:46:22.938985 kubelet[2864]: E1031 02:46:22.936660 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krkfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-75444f6b5f-gp7s4_calico-apiserver(51cf0923-33f8-4c41-8a1b-075035cdaba9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:22.938985 kubelet[2864]: E1031 02:46:22.938246 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:46:22.943765 containerd[1609]: time="2025-10-31T02:46:22.941083189Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 31 02:46:23.268771 containerd[1609]: time="2025-10-31T02:46:23.268505470Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:23.271413 containerd[1609]: time="2025-10-31T02:46:23.270960933Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 31 02:46:23.271413 containerd[1609]: time="2025-10-31T02:46:23.271108197Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 31 02:46:23.271925 kubelet[2864]: E1031 02:46:23.271315 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:46:23.271925 kubelet[2864]: E1031 02:46:23.271405 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 31 02:46:23.271925 kubelet[2864]: E1031 02:46:23.271587 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-455vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-74fbc74c8b-kb8vb_calico-system(1e5f0f90-f710-4f9a-b757-2db9fe751f0e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:23.273063 kubelet[2864]: E1031 02:46:23.272907 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:46:23.819058 sshd[5930]: Accepted publickey for core from 147.75.109.163 port 56966 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:23.822887 sshd[5930]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:23.841069 systemd-logind[1591]: New session 23 of user core. Oct 31 02:46:23.847501 systemd[1]: Started session-23.scope - Session 23 of User core. Oct 31 02:46:24.578756 containerd[1609]: time="2025-10-31T02:46:24.578538028Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:46:24.730849 sshd[5930]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:24.744896 systemd[1]: sshd@24-10.230.25.230:22-147.75.109.163:56966.service: Deactivated successfully. Oct 31 02:46:24.754993 systemd[1]: session-23.scope: Deactivated successfully. Oct 31 02:46:24.757820 systemd-logind[1591]: Session 23 logged out. Waiting for processes to exit. Oct 31 02:46:24.761574 systemd-logind[1591]: Removed session 23. Oct 31 02:46:24.898136 containerd[1609]: time="2025-10-31T02:46:24.897410582Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:24.898874 containerd[1609]: time="2025-10-31T02:46:24.898823596Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:46:24.899145 containerd[1609]: time="2025-10-31T02:46:24.898925223Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:46:24.899238 kubelet[2864]: E1031 02:46:24.899174 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:46:24.899825 kubelet[2864]: E1031 02:46:24.899260 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:46:24.899825 kubelet[2864]: E1031 02:46:24.899461 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tffjx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-ml2bp_calico-apiserver(2c40f225-7c8b-464e-9dd2-615c17de403d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:24.901011 kubelet[2864]: E1031 02:46:24.900625 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:46:26.583059 containerd[1609]: time="2025-10-31T02:46:26.582603751Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 31 02:46:26.914487 containerd[1609]: time="2025-10-31T02:46:26.914139713Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:26.915861 containerd[1609]: time="2025-10-31T02:46:26.915764762Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 31 02:46:26.915991 containerd[1609]: time="2025-10-31T02:46:26.915817746Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 31 02:46:26.916359 kubelet[2864]: E1031 02:46:26.916238 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:46:26.917071 kubelet[2864]: E1031 02:46:26.916404 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 31 02:46:26.918581 containerd[1609]: time="2025-10-31T02:46:26.917672008Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 31 02:46:26.918829 kubelet[2864]: E1031 02:46:26.918208 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xqpd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-589687cff4-5xd2q_calico-apiserver(1d3e6a01-de45-45a9-bd90-9cad31c5290e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:26.920059 kubelet[2864]: E1031 02:46:26.919937 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:46:27.244237 containerd[1609]: time="2025-10-31T02:46:27.243772712Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:27.245549 containerd[1609]: time="2025-10-31T02:46:27.245366080Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 31 02:46:27.245945 containerd[1609]: time="2025-10-31T02:46:27.245424369Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 31 02:46:27.246518 kubelet[2864]: E1031 02:46:27.246415 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:46:27.246518 kubelet[2864]: E1031 02:46:27.246508 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 31 02:46:27.246903 kubelet[2864]: E1031 02:46:27.246687 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:27.250056 containerd[1609]: time="2025-10-31T02:46:27.249331135Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 31 02:46:27.560073 containerd[1609]: time="2025-10-31T02:46:27.559901428Z" level=info msg="trying next host - response was http.StatusNotFound" host=ghcr.io Oct 31 02:46:27.562307 containerd[1609]: time="2025-10-31T02:46:27.562190566Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 31 02:46:27.562307 containerd[1609]: time="2025-10-31T02:46:27.562239837Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 31 02:46:27.563307 kubelet[2864]: E1031 02:46:27.562935 2864 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:46:27.563307 kubelet[2864]: E1031 02:46:27.563098 2864 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 31 02:46:27.565217 kubelet[2864]: E1031 02:46:27.564087 2864 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8nr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-d2znc_calico-system(7293ccde-9a5c-4930-845b-a0c0b00c9ac3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 31 02:46:27.566376 kubelet[2864]: E1031 02:46:27.566302 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:46:29.890687 systemd[1]: Started sshd@25-10.230.25.230:22-147.75.109.163:56980.service - OpenSSH per-connection server daemon (147.75.109.163:56980). Oct 31 02:46:30.581138 kubelet[2864]: E1031 02:46:30.579438 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:46:30.799527 sshd[5978]: Accepted publickey for core from 147.75.109.163 port 56980 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:30.803161 sshd[5978]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:30.818837 systemd-logind[1591]: New session 24 of user core. Oct 31 02:46:30.825558 systemd[1]: Started session-24.scope - Session 24 of User core. Oct 31 02:46:31.713316 sshd[5978]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:31.718381 systemd[1]: sshd@25-10.230.25.230:22-147.75.109.163:56980.service: Deactivated successfully. Oct 31 02:46:31.727369 systemd[1]: session-24.scope: Deactivated successfully. Oct 31 02:46:31.731126 systemd-logind[1591]: Session 24 logged out. Waiting for processes to exit. Oct 31 02:46:31.733402 systemd-logind[1591]: Removed session 24. Oct 31 02:46:32.575763 kubelet[2864]: E1031 02:46:32.574802 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:46:35.605128 kubelet[2864]: E1031 02:46:35.604275 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d" Oct 31 02:46:36.882507 systemd[1]: Started sshd@26-10.230.25.230:22-147.75.109.163:59064.service - OpenSSH per-connection server daemon (147.75.109.163:59064). Oct 31 02:46:37.578072 kubelet[2864]: E1031 02:46:37.577988 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-75444f6b5f-gp7s4" podUID="51cf0923-33f8-4c41-8a1b-075035cdaba9" Oct 31 02:46:37.822457 sshd[6001]: Accepted publickey for core from 147.75.109.163 port 59064 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:37.827711 sshd[6001]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:37.841947 systemd-logind[1591]: New session 25 of user core. Oct 31 02:46:37.850604 systemd[1]: Started session-25.scope - Session 25 of User core. Oct 31 02:46:38.575680 kubelet[2864]: E1031 02:46:38.575197 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-74fbc74c8b-kb8vb" podUID="1e5f0f90-f710-4f9a-b757-2db9fe751f0e" Oct 31 02:46:38.621394 sshd[6001]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:38.629922 systemd-logind[1591]: Session 25 logged out. Waiting for processes to exit. Oct 31 02:46:38.633454 systemd[1]: sshd@26-10.230.25.230:22-147.75.109.163:59064.service: Deactivated successfully. Oct 31 02:46:38.649343 systemd[1]: session-25.scope: Deactivated successfully. Oct 31 02:46:38.653482 systemd-logind[1591]: Removed session 25. Oct 31 02:46:40.599736 kubelet[2864]: E1031 02:46:40.599528 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-d2znc" podUID="7293ccde-9a5c-4930-845b-a0c0b00c9ac3" Oct 31 02:46:41.579661 kubelet[2864]: E1031 02:46:41.578623 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-5xd2q" podUID="1d3e6a01-de45-45a9-bd90-9cad31c5290e" Oct 31 02:46:42.576351 kubelet[2864]: E1031 02:46:42.576262 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5d659c4594-7qhbb" podUID="ec0f842a-be78-41a7-a220-2e129c052a42" Oct 31 02:46:43.779516 systemd[1]: Started sshd@27-10.230.25.230:22-147.75.109.163:41700.service - OpenSSH per-connection server daemon (147.75.109.163:41700). Oct 31 02:46:44.702933 sshd[6017]: Accepted publickey for core from 147.75.109.163 port 41700 ssh2: RSA SHA256:d+nLrY8Dsc9/yJeymnhT6SHXxGEkOkD6rfqu967eLjU Oct 31 02:46:44.704015 sshd[6017]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 31 02:46:44.716663 systemd-logind[1591]: New session 26 of user core. Oct 31 02:46:44.722567 systemd[1]: Started session-26.scope - Session 26 of User core. Oct 31 02:46:45.509750 sshd[6017]: pam_unix(sshd:session): session closed for user core Oct 31 02:46:45.524017 systemd[1]: sshd@27-10.230.25.230:22-147.75.109.163:41700.service: Deactivated successfully. Oct 31 02:46:45.526596 systemd-logind[1591]: Session 26 logged out. Waiting for processes to exit. Oct 31 02:46:45.534851 systemd[1]: session-26.scope: Deactivated successfully. Oct 31 02:46:45.540085 systemd-logind[1591]: Removed session 26. Oct 31 02:46:45.587925 kubelet[2864]: E1031 02:46:45.586083 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-n7gvx" podUID="c44deaed-49c1-414b-a77a-e9b3a43a004d" Oct 31 02:46:48.574582 kubelet[2864]: E1031 02:46:48.573748 2864 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-589687cff4-ml2bp" podUID="2c40f225-7c8b-464e-9dd2-615c17de403d"