Oct 27 08:23:47.107161 kernel: Linux version 6.12.54-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Mon Oct 27 06:24:35 -00 2025 Oct 27 08:23:47.107198 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=e6ac205aca0358d0b739fe2cba6f8244850dbdc9027fd8e7442161fce065515e Oct 27 08:23:47.107214 kernel: BIOS-provided physical RAM map: Oct 27 08:23:47.107222 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Oct 27 08:23:47.107229 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Oct 27 08:23:47.107236 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Oct 27 08:23:47.107244 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Oct 27 08:23:47.107254 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Oct 27 08:23:47.107262 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Oct 27 08:23:47.107274 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Oct 27 08:23:47.107282 kernel: NX (Execute Disable) protection: active Oct 27 08:23:47.107289 kernel: APIC: Static calls initialized Oct 27 08:23:47.107296 kernel: SMBIOS 2.8 present. Oct 27 08:23:47.107304 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Oct 27 08:23:47.107313 kernel: DMI: Memory slots populated: 1/1 Oct 27 08:23:47.107332 kernel: Hypervisor detected: KVM Oct 27 08:23:47.107347 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Oct 27 08:23:47.107356 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Oct 27 08:23:47.107365 kernel: kvm-clock: using sched offset of 3779142484 cycles Oct 27 08:23:47.107375 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Oct 27 08:23:47.107383 kernel: tsc: Detected 2494.138 MHz processor Oct 27 08:23:47.107392 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Oct 27 08:23:47.107401 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Oct 27 08:23:47.107417 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Oct 27 08:23:47.107430 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Oct 27 08:23:47.107438 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Oct 27 08:23:47.107447 kernel: ACPI: Early table checksum verification disabled Oct 27 08:23:47.107456 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Oct 27 08:23:47.107464 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107473 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107488 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107496 kernel: ACPI: FACS 0x000000007FFE0000 000040 Oct 27 08:23:47.107505 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107514 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107522 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107531 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 27 08:23:47.107539 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Oct 27 08:23:47.107553 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Oct 27 08:23:47.107562 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Oct 27 08:23:47.107571 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Oct 27 08:23:47.107622 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Oct 27 08:23:47.107631 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Oct 27 08:23:47.107647 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Oct 27 08:23:47.107656 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Oct 27 08:23:47.107665 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Oct 27 08:23:47.107674 kernel: NUMA: Node 0 [mem 0x00001000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00001000-0x7ffdafff] Oct 27 08:23:47.107683 kernel: NODE_DATA(0) allocated [mem 0x7ffd3dc0-0x7ffdafff] Oct 27 08:23:47.107692 kernel: Zone ranges: Oct 27 08:23:47.107707 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Oct 27 08:23:47.107716 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Oct 27 08:23:47.107725 kernel: Normal empty Oct 27 08:23:47.107734 kernel: Device empty Oct 27 08:23:47.107743 kernel: Movable zone start for each node Oct 27 08:23:47.107752 kernel: Early memory node ranges Oct 27 08:23:47.107761 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Oct 27 08:23:47.107769 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Oct 27 08:23:47.107784 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Oct 27 08:23:47.107793 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 27 08:23:47.107801 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Oct 27 08:23:47.107811 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Oct 27 08:23:47.107820 kernel: ACPI: PM-Timer IO Port: 0x608 Oct 27 08:23:47.107831 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Oct 27 08:23:47.107840 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Oct 27 08:23:47.107868 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Oct 27 08:23:47.107881 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Oct 27 08:23:47.107894 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Oct 27 08:23:47.107910 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Oct 27 08:23:47.107923 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Oct 27 08:23:47.107935 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Oct 27 08:23:47.107948 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Oct 27 08:23:47.107969 kernel: TSC deadline timer available Oct 27 08:23:47.107978 kernel: CPU topo: Max. logical packages: 1 Oct 27 08:23:47.107987 kernel: CPU topo: Max. logical dies: 1 Oct 27 08:23:47.107996 kernel: CPU topo: Max. dies per package: 1 Oct 27 08:23:47.108005 kernel: CPU topo: Max. threads per core: 1 Oct 27 08:23:47.108014 kernel: CPU topo: Num. cores per package: 2 Oct 27 08:23:47.108023 kernel: CPU topo: Num. threads per package: 2 Oct 27 08:23:47.108032 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Oct 27 08:23:47.108047 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Oct 27 08:23:47.108056 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Oct 27 08:23:47.108065 kernel: Booting paravirtualized kernel on KVM Oct 27 08:23:47.108074 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Oct 27 08:23:47.108083 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Oct 27 08:23:47.108093 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Oct 27 08:23:47.108101 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Oct 27 08:23:47.108117 kernel: pcpu-alloc: [0] 0 1 Oct 27 08:23:47.108126 kernel: kvm-guest: PV spinlocks disabled, no host support Oct 27 08:23:47.108136 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=e6ac205aca0358d0b739fe2cba6f8244850dbdc9027fd8e7442161fce065515e Oct 27 08:23:47.108146 kernel: random: crng init done Oct 27 08:23:47.108155 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 27 08:23:47.108164 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Oct 27 08:23:47.108173 kernel: Fallback order for Node 0: 0 Oct 27 08:23:47.108188 kernel: Built 1 zonelists, mobility grouping on. Total pages: 524153 Oct 27 08:23:47.108197 kernel: Policy zone: DMA32 Oct 27 08:23:47.108206 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 27 08:23:47.108215 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Oct 27 08:23:47.108229 kernel: Kernel/User page tables isolation: enabled Oct 27 08:23:47.108242 kernel: ftrace: allocating 40092 entries in 157 pages Oct 27 08:23:47.108255 kernel: ftrace: allocated 157 pages with 5 groups Oct 27 08:23:47.108278 kernel: Dynamic Preempt: voluntary Oct 27 08:23:47.108289 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 27 08:23:47.108300 kernel: rcu: RCU event tracing is enabled. Oct 27 08:23:47.108309 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Oct 27 08:23:47.108318 kernel: Trampoline variant of Tasks RCU enabled. Oct 27 08:23:47.108327 kernel: Rude variant of Tasks RCU enabled. Oct 27 08:23:47.108335 kernel: Tracing variant of Tasks RCU enabled. Oct 27 08:23:47.108344 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 27 08:23:47.108360 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Oct 27 08:23:47.108373 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 27 08:23:47.108391 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 27 08:23:47.108404 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 27 08:23:47.108419 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Oct 27 08:23:47.108429 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 27 08:23:47.108439 kernel: Console: colour VGA+ 80x25 Oct 27 08:23:47.108455 kernel: printk: legacy console [tty0] enabled Oct 27 08:23:47.108464 kernel: printk: legacy console [ttyS0] enabled Oct 27 08:23:47.108474 kernel: ACPI: Core revision 20240827 Oct 27 08:23:47.108483 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Oct 27 08:23:47.108509 kernel: APIC: Switch to symmetric I/O mode setup Oct 27 08:23:47.108524 kernel: x2apic enabled Oct 27 08:23:47.108533 kernel: APIC: Switched APIC routing to: physical x2apic Oct 27 08:23:47.108543 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Oct 27 08:23:47.108553 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39838d43, max_idle_ns: 440795267131 ns Oct 27 08:23:47.108569 kernel: Calibrating delay loop (skipped) preset value.. 4988.27 BogoMIPS (lpj=2494138) Oct 27 08:23:47.108579 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Oct 27 08:23:47.108606 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Oct 27 08:23:47.108615 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Oct 27 08:23:47.108631 kernel: Spectre V2 : Mitigation: Retpolines Oct 27 08:23:47.108641 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Oct 27 08:23:47.108650 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Oct 27 08:23:47.108660 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Oct 27 08:23:47.108669 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Oct 27 08:23:47.108679 kernel: MDS: Mitigation: Clear CPU buffers Oct 27 08:23:47.108688 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Oct 27 08:23:47.108704 kernel: active return thunk: its_return_thunk Oct 27 08:23:47.108713 kernel: ITS: Mitigation: Aligned branch/return thunks Oct 27 08:23:47.108723 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Oct 27 08:23:47.108732 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Oct 27 08:23:47.108742 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Oct 27 08:23:47.108751 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Oct 27 08:23:47.108761 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Oct 27 08:23:47.108778 kernel: Freeing SMP alternatives memory: 32K Oct 27 08:23:47.108936 kernel: pid_max: default: 32768 minimum: 301 Oct 27 08:23:47.108946 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Oct 27 08:23:47.108956 kernel: landlock: Up and running. Oct 27 08:23:47.108965 kernel: SELinux: Initializing. Oct 27 08:23:47.108975 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 27 08:23:47.108985 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 27 08:23:47.109008 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Oct 27 08:23:47.109021 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Oct 27 08:23:47.109033 kernel: signal: max sigframe size: 1776 Oct 27 08:23:47.109048 kernel: rcu: Hierarchical SRCU implementation. Oct 27 08:23:47.109062 kernel: rcu: Max phase no-delay instances is 400. Oct 27 08:23:47.109075 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Oct 27 08:23:47.109088 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Oct 27 08:23:47.109110 kernel: smp: Bringing up secondary CPUs ... Oct 27 08:23:47.109122 kernel: smpboot: x86: Booting SMP configuration: Oct 27 08:23:47.109132 kernel: .... node #0, CPUs: #1 Oct 27 08:23:47.109142 kernel: smp: Brought up 1 node, 2 CPUs Oct 27 08:23:47.109151 kernel: smpboot: Total of 2 processors activated (9976.55 BogoMIPS) Oct 27 08:23:47.109162 kernel: Memory: 1989436K/2096612K available (14336K kernel code, 2443K rwdata, 26064K rodata, 15964K init, 2080K bss, 102612K reserved, 0K cma-reserved) Oct 27 08:23:47.109172 kernel: devtmpfs: initialized Oct 27 08:23:47.109187 kernel: x86/mm: Memory block size: 128MB Oct 27 08:23:47.109227 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 27 08:23:47.109241 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Oct 27 08:23:47.109251 kernel: pinctrl core: initialized pinctrl subsystem Oct 27 08:23:47.109261 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 27 08:23:47.109271 kernel: audit: initializing netlink subsys (disabled) Oct 27 08:23:47.109280 kernel: audit: type=2000 audit(1761553424.749:1): state=initialized audit_enabled=0 res=1 Oct 27 08:23:47.109298 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 27 08:23:47.109308 kernel: thermal_sys: Registered thermal governor 'user_space' Oct 27 08:23:47.109318 kernel: cpuidle: using governor menu Oct 27 08:23:47.109327 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 27 08:23:47.114022 kernel: dca service started, version 1.12.1 Oct 27 08:23:47.114044 kernel: PCI: Using configuration type 1 for base access Oct 27 08:23:47.114055 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Oct 27 08:23:47.114065 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 27 08:23:47.114085 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Oct 27 08:23:47.114095 kernel: ACPI: Added _OSI(Module Device) Oct 27 08:23:47.114105 kernel: ACPI: Added _OSI(Processor Device) Oct 27 08:23:47.114115 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 27 08:23:47.114125 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Oct 27 08:23:47.114134 kernel: ACPI: Interpreter enabled Oct 27 08:23:47.114143 kernel: ACPI: PM: (supports S0 S5) Oct 27 08:23:47.114159 kernel: ACPI: Using IOAPIC for interrupt routing Oct 27 08:23:47.114169 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Oct 27 08:23:47.114182 kernel: PCI: Using E820 reservations for host bridge windows Oct 27 08:23:47.114199 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Oct 27 08:23:47.114212 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 27 08:23:47.114463 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Oct 27 08:23:47.114647 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Oct 27 08:23:47.114790 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Oct 27 08:23:47.114803 kernel: acpiphp: Slot [3] registered Oct 27 08:23:47.114813 kernel: acpiphp: Slot [4] registered Oct 27 08:23:47.114823 kernel: acpiphp: Slot [5] registered Oct 27 08:23:47.114832 kernel: acpiphp: Slot [6] registered Oct 27 08:23:47.114842 kernel: acpiphp: Slot [7] registered Oct 27 08:23:47.114863 kernel: acpiphp: Slot [8] registered Oct 27 08:23:47.114872 kernel: acpiphp: Slot [9] registered Oct 27 08:23:47.114882 kernel: acpiphp: Slot [10] registered Oct 27 08:23:47.114892 kernel: acpiphp: Slot [11] registered Oct 27 08:23:47.114902 kernel: acpiphp: Slot [12] registered Oct 27 08:23:47.114912 kernel: acpiphp: Slot [13] registered Oct 27 08:23:47.114922 kernel: acpiphp: Slot [14] registered Oct 27 08:23:47.114937 kernel: acpiphp: Slot [15] registered Oct 27 08:23:47.114947 kernel: acpiphp: Slot [16] registered Oct 27 08:23:47.114957 kernel: acpiphp: Slot [17] registered Oct 27 08:23:47.114967 kernel: acpiphp: Slot [18] registered Oct 27 08:23:47.114976 kernel: acpiphp: Slot [19] registered Oct 27 08:23:47.114986 kernel: acpiphp: Slot [20] registered Oct 27 08:23:47.114995 kernel: acpiphp: Slot [21] registered Oct 27 08:23:47.115012 kernel: acpiphp: Slot [22] registered Oct 27 08:23:47.115022 kernel: acpiphp: Slot [23] registered Oct 27 08:23:47.115031 kernel: acpiphp: Slot [24] registered Oct 27 08:23:47.115041 kernel: acpiphp: Slot [25] registered Oct 27 08:23:47.115051 kernel: acpiphp: Slot [26] registered Oct 27 08:23:47.115060 kernel: acpiphp: Slot [27] registered Oct 27 08:23:47.115070 kernel: acpiphp: Slot [28] registered Oct 27 08:23:47.115080 kernel: acpiphp: Slot [29] registered Oct 27 08:23:47.115095 kernel: acpiphp: Slot [30] registered Oct 27 08:23:47.115104 kernel: acpiphp: Slot [31] registered Oct 27 08:23:47.115115 kernel: PCI host bridge to bus 0000:00 Oct 27 08:23:47.115261 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Oct 27 08:23:47.115387 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Oct 27 08:23:47.115510 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Oct 27 08:23:47.116312 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Oct 27 08:23:47.116455 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Oct 27 08:23:47.116629 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 27 08:23:47.116788 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint Oct 27 08:23:47.116931 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint Oct 27 08:23:47.117090 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 conventional PCI endpoint Oct 27 08:23:47.117246 kernel: pci 0000:00:01.1: BAR 4 [io 0xc1e0-0xc1ef] Oct 27 08:23:47.117412 kernel: pci 0000:00:01.1: BAR 0 [io 0x01f0-0x01f7]: legacy IDE quirk Oct 27 08:23:47.117582 kernel: pci 0000:00:01.1: BAR 1 [io 0x03f6]: legacy IDE quirk Oct 27 08:23:47.117756 kernel: pci 0000:00:01.1: BAR 2 [io 0x0170-0x0177]: legacy IDE quirk Oct 27 08:23:47.117889 kernel: pci 0000:00:01.1: BAR 3 [io 0x0376]: legacy IDE quirk Oct 27 08:23:47.118045 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 conventional PCI endpoint Oct 27 08:23:47.118179 kernel: pci 0000:00:01.2: BAR 4 [io 0xc180-0xc19f] Oct 27 08:23:47.118319 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 conventional PCI endpoint Oct 27 08:23:47.118488 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Oct 27 08:23:47.118649 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Oct 27 08:23:47.118808 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint Oct 27 08:23:47.118944 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfe000000-0xfe7fffff pref] Oct 27 08:23:47.120854 kernel: pci 0000:00:02.0: BAR 2 [mem 0xfe800000-0xfe803fff 64bit pref] Oct 27 08:23:47.121006 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfebf0000-0xfebf0fff] Oct 27 08:23:47.122566 kernel: pci 0000:00:02.0: ROM [mem 0xfebe0000-0xfebeffff pref] Oct 27 08:23:47.123059 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Oct 27 08:23:47.123236 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Oct 27 08:23:47.123370 kernel: pci 0000:00:03.0: BAR 0 [io 0xc1a0-0xc1bf] Oct 27 08:23:47.123501 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebf1000-0xfebf1fff] Oct 27 08:23:47.124381 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe804000-0xfe807fff 64bit pref] Oct 27 08:23:47.124554 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Oct 27 08:23:47.124723 kernel: pci 0000:00:04.0: BAR 0 [io 0xc1c0-0xc1df] Oct 27 08:23:47.128290 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebf2000-0xfebf2fff] Oct 27 08:23:47.128470 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe808000-0xfe80bfff 64bit pref] Oct 27 08:23:47.132781 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint Oct 27 08:23:47.132951 kernel: pci 0000:00:05.0: BAR 0 [io 0xc100-0xc13f] Oct 27 08:23:47.133087 kernel: pci 0000:00:05.0: BAR 1 [mem 0xfebf3000-0xfebf3fff] Oct 27 08:23:47.133263 kernel: pci 0000:00:05.0: BAR 4 [mem 0xfe80c000-0xfe80ffff 64bit pref] Oct 27 08:23:47.133407 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Oct 27 08:23:47.133541 kernel: pci 0000:00:06.0: BAR 0 [io 0xc000-0xc07f] Oct 27 08:23:47.135468 kernel: pci 0000:00:06.0: BAR 1 [mem 0xfebf4000-0xfebf4fff] Oct 27 08:23:47.135661 kernel: pci 0000:00:06.0: BAR 4 [mem 0xfe810000-0xfe813fff 64bit pref] Oct 27 08:23:47.135808 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Oct 27 08:23:47.136042 kernel: pci 0000:00:07.0: BAR 0 [io 0xc080-0xc0ff] Oct 27 08:23:47.136247 kernel: pci 0000:00:07.0: BAR 1 [mem 0xfebf5000-0xfebf5fff] Oct 27 08:23:47.136449 kernel: pci 0000:00:07.0: BAR 4 [mem 0xfe814000-0xfe817fff 64bit pref] Oct 27 08:23:47.138909 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 conventional PCI endpoint Oct 27 08:23:47.139092 kernel: pci 0000:00:08.0: BAR 0 [io 0xc140-0xc17f] Oct 27 08:23:47.139252 kernel: pci 0000:00:08.0: BAR 4 [mem 0xfe818000-0xfe81bfff 64bit pref] Oct 27 08:23:47.139265 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Oct 27 08:23:47.139276 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Oct 27 08:23:47.139286 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Oct 27 08:23:47.139295 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Oct 27 08:23:47.139306 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Oct 27 08:23:47.139316 kernel: iommu: Default domain type: Translated Oct 27 08:23:47.139333 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Oct 27 08:23:47.139343 kernel: PCI: Using ACPI for IRQ routing Oct 27 08:23:47.139353 kernel: PCI: pci_cache_line_size set to 64 bytes Oct 27 08:23:47.139362 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Oct 27 08:23:47.139372 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Oct 27 08:23:47.139512 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Oct 27 08:23:47.141741 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Oct 27 08:23:47.141923 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Oct 27 08:23:47.141937 kernel: vgaarb: loaded Oct 27 08:23:47.141947 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Oct 27 08:23:47.141957 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Oct 27 08:23:47.141967 kernel: clocksource: Switched to clocksource kvm-clock Oct 27 08:23:47.141977 kernel: VFS: Disk quotas dquot_6.6.0 Oct 27 08:23:47.141987 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 27 08:23:47.142005 kernel: pnp: PnP ACPI init Oct 27 08:23:47.142015 kernel: pnp: PnP ACPI: found 4 devices Oct 27 08:23:47.142025 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Oct 27 08:23:47.142034 kernel: NET: Registered PF_INET protocol family Oct 27 08:23:47.142044 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 27 08:23:47.142054 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Oct 27 08:23:47.142064 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 27 08:23:47.142079 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Oct 27 08:23:47.142089 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Oct 27 08:23:47.142099 kernel: TCP: Hash tables configured (established 16384 bind 16384) Oct 27 08:23:47.142109 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 27 08:23:47.142119 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 27 08:23:47.142128 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 27 08:23:47.142138 kernel: NET: Registered PF_XDP protocol family Oct 27 08:23:47.144722 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Oct 27 08:23:47.144859 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Oct 27 08:23:47.144982 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Oct 27 08:23:47.145103 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Oct 27 08:23:47.145244 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Oct 27 08:23:47.145393 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Oct 27 08:23:47.147655 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Oct 27 08:23:47.147678 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Oct 27 08:23:47.147836 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x720 took 41672 usecs Oct 27 08:23:47.147850 kernel: PCI: CLS 0 bytes, default 64 Oct 27 08:23:47.147861 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Oct 27 08:23:47.147871 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39838d43, max_idle_ns: 440795267131 ns Oct 27 08:23:47.147881 kernel: Initialise system trusted keyrings Oct 27 08:23:47.147907 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Oct 27 08:23:47.147917 kernel: Key type asymmetric registered Oct 27 08:23:47.147927 kernel: Asymmetric key parser 'x509' registered Oct 27 08:23:47.147937 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Oct 27 08:23:47.147947 kernel: io scheduler mq-deadline registered Oct 27 08:23:47.147957 kernel: io scheduler kyber registered Oct 27 08:23:47.147967 kernel: io scheduler bfq registered Oct 27 08:23:47.147982 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Oct 27 08:23:47.147992 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Oct 27 08:23:47.148002 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Oct 27 08:23:47.148012 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Oct 27 08:23:47.148022 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 27 08:23:47.148032 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Oct 27 08:23:47.148042 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Oct 27 08:23:47.148057 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Oct 27 08:23:47.148067 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Oct 27 08:23:47.148077 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Oct 27 08:23:47.148234 kernel: rtc_cmos 00:03: RTC can wake from S4 Oct 27 08:23:47.148406 kernel: rtc_cmos 00:03: registered as rtc0 Oct 27 08:23:47.148603 kernel: rtc_cmos 00:03: setting system clock to 2025-10-27T08:23:45 UTC (1761553425) Oct 27 08:23:47.148760 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Oct 27 08:23:47.148773 kernel: intel_pstate: CPU model not supported Oct 27 08:23:47.148783 kernel: NET: Registered PF_INET6 protocol family Oct 27 08:23:47.148793 kernel: Segment Routing with IPv6 Oct 27 08:23:47.148804 kernel: In-situ OAM (IOAM) with IPv6 Oct 27 08:23:47.148813 kernel: NET: Registered PF_PACKET protocol family Oct 27 08:23:47.148823 kernel: Key type dns_resolver registered Oct 27 08:23:47.148840 kernel: IPI shorthand broadcast: enabled Oct 27 08:23:47.148850 kernel: sched_clock: Marking stable (1272006538, 144478087)->(1437417628, -20933003) Oct 27 08:23:47.148860 kernel: registered taskstats version 1 Oct 27 08:23:47.148869 kernel: Loading compiled-in X.509 certificates Oct 27 08:23:47.148879 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.54-flatcar: 6c7ef547b8d769f7afd2708799fb9c3145695bfb' Oct 27 08:23:47.148889 kernel: Demotion targets for Node 0: null Oct 27 08:23:47.148898 kernel: Key type .fscrypt registered Oct 27 08:23:47.148914 kernel: Key type fscrypt-provisioning registered Oct 27 08:23:47.148957 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 27 08:23:47.148972 kernel: ima: Allocated hash algorithm: sha1 Oct 27 08:23:47.148982 kernel: ima: No architecture policies found Oct 27 08:23:47.148992 kernel: clk: Disabling unused clocks Oct 27 08:23:47.149002 kernel: Freeing unused kernel image (initmem) memory: 15964K Oct 27 08:23:47.149013 kernel: Write protecting the kernel read-only data: 40960k Oct 27 08:23:47.149023 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Oct 27 08:23:47.149039 kernel: Run /init as init process Oct 27 08:23:47.149049 kernel: with arguments: Oct 27 08:23:47.149059 kernel: /init Oct 27 08:23:47.149069 kernel: with environment: Oct 27 08:23:47.149079 kernel: HOME=/ Oct 27 08:23:47.149089 kernel: TERM=linux Oct 27 08:23:47.149099 kernel: SCSI subsystem initialized Oct 27 08:23:47.149114 kernel: libata version 3.00 loaded. Oct 27 08:23:47.149361 kernel: ata_piix 0000:00:01.1: version 2.13 Oct 27 08:23:47.149570 kernel: scsi host0: ata_piix Oct 27 08:23:47.150380 kernel: scsi host1: ata_piix Oct 27 08:23:47.150403 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 lpm-pol 0 Oct 27 08:23:47.150435 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 lpm-pol 0 Oct 27 08:23:47.150446 kernel: ACPI: bus type USB registered Oct 27 08:23:47.150456 kernel: usbcore: registered new interface driver usbfs Oct 27 08:23:47.150467 kernel: usbcore: registered new interface driver hub Oct 27 08:23:47.150477 kernel: usbcore: registered new device driver usb Oct 27 08:23:47.150700 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Oct 27 08:23:47.150845 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Oct 27 08:23:47.150995 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Oct 27 08:23:47.151133 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Oct 27 08:23:47.151301 kernel: hub 1-0:1.0: USB hub found Oct 27 08:23:47.151446 kernel: hub 1-0:1.0: 2 ports detected Oct 27 08:23:47.151637 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Oct 27 08:23:47.151783 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Oct 27 08:23:47.151799 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 27 08:23:47.151809 kernel: GPT:16515071 != 125829119 Oct 27 08:23:47.151819 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 27 08:23:47.151829 kernel: GPT:16515071 != 125829119 Oct 27 08:23:47.151848 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 27 08:23:47.151858 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 27 08:23:47.152006 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Oct 27 08:23:47.152139 kernel: virtio_blk virtio5: [vdb] 980 512-byte logical blocks (502 kB/490 KiB) Oct 27 08:23:47.152280 kernel: virtio_scsi virtio3: 2/0/0 default/read/poll queues Oct 27 08:23:47.152434 kernel: scsi host2: Virtio SCSI HBA Oct 27 08:23:47.152459 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152470 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 27 08:23:47.152480 kernel: device-mapper: uevent: version 1.0.3 Oct 27 08:23:47.152491 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Oct 27 08:23:47.152501 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Oct 27 08:23:47.152512 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152527 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152537 kernel: raid6: avx2x4 gen() 14251 MB/s Oct 27 08:23:47.152548 kernel: raid6: avx2x2 gen() 15158 MB/s Oct 27 08:23:47.152558 kernel: raid6: avx2x1 gen() 11752 MB/s Oct 27 08:23:47.152568 kernel: raid6: using algorithm avx2x2 gen() 15158 MB/s Oct 27 08:23:47.152579 kernel: raid6: .... xor() 11785 MB/s, rmw enabled Oct 27 08:23:47.152603 kernel: raid6: using avx2x2 recovery algorithm Oct 27 08:23:47.152620 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152637 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152647 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152656 kernel: xor: automatically using best checksumming function avx Oct 27 08:23:47.152667 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152676 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 27 08:23:47.152687 kernel: BTRFS: device fsid bf514789-bcec-4c15-ac9d-e4c3d19a42b2 devid 1 transid 36 /dev/mapper/usr (253:0) scanned by mount (157) Oct 27 08:23:47.152698 kernel: BTRFS info (device dm-0): first mount of filesystem bf514789-bcec-4c15-ac9d-e4c3d19a42b2 Oct 27 08:23:47.152714 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Oct 27 08:23:47.152724 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 27 08:23:47.152734 kernel: BTRFS info (device dm-0): enabling free space tree Oct 27 08:23:47.152745 kernel: Invalid ELF header magic: != \u007fELF Oct 27 08:23:47.152755 kernel: loop: module loaded Oct 27 08:23:47.152765 kernel: loop0: detected capacity change from 0 to 100120 Oct 27 08:23:47.152775 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 27 08:23:47.152787 systemd[1]: Successfully made /usr/ read-only. Oct 27 08:23:47.152806 systemd[1]: systemd 257.7 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 27 08:23:47.152817 systemd[1]: Detected virtualization kvm. Oct 27 08:23:47.152827 systemd[1]: Detected architecture x86-64. Oct 27 08:23:47.152838 systemd[1]: Running in initrd. Oct 27 08:23:47.152848 systemd[1]: No hostname configured, using default hostname. Oct 27 08:23:47.152864 systemd[1]: Hostname set to . Oct 27 08:23:47.152875 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Oct 27 08:23:47.152886 systemd[1]: Queued start job for default target initrd.target. Oct 27 08:23:47.152897 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Oct 27 08:23:47.152908 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 27 08:23:47.152918 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 27 08:23:47.152929 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 27 08:23:47.152946 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 27 08:23:47.152958 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 27 08:23:47.152969 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 27 08:23:47.152980 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 27 08:23:47.152990 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 27 08:23:47.153007 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Oct 27 08:23:47.153018 systemd[1]: Reached target paths.target - Path Units. Oct 27 08:23:47.153029 systemd[1]: Reached target slices.target - Slice Units. Oct 27 08:23:47.153040 systemd[1]: Reached target swap.target - Swaps. Oct 27 08:23:47.153051 systemd[1]: Reached target timers.target - Timer Units. Oct 27 08:23:47.153061 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 27 08:23:47.153072 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 27 08:23:47.153088 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 27 08:23:47.153099 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Oct 27 08:23:47.153110 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 27 08:23:47.153120 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 27 08:23:47.153131 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 27 08:23:47.153142 systemd[1]: Reached target sockets.target - Socket Units. Oct 27 08:23:47.153153 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 27 08:23:47.153169 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 27 08:23:47.153180 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 27 08:23:47.153191 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 27 08:23:47.153290 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Oct 27 08:23:47.153309 systemd[1]: Starting systemd-fsck-usr.service... Oct 27 08:23:47.153322 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 27 08:23:47.153333 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 27 08:23:47.153359 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:47.153371 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 27 08:23:47.153382 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 27 08:23:47.153399 systemd[1]: Finished systemd-fsck-usr.service. Oct 27 08:23:47.153410 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 27 08:23:47.153477 systemd-journald[292]: Collecting audit messages is disabled. Oct 27 08:23:47.153511 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 27 08:23:47.153522 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 27 08:23:47.153534 kernel: Bridge firewalling registered Oct 27 08:23:47.153546 systemd-journald[292]: Journal started Oct 27 08:23:47.153568 systemd-journald[292]: Runtime Journal (/run/log/journal/04b56fd3b9a34f6e9ca1ad00ac0d386b) is 4.9M, max 39.2M, 34.3M free. Oct 27 08:23:47.149493 systemd-modules-load[294]: Inserted module 'br_netfilter' Oct 27 08:23:47.202623 systemd[1]: Started systemd-journald.service - Journal Service. Oct 27 08:23:47.203327 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 27 08:23:47.204177 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:47.208290 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 27 08:23:47.209747 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 27 08:23:47.212802 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 27 08:23:47.215755 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 27 08:23:47.238545 systemd-tmpfiles[315]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Oct 27 08:23:47.238701 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 27 08:23:47.243504 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 27 08:23:47.247172 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 27 08:23:47.250736 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 27 08:23:47.253456 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 27 08:23:47.256795 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 27 08:23:47.284737 dracut-cmdline[333]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=e6ac205aca0358d0b739fe2cba6f8244850dbdc9027fd8e7442161fce065515e Oct 27 08:23:47.306952 systemd-resolved[327]: Positive Trust Anchors: Oct 27 08:23:47.306969 systemd-resolved[327]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 27 08:23:47.306973 systemd-resolved[327]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Oct 27 08:23:47.307010 systemd-resolved[327]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 27 08:23:47.341553 systemd-resolved[327]: Defaulting to hostname 'linux'. Oct 27 08:23:47.343179 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 27 08:23:47.343803 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 27 08:23:47.407635 kernel: Loading iSCSI transport class v2.0-870. Oct 27 08:23:47.423619 kernel: iscsi: registered transport (tcp) Oct 27 08:23:47.447727 kernel: iscsi: registered transport (qla4xxx) Oct 27 08:23:47.447810 kernel: QLogic iSCSI HBA Driver Oct 27 08:23:47.479439 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 27 08:23:47.505020 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 27 08:23:47.506062 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 27 08:23:47.562489 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 27 08:23:47.565007 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 27 08:23:47.566298 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 27 08:23:47.606976 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 27 08:23:47.609788 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 27 08:23:47.640683 systemd-udevd[572]: Using default interface naming scheme 'v257'. Oct 27 08:23:47.652470 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 27 08:23:47.656482 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 27 08:23:47.692291 dracut-pre-trigger[642]: rd.md=0: removing MD RAID activation Oct 27 08:23:47.699650 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 27 08:23:47.704893 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 27 08:23:47.726622 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 27 08:23:47.730760 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 27 08:23:47.760570 systemd-networkd[691]: lo: Link UP Oct 27 08:23:47.760580 systemd-networkd[691]: lo: Gained carrier Oct 27 08:23:47.761821 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 27 08:23:47.762476 systemd[1]: Reached target network.target - Network. Oct 27 08:23:47.815886 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 27 08:23:47.818868 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 27 08:23:47.950994 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Oct 27 08:23:47.963371 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Oct 27 08:23:47.974397 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Oct 27 08:23:47.985418 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 27 08:23:47.987975 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 27 08:23:47.996631 kernel: cryptd: max_cpu_qlen set to 1000 Oct 27 08:23:48.015005 disk-uuid[744]: Primary Header is updated. Oct 27 08:23:48.015005 disk-uuid[744]: Secondary Entries is updated. Oct 27 08:23:48.015005 disk-uuid[744]: Secondary Header is updated. Oct 27 08:23:48.031638 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Oct 27 08:23:48.094671 systemd-networkd[691]: eth1: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Oct 27 08:23:48.094680 systemd-networkd[691]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 27 08:23:48.096770 systemd-networkd[691]: eth1: Link UP Oct 27 08:23:48.097456 systemd-networkd[691]: eth1: Gained carrier Oct 27 08:23:48.097476 systemd-networkd[691]: eth1: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Oct 27 08:23:48.114649 kernel: AES CTR mode by8 optimization enabled Oct 27 08:23:48.111672 systemd-networkd[691]: eth1: DHCPv4 address 10.124.0.33/20 acquired from 169.254.169.253 Oct 27 08:23:48.129044 systemd-networkd[691]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/yy-digitalocean.network Oct 27 08:23:48.129059 systemd-networkd[691]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Oct 27 08:23:48.131067 systemd-networkd[691]: eth0: Link UP Oct 27 08:23:48.131272 systemd-networkd[691]: eth0: Gained carrier Oct 27 08:23:48.131291 systemd-networkd[691]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/yy-digitalocean.network Oct 27 08:23:48.142298 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 27 08:23:48.142424 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:48.145669 systemd-networkd[691]: eth0: DHCPv4 address 64.23.205.201/19, gateway 64.23.192.1 acquired from 169.254.169.253 Oct 27 08:23:48.147838 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:48.153665 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:48.222740 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 27 08:23:48.273675 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:48.276374 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 27 08:23:48.277034 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 27 08:23:48.278174 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 27 08:23:48.280476 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 27 08:23:48.307335 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 27 08:23:49.082997 disk-uuid[745]: Warning: The kernel is still using the old partition table. Oct 27 08:23:49.082997 disk-uuid[745]: The new table will be used at the next reboot or after you Oct 27 08:23:49.082997 disk-uuid[745]: run partprobe(8) or kpartx(8) Oct 27 08:23:49.082997 disk-uuid[745]: The operation has completed successfully. Oct 27 08:23:49.091100 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 27 08:23:49.091314 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 27 08:23:49.094731 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 27 08:23:49.132633 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (834) Oct 27 08:23:49.137895 kernel: BTRFS info (device vda6): first mount of filesystem 3c7e1d30-69bc-4811-963d-029e55854883 Oct 27 08:23:49.137983 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 27 08:23:49.142517 kernel: BTRFS info (device vda6): turning on async discard Oct 27 08:23:49.142618 kernel: BTRFS info (device vda6): enabling free space tree Oct 27 08:23:49.152651 kernel: BTRFS info (device vda6): last unmount of filesystem 3c7e1d30-69bc-4811-963d-029e55854883 Oct 27 08:23:49.152996 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 27 08:23:49.155838 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 27 08:23:49.357173 ignition[853]: Ignition 2.22.0 Oct 27 08:23:49.357207 ignition[853]: Stage: fetch-offline Oct 27 08:23:49.357257 ignition[853]: no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:49.357273 ignition[853]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:49.357424 ignition[853]: parsed url from cmdline: "" Oct 27 08:23:49.357430 ignition[853]: no config URL provided Oct 27 08:23:49.360561 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 27 08:23:49.357438 ignition[853]: reading system config file "/usr/lib/ignition/user.ign" Oct 27 08:23:49.357451 ignition[853]: no config at "/usr/lib/ignition/user.ign" Oct 27 08:23:49.357460 ignition[853]: failed to fetch config: resource requires networking Oct 27 08:23:49.358043 ignition[853]: Ignition finished successfully Oct 27 08:23:49.363812 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Oct 27 08:23:49.417672 ignition[862]: Ignition 2.22.0 Oct 27 08:23:49.417689 ignition[862]: Stage: fetch Oct 27 08:23:49.417908 ignition[862]: no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:49.417935 ignition[862]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:49.418102 ignition[862]: parsed url from cmdline: "" Oct 27 08:23:49.418111 ignition[862]: no config URL provided Oct 27 08:23:49.418121 ignition[862]: reading system config file "/usr/lib/ignition/user.ign" Oct 27 08:23:49.418136 ignition[862]: no config at "/usr/lib/ignition/user.ign" Oct 27 08:23:49.418179 ignition[862]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Oct 27 08:23:49.430868 systemd-networkd[691]: eth0: Gained IPv6LL Oct 27 08:23:49.435759 ignition[862]: GET result: OK Oct 27 08:23:49.435925 ignition[862]: parsing config with SHA512: e893d7a8f08f7de012f66b490c8b53d54a7ab10d657546c2ab6443e6dee2166a49201101b681cd0c4eab830f102f17685403d4d53922b22c930c9c7cf78b0f53 Oct 27 08:23:49.442574 unknown[862]: fetched base config from "system" Oct 27 08:23:49.442607 unknown[862]: fetched base config from "system" Oct 27 08:23:49.443125 ignition[862]: fetch: fetch complete Oct 27 08:23:49.442617 unknown[862]: fetched user config from "digitalocean" Oct 27 08:23:49.443134 ignition[862]: fetch: fetch passed Oct 27 08:23:49.443202 ignition[862]: Ignition finished successfully Oct 27 08:23:49.446611 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Oct 27 08:23:49.448214 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 27 08:23:49.494084 ignition[868]: Ignition 2.22.0 Oct 27 08:23:49.494098 ignition[868]: Stage: kargs Oct 27 08:23:49.494257 ignition[868]: no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:49.494266 ignition[868]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:49.498097 ignition[868]: kargs: kargs passed Oct 27 08:23:49.498185 ignition[868]: Ignition finished successfully Oct 27 08:23:49.500544 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 27 08:23:49.502731 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 27 08:23:49.544051 ignition[874]: Ignition 2.22.0 Oct 27 08:23:49.544078 ignition[874]: Stage: disks Oct 27 08:23:49.544228 ignition[874]: no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:49.544236 ignition[874]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:49.545406 ignition[874]: disks: disks passed Oct 27 08:23:49.545460 ignition[874]: Ignition finished successfully Oct 27 08:23:49.547279 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 27 08:23:49.548784 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 27 08:23:49.549366 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 27 08:23:49.550255 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 27 08:23:49.551289 systemd[1]: Reached target sysinit.target - System Initialization. Oct 27 08:23:49.552223 systemd[1]: Reached target basic.target - Basic System. Oct 27 08:23:49.554327 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 27 08:23:49.558794 systemd-networkd[691]: eth1: Gained IPv6LL Oct 27 08:23:49.595918 systemd-fsck[882]: ROOT: clean, 15/456736 files, 38230/456704 blocks Oct 27 08:23:49.598195 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 27 08:23:49.601086 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 27 08:23:49.724604 kernel: EXT4-fs (vda9): mounted filesystem e90e2fe3-e1db-4bff-abac-c8d1d032f674 r/w with ordered data mode. Quota mode: none. Oct 27 08:23:49.725565 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 27 08:23:49.727183 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 27 08:23:49.729962 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 27 08:23:49.732093 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 27 08:23:49.742847 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... Oct 27 08:23:49.749797 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Oct 27 08:23:49.751846 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (891) Oct 27 08:23:49.754819 kernel: BTRFS info (device vda6): first mount of filesystem 3c7e1d30-69bc-4811-963d-029e55854883 Oct 27 08:23:49.754917 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 27 08:23:49.759532 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 27 08:23:49.760349 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 27 08:23:49.763728 kernel: BTRFS info (device vda6): turning on async discard Oct 27 08:23:49.763797 kernel: BTRFS info (device vda6): enabling free space tree Oct 27 08:23:49.770384 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 27 08:23:49.771226 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 27 08:23:49.774836 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 27 08:23:49.883566 initrd-setup-root[921]: cut: /sysroot/etc/passwd: No such file or directory Oct 27 08:23:49.892491 coreos-metadata[893]: Oct 27 08:23:49.892 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 27 08:23:49.898456 coreos-metadata[894]: Oct 27 08:23:49.898 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 27 08:23:49.900832 initrd-setup-root[928]: cut: /sysroot/etc/group: No such file or directory Oct 27 08:23:49.905164 coreos-metadata[893]: Oct 27 08:23:49.905 INFO Fetch successful Oct 27 08:23:49.909647 initrd-setup-root[935]: cut: /sysroot/etc/shadow: No such file or directory Oct 27 08:23:49.912690 coreos-metadata[894]: Oct 27 08:23:49.912 INFO Fetch successful Oct 27 08:23:49.914941 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. Oct 27 08:23:49.916780 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. Oct 27 08:23:49.920522 coreos-metadata[894]: Oct 27 08:23:49.920 INFO wrote hostname ci-9999.9.9-k-4f7b9e69b1 to /sysroot/etc/hostname Oct 27 08:23:49.923406 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Oct 27 08:23:49.925280 initrd-setup-root[943]: cut: /sysroot/etc/gshadow: No such file or directory Oct 27 08:23:50.056901 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 27 08:23:50.059091 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 27 08:23:50.060506 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 27 08:23:50.085464 kernel: BTRFS info (device vda6): last unmount of filesystem 3c7e1d30-69bc-4811-963d-029e55854883 Oct 27 08:23:50.100882 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 27 08:23:50.118713 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 27 08:23:50.128393 ignition[1012]: INFO : Ignition 2.22.0 Oct 27 08:23:50.128393 ignition[1012]: INFO : Stage: mount Oct 27 08:23:50.129924 ignition[1012]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:50.129924 ignition[1012]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:50.132276 ignition[1012]: INFO : mount: mount passed Oct 27 08:23:50.132276 ignition[1012]: INFO : Ignition finished successfully Oct 27 08:23:50.132964 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 27 08:23:50.136390 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 27 08:23:50.163062 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 27 08:23:50.187617 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1023) Oct 27 08:23:50.190728 kernel: BTRFS info (device vda6): first mount of filesystem 3c7e1d30-69bc-4811-963d-029e55854883 Oct 27 08:23:50.190822 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 27 08:23:50.196102 kernel: BTRFS info (device vda6): turning on async discard Oct 27 08:23:50.196194 kernel: BTRFS info (device vda6): enabling free space tree Oct 27 08:23:50.198647 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 27 08:23:50.236079 ignition[1039]: INFO : Ignition 2.22.0 Oct 27 08:23:50.236079 ignition[1039]: INFO : Stage: files Oct 27 08:23:50.237900 ignition[1039]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:50.237900 ignition[1039]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:50.239599 ignition[1039]: DEBUG : files: compiled without relabeling support, skipping Oct 27 08:23:50.239599 ignition[1039]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 27 08:23:50.239599 ignition[1039]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 27 08:23:50.244061 ignition[1039]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 27 08:23:50.244927 ignition[1039]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 27 08:23:50.245740 unknown[1039]: wrote ssh authorized keys file for user: core Oct 27 08:23:50.246484 ignition[1039]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 27 08:23:50.247611 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 27 08:23:50.248342 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Oct 27 08:23:50.288068 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Oct 27 08:23:50.345921 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 27 08:23:50.345921 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 27 08:23:50.348749 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 27 08:23:50.365310 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 27 08:23:50.365310 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 27 08:23:50.365310 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Oct 27 08:23:50.789793 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Oct 27 08:23:51.113543 ignition[1039]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 27 08:23:51.113543 ignition[1039]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Oct 27 08:23:51.116004 ignition[1039]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 27 08:23:51.118678 ignition[1039]: INFO : files: files passed Oct 27 08:23:51.118678 ignition[1039]: INFO : Ignition finished successfully Oct 27 08:23:51.121688 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 27 08:23:51.123877 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 27 08:23:51.127786 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 27 08:23:51.142908 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 27 08:23:51.143034 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 27 08:23:51.155040 initrd-setup-root-after-ignition[1071]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 27 08:23:51.156168 initrd-setup-root-after-ignition[1075]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 27 08:23:51.157317 initrd-setup-root-after-ignition[1071]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 27 08:23:51.158244 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 27 08:23:51.159073 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 27 08:23:51.160765 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 27 08:23:51.216548 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 27 08:23:51.216719 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 27 08:23:51.217886 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 27 08:23:51.218541 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 27 08:23:51.219647 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 27 08:23:51.220759 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 27 08:23:51.247057 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 27 08:23:51.249711 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 27 08:23:51.278344 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Oct 27 08:23:51.279814 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 27 08:23:51.280568 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 27 08:23:51.282838 systemd[1]: Stopped target timers.target - Timer Units. Oct 27 08:23:51.283455 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 27 08:23:51.283636 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 27 08:23:51.284791 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 27 08:23:51.285432 systemd[1]: Stopped target basic.target - Basic System. Oct 27 08:23:51.286716 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 27 08:23:51.287789 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 27 08:23:51.288976 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 27 08:23:51.290345 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Oct 27 08:23:51.291643 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 27 08:23:51.292894 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 27 08:23:51.294139 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 27 08:23:51.295223 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 27 08:23:51.296273 systemd[1]: Stopped target swap.target - Swaps. Oct 27 08:23:51.297483 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 27 08:23:51.297669 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 27 08:23:51.298985 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 27 08:23:51.299835 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 27 08:23:51.300772 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 27 08:23:51.301031 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 27 08:23:51.302014 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 27 08:23:51.302156 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 27 08:23:51.303646 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 27 08:23:51.303815 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 27 08:23:51.304411 systemd[1]: ignition-files.service: Deactivated successfully. Oct 27 08:23:51.304513 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 27 08:23:51.305480 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Oct 27 08:23:51.305611 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Oct 27 08:23:51.308801 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 27 08:23:51.312277 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 27 08:23:51.313357 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 27 08:23:51.313545 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 27 08:23:51.317652 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 27 08:23:51.317814 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 27 08:23:51.319034 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 27 08:23:51.319152 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 27 08:23:51.326092 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 27 08:23:51.327099 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 27 08:23:51.349837 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 27 08:23:51.353436 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 27 08:23:51.353548 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 27 08:23:51.363212 ignition[1095]: INFO : Ignition 2.22.0 Oct 27 08:23:51.364023 ignition[1095]: INFO : Stage: umount Oct 27 08:23:51.364727 ignition[1095]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 27 08:23:51.366663 ignition[1095]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 27 08:23:51.366663 ignition[1095]: INFO : umount: umount passed Oct 27 08:23:51.367722 ignition[1095]: INFO : Ignition finished successfully Oct 27 08:23:51.369717 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 27 08:23:51.369928 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 27 08:23:51.371538 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 27 08:23:51.371716 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 27 08:23:51.372658 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 27 08:23:51.372720 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 27 08:23:51.373486 systemd[1]: ignition-fetch.service: Deactivated successfully. Oct 27 08:23:51.373538 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Oct 27 08:23:51.374324 systemd[1]: Stopped target network.target - Network. Oct 27 08:23:51.375067 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 27 08:23:51.375126 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 27 08:23:51.375941 systemd[1]: Stopped target paths.target - Path Units. Oct 27 08:23:51.376677 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 27 08:23:51.378688 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 27 08:23:51.379216 systemd[1]: Stopped target slices.target - Slice Units. Oct 27 08:23:51.380213 systemd[1]: Stopped target sockets.target - Socket Units. Oct 27 08:23:51.381382 systemd[1]: iscsid.socket: Deactivated successfully. Oct 27 08:23:51.381437 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 27 08:23:51.382140 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 27 08:23:51.382178 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 27 08:23:51.382906 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 27 08:23:51.382977 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 27 08:23:51.383688 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 27 08:23:51.383732 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 27 08:23:51.384504 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 27 08:23:51.384555 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 27 08:23:51.385535 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 27 08:23:51.386367 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 27 08:23:51.395890 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 27 08:23:51.396039 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 27 08:23:51.399280 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 27 08:23:51.399387 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 27 08:23:51.403467 systemd[1]: Stopped target network-pre.target - Preparation for Network. Oct 27 08:23:51.404049 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 27 08:23:51.404098 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 27 08:23:51.406065 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 27 08:23:51.406890 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 27 08:23:51.406975 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 27 08:23:51.408991 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 27 08:23:51.409050 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 27 08:23:51.409628 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 27 08:23:51.409676 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 27 08:23:51.410195 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 27 08:23:51.425612 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 27 08:23:51.425770 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 27 08:23:51.426527 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 27 08:23:51.426572 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 27 08:23:51.431142 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 27 08:23:51.431203 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 27 08:23:51.432881 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 27 08:23:51.432949 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 27 08:23:51.434650 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 27 08:23:51.434716 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 27 08:23:51.435291 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 27 08:23:51.435346 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 27 08:23:51.439247 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 27 08:23:51.440729 systemd[1]: systemd-network-generator.service: Deactivated successfully. Oct 27 08:23:51.440851 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Oct 27 08:23:51.441699 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 27 08:23:51.441778 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 27 08:23:51.442453 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Oct 27 08:23:51.442519 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 27 08:23:51.443560 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 27 08:23:51.443675 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 27 08:23:51.444678 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 27 08:23:51.444745 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:51.456116 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 27 08:23:51.456249 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 27 08:23:51.467461 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 27 08:23:51.467605 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 27 08:23:51.468801 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 27 08:23:51.470314 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 27 08:23:51.493534 systemd[1]: Switching root. Oct 27 08:23:51.529944 systemd-journald[292]: Journal stopped Oct 27 08:23:52.807123 systemd-journald[292]: Received SIGTERM from PID 1 (systemd). Oct 27 08:23:52.807228 kernel: SELinux: policy capability network_peer_controls=1 Oct 27 08:23:52.807284 kernel: SELinux: policy capability open_perms=1 Oct 27 08:23:52.807302 kernel: SELinux: policy capability extended_socket_class=1 Oct 27 08:23:52.807320 kernel: SELinux: policy capability always_check_network=0 Oct 27 08:23:52.807338 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 27 08:23:52.807354 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 27 08:23:52.807367 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 27 08:23:52.807381 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 27 08:23:52.807401 kernel: SELinux: policy capability userspace_initial_context=0 Oct 27 08:23:52.807415 kernel: audit: type=1403 audit(1761553431.708:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 27 08:23:52.807429 systemd[1]: Successfully loaded SELinux policy in 79.660ms. Oct 27 08:23:52.807449 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.490ms. Oct 27 08:23:52.807465 systemd[1]: systemd 257.7 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 27 08:23:52.807479 systemd[1]: Detected virtualization kvm. Oct 27 08:23:52.807510 systemd[1]: Detected architecture x86-64. Oct 27 08:23:52.807531 systemd[1]: Detected first boot. Oct 27 08:23:52.807554 systemd[1]: Hostname set to . Oct 27 08:23:52.807575 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Oct 27 08:23:52.807625 zram_generator::config[1141]: No configuration found. Oct 27 08:23:52.807649 kernel: Guest personality initialized and is inactive Oct 27 08:23:52.807671 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Oct 27 08:23:52.807699 kernel: Initialized host personality Oct 27 08:23:52.807718 kernel: NET: Registered PF_VSOCK protocol family Oct 27 08:23:52.807732 systemd[1]: Populated /etc with preset unit settings. Oct 27 08:23:52.807750 systemd[1]: initrd-switch-root.service: Deactivated successfully. Oct 27 08:23:52.807771 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Oct 27 08:23:52.807787 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Oct 27 08:23:52.807802 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 27 08:23:52.807822 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 27 08:23:52.807841 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 27 08:23:52.807854 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 27 08:23:52.807868 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 27 08:23:52.807882 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 27 08:23:52.807896 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 27 08:23:52.807909 systemd[1]: Created slice user.slice - User and Session Slice. Oct 27 08:23:52.807930 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 27 08:23:52.807951 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 27 08:23:52.807973 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 27 08:23:52.807995 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 27 08:23:52.808017 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 27 08:23:52.808046 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 27 08:23:52.808069 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Oct 27 08:23:52.808091 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 27 08:23:52.808111 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 27 08:23:52.808130 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Oct 27 08:23:52.808145 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Oct 27 08:23:52.808159 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Oct 27 08:23:52.808188 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 27 08:23:52.808202 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 27 08:23:52.808216 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 27 08:23:52.808230 systemd[1]: Reached target slices.target - Slice Units. Oct 27 08:23:52.808243 systemd[1]: Reached target swap.target - Swaps. Oct 27 08:23:52.808256 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 27 08:23:52.808270 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 27 08:23:52.808289 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Oct 27 08:23:52.808302 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 27 08:23:52.808316 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 27 08:23:52.808329 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 27 08:23:52.808346 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 27 08:23:52.808369 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 27 08:23:52.808391 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 27 08:23:52.808421 systemd[1]: Mounting media.mount - External Media Directory... Oct 27 08:23:52.808443 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:52.808464 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 27 08:23:52.808486 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 27 08:23:52.808509 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 27 08:23:52.808530 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 27 08:23:52.808550 systemd[1]: Reached target machines.target - Containers. Oct 27 08:23:52.808570 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 27 08:23:52.813690 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 27 08:23:52.813758 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 27 08:23:52.813791 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 27 08:23:52.813822 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 27 08:23:52.813855 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 27 08:23:52.813882 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 27 08:23:52.813946 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 27 08:23:52.813978 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 27 08:23:52.814012 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 27 08:23:52.814031 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Oct 27 08:23:52.814051 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Oct 27 08:23:52.814071 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Oct 27 08:23:52.814095 systemd[1]: Stopped systemd-fsck-usr.service. Oct 27 08:23:52.814112 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 27 08:23:52.814127 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 27 08:23:52.814141 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 27 08:23:52.814156 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 27 08:23:52.814170 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 27 08:23:52.814189 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Oct 27 08:23:52.814221 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 27 08:23:52.814242 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:52.814262 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 27 08:23:52.814281 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 27 08:23:52.814307 systemd[1]: Mounted media.mount - External Media Directory. Oct 27 08:23:52.814321 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 27 08:23:52.814335 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 27 08:23:52.814350 kernel: fuse: init (API version 7.41) Oct 27 08:23:52.814365 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 27 08:23:52.814378 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 27 08:23:52.814398 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 27 08:23:52.814414 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 27 08:23:52.814427 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 27 08:23:52.814441 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 27 08:23:52.814454 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 27 08:23:52.814474 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 27 08:23:52.814488 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 27 08:23:52.814502 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 27 08:23:52.814515 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 27 08:23:52.814529 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 27 08:23:52.814542 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 27 08:23:52.814556 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 27 08:23:52.814574 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 27 08:23:52.814622 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 27 08:23:52.814636 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 27 08:23:52.814650 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 27 08:23:52.814663 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 27 08:23:52.814676 kernel: ACPI: bus type drm_connector registered Oct 27 08:23:52.814689 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 27 08:23:52.814710 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 27 08:23:52.814730 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 27 08:23:52.814744 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 27 08:23:52.814758 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 27 08:23:52.814771 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 27 08:23:52.814785 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Oct 27 08:23:52.814799 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 27 08:23:52.814812 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 27 08:23:52.814831 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Oct 27 08:23:52.814845 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 27 08:23:52.814902 systemd-journald[1213]: Collecting audit messages is disabled. Oct 27 08:23:52.814928 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 27 08:23:52.814941 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 27 08:23:52.814962 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 27 08:23:52.814976 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 27 08:23:52.814992 systemd-journald[1213]: Journal started Oct 27 08:23:52.815017 systemd-journald[1213]: Runtime Journal (/run/log/journal/04b56fd3b9a34f6e9ca1ad00ac0d386b) is 4.9M, max 39.2M, 34.3M free. Oct 27 08:23:52.362648 systemd[1]: Queued start job for default target multi-user.target. Oct 27 08:23:52.387774 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Oct 27 08:23:52.388352 systemd[1]: systemd-journald.service: Deactivated successfully. Oct 27 08:23:52.811803 systemd-tmpfiles[1242]: ACLs are not supported, ignoring. Oct 27 08:23:52.822715 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 27 08:23:52.811819 systemd-tmpfiles[1242]: ACLs are not supported, ignoring. Oct 27 08:23:52.827731 systemd[1]: Started systemd-journald.service - Journal Service. Oct 27 08:23:52.830873 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Oct 27 08:23:52.832529 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 27 08:23:52.853838 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 27 08:23:52.857876 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 27 08:23:52.859508 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 27 08:23:52.861839 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 27 08:23:52.865918 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Oct 27 08:23:52.883612 kernel: loop1: detected capacity change from 0 to 110984 Oct 27 08:23:52.884955 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 27 08:23:52.898045 systemd-journald[1213]: Time spent on flushing to /var/log/journal/04b56fd3b9a34f6e9ca1ad00ac0d386b is 41.602ms for 1012 entries. Oct 27 08:23:52.898045 systemd-journald[1213]: System Journal (/var/log/journal/04b56fd3b9a34f6e9ca1ad00ac0d386b) is 8M, max 163.5M, 155.5M free. Oct 27 08:23:52.955724 systemd-journald[1213]: Received client request to flush runtime journal. Oct 27 08:23:52.955858 kernel: loop2: detected capacity change from 0 to 8 Oct 27 08:23:52.955902 kernel: loop3: detected capacity change from 0 to 128048 Oct 27 08:23:52.928056 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Oct 27 08:23:52.957547 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 27 08:23:52.960181 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 27 08:23:52.961121 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 27 08:23:52.967832 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 27 08:23:52.971874 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 27 08:23:52.988805 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 27 08:23:52.991609 kernel: loop4: detected capacity change from 0 to 224512 Oct 27 08:23:53.023104 systemd-tmpfiles[1288]: ACLs are not supported, ignoring. Oct 27 08:23:53.023508 systemd-tmpfiles[1288]: ACLs are not supported, ignoring. Oct 27 08:23:53.026620 kernel: loop5: detected capacity change from 0 to 110984 Oct 27 08:23:53.034912 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 27 08:23:53.043621 kernel: loop6: detected capacity change from 0 to 8 Oct 27 08:23:53.049615 kernel: loop7: detected capacity change from 0 to 128048 Oct 27 08:23:53.064632 kernel: loop1: detected capacity change from 0 to 224512 Oct 27 08:23:53.073390 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 27 08:23:53.082365 (sd-merge)[1292]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw', 'oem-digitalocean.raw'. Oct 27 08:23:53.091087 (sd-merge)[1292]: Merged extensions into '/usr'. Oct 27 08:23:53.100082 systemd[1]: Reload requested from client PID 1248 ('systemd-sysext') (unit systemd-sysext.service)... Oct 27 08:23:53.100101 systemd[1]: Reloading... Oct 27 08:23:53.167015 zram_generator::config[1323]: No configuration found. Oct 27 08:23:53.227318 systemd-resolved[1286]: Positive Trust Anchors: Oct 27 08:23:53.227639 systemd-resolved[1286]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 27 08:23:53.227646 systemd-resolved[1286]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Oct 27 08:23:53.227684 systemd-resolved[1286]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 27 08:23:53.251003 systemd-resolved[1286]: Using system hostname 'ci-9999.9.9-k-4f7b9e69b1'. Oct 27 08:23:53.447057 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 27 08:23:53.447244 systemd[1]: Reloading finished in 346 ms. Oct 27 08:23:53.486164 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 27 08:23:53.487212 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 27 08:23:53.491486 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 27 08:23:53.498835 systemd[1]: Starting ensure-sysext.service... Oct 27 08:23:53.503840 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 27 08:23:53.536757 systemd[1]: Reload requested from client PID 1369 ('systemctl') (unit ensure-sysext.service)... Oct 27 08:23:53.536915 systemd[1]: Reloading... Oct 27 08:23:53.565860 systemd-tmpfiles[1370]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Oct 27 08:23:53.566314 systemd-tmpfiles[1370]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Oct 27 08:23:53.568952 systemd-tmpfiles[1370]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 27 08:23:53.569324 systemd-tmpfiles[1370]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 27 08:23:53.571910 systemd-tmpfiles[1370]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 27 08:23:53.572999 systemd-tmpfiles[1370]: ACLs are not supported, ignoring. Oct 27 08:23:53.573792 systemd-tmpfiles[1370]: ACLs are not supported, ignoring. Oct 27 08:23:53.585964 systemd-tmpfiles[1370]: Detected autofs mount point /boot during canonicalization of boot. Oct 27 08:23:53.585979 systemd-tmpfiles[1370]: Skipping /boot Oct 27 08:23:53.620521 systemd-tmpfiles[1370]: Detected autofs mount point /boot during canonicalization of boot. Oct 27 08:23:53.623641 systemd-tmpfiles[1370]: Skipping /boot Oct 27 08:23:53.652626 zram_generator::config[1396]: No configuration found. Oct 27 08:23:53.878437 systemd[1]: Reloading finished in 341 ms. Oct 27 08:23:53.899020 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 27 08:23:53.907170 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 27 08:23:53.916822 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 27 08:23:53.919771 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 27 08:23:53.921861 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 27 08:23:53.925445 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 27 08:23:53.932639 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 27 08:23:53.934722 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 27 08:23:53.945258 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:53.945461 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 27 08:23:53.950775 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 27 08:23:53.954709 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 27 08:23:53.958956 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 27 08:23:53.960050 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 27 08:23:53.960177 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 27 08:23:53.960272 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:53.964699 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:53.964891 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 27 08:23:53.965079 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 27 08:23:53.965177 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 27 08:23:53.965264 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:53.973776 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:53.974074 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 27 08:23:53.988604 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 27 08:23:53.989266 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 27 08:23:53.989387 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 27 08:23:53.989522 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:54.006605 systemd[1]: Finished ensure-sysext.service. Oct 27 08:23:54.010499 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 27 08:23:54.016286 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Oct 27 08:23:54.040447 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 27 08:23:54.041989 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 27 08:23:54.044029 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 27 08:23:54.083582 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 27 08:23:54.085881 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 27 08:23:54.090753 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 27 08:23:54.096380 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 27 08:23:54.097252 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 27 08:23:54.106659 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 27 08:23:54.107243 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 27 08:23:54.110385 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 27 08:23:54.130882 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 27 08:23:54.132912 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 27 08:23:54.137762 systemd-udevd[1449]: Using default interface naming scheme 'v257'. Oct 27 08:23:54.183046 augenrules[1485]: No rules Oct 27 08:23:54.196020 systemd[1]: audit-rules.service: Deactivated successfully. Oct 27 08:23:54.196737 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 27 08:23:54.206771 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 27 08:23:54.211029 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 27 08:23:54.264719 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Oct 27 08:23:54.265725 systemd[1]: Reached target time-set.target - System Time Set. Oct 27 08:23:54.343702 systemd-networkd[1494]: lo: Link UP Oct 27 08:23:54.344017 systemd-networkd[1494]: lo: Gained carrier Oct 27 08:23:54.346470 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 27 08:23:54.347169 systemd[1]: Reached target network.target - Network. Oct 27 08:23:54.350305 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Oct 27 08:23:54.356907 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 27 08:23:54.421269 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Oct 27 08:23:54.446461 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. Oct 27 08:23:54.449922 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Oct 27 08:23:54.450872 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:54.451154 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 27 08:23:54.453231 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 27 08:23:54.459216 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 27 08:23:54.462550 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 27 08:23:54.463115 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 27 08:23:54.463150 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 27 08:23:54.463185 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 27 08:23:54.463200 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 27 08:23:54.500618 kernel: ISO 9660 Extensions: RRIP_1991A Oct 27 08:23:54.500552 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Oct 27 08:23:54.510471 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 27 08:23:54.510939 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 27 08:23:54.513241 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 27 08:23:54.520053 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 27 08:23:54.521386 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 27 08:23:54.522844 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 27 08:23:54.523846 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 27 08:23:54.541066 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Oct 27 08:23:54.554102 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 27 08:23:54.596695 systemd-networkd[1494]: eth0: Configuring with /run/systemd/network/10-a6:21:b7:4b:09:13.network. Oct 27 08:23:54.598135 systemd-networkd[1494]: eth0: Link UP Oct 27 08:23:54.598988 systemd-networkd[1494]: eth0: Gained carrier Oct 27 08:23:54.604936 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:54.637725 kernel: mousedev: PS/2 mouse device common for all mice Oct 27 08:23:54.656525 systemd-networkd[1494]: eth1: Configuring with /run/systemd/network/10-82:ef:40:bb:68:49.network. Oct 27 08:23:54.658063 systemd-networkd[1494]: eth1: Link UP Oct 27 08:23:54.658152 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:54.660706 systemd-networkd[1494]: eth1: Gained carrier Oct 27 08:23:54.660935 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:54.665945 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:54.667339 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:54.690615 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Oct 27 08:23:54.700677 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Oct 27 08:23:54.704719 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 27 08:23:54.710304 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 27 08:23:54.727628 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Oct 27 08:23:54.765622 kernel: ACPI: button: Power Button [PWRF] Oct 27 08:23:54.782205 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 27 08:23:54.785624 ldconfig[1447]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 27 08:23:54.792358 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 27 08:23:54.797853 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 27 08:23:54.817621 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Oct 27 08:23:54.826620 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Oct 27 08:23:54.833408 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 27 08:23:54.858443 kernel: Console: switching to colour dummy device 80x25 Oct 27 08:23:54.858344 systemd[1]: Reached target sysinit.target - System Initialization. Oct 27 08:23:54.861683 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Oct 27 08:23:54.861758 kernel: [drm] features: -context_init Oct 27 08:23:54.860787 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 27 08:23:54.860929 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 27 08:23:54.860998 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Oct 27 08:23:54.861357 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 27 08:23:54.861625 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 27 08:23:54.862080 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 27 08:23:54.862161 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 27 08:23:54.862189 systemd[1]: Reached target paths.target - Path Units. Oct 27 08:23:54.862254 systemd[1]: Reached target timers.target - Timer Units. Oct 27 08:23:54.863537 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 27 08:23:54.866423 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 27 08:23:54.876383 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Oct 27 08:23:54.877104 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Oct 27 08:23:54.878906 systemd[1]: Reached target ssh-access.target - SSH Access Available. Oct 27 08:23:54.897630 kernel: [drm] number of scanouts: 1 Oct 27 08:23:54.897849 kernel: [drm] number of cap sets: 0 Oct 27 08:23:54.925918 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:02.0 on minor 0 Oct 27 08:23:54.924030 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 27 08:23:54.925792 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Oct 27 08:23:54.930641 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 27 08:23:54.932552 systemd[1]: Reached target sockets.target - Socket Units. Oct 27 08:23:54.933579 systemd[1]: Reached target basic.target - Basic System. Oct 27 08:23:54.933740 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 27 08:23:54.933769 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 27 08:23:54.936754 systemd[1]: Starting containerd.service - containerd container runtime... Oct 27 08:23:54.941299 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Oct 27 08:23:54.945957 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 27 08:23:54.950140 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 27 08:23:54.954859 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 27 08:23:54.958908 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 27 08:23:54.959060 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 27 08:23:54.968407 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Oct 27 08:23:54.978185 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 27 08:23:54.981883 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 27 08:23:54.987299 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 27 08:23:54.998046 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 27 08:23:55.006841 extend-filesystems[1561]: Found /dev/vda6 Oct 27 08:23:55.008831 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 27 08:23:55.009568 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 27 08:23:55.010161 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 27 08:23:55.017917 extend-filesystems[1561]: Found /dev/vda9 Oct 27 08:23:55.014943 systemd[1]: Starting update-engine.service - Update Engine... Oct 27 08:23:55.023077 extend-filesystems[1561]: Checking size of /dev/vda9 Oct 27 08:23:55.024850 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 27 08:23:55.039303 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 27 08:23:55.049863 oslogin_cache_refresh[1562]: Refreshing passwd entry cache Oct 27 08:23:55.051994 google_oslogin_nss_cache[1562]: oslogin_cache_refresh[1562]: Refreshing passwd entry cache Oct 27 08:23:55.069695 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Oct 27 08:23:55.070671 jq[1574]: true Oct 27 08:23:55.075025 kernel: Console: switching to colour frame buffer device 128x48 Oct 27 08:23:55.083631 google_oslogin_nss_cache[1562]: oslogin_cache_refresh[1562]: Failure getting users, quitting Oct 27 08:23:55.083631 google_oslogin_nss_cache[1562]: oslogin_cache_refresh[1562]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Oct 27 08:23:55.083631 google_oslogin_nss_cache[1562]: oslogin_cache_refresh[1562]: Refreshing group entry cache Oct 27 08:23:55.081050 oslogin_cache_refresh[1562]: Failure getting users, quitting Oct 27 08:23:55.081081 oslogin_cache_refresh[1562]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Oct 27 08:23:55.081171 oslogin_cache_refresh[1562]: Refreshing group entry cache Oct 27 08:23:55.089801 jq[1560]: false Oct 27 08:23:55.087775 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 27 08:23:55.088318 oslogin_cache_refresh[1562]: Failure getting groups, quitting Oct 27 08:23:55.090426 google_oslogin_nss_cache[1562]: oslogin_cache_refresh[1562]: Failure getting groups, quitting Oct 27 08:23:55.090426 google_oslogin_nss_cache[1562]: oslogin_cache_refresh[1562]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Oct 27 08:23:55.088699 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 27 08:23:55.088333 oslogin_cache_refresh[1562]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Oct 27 08:23:55.099540 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Oct 27 08:23:55.100411 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Oct 27 08:23:55.102161 extend-filesystems[1561]: Resized partition /dev/vda9 Oct 27 08:23:55.166977 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 14138363 blocks Oct 27 08:23:55.167017 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Oct 27 08:23:55.167318 update_engine[1572]: I20251027 08:23:55.128278 1572 main.cc:92] Flatcar Update Engine starting Oct 27 08:23:55.167318 update_engine[1572]: I20251027 08:23:55.134680 1572 update_check_scheduler.cc:74] Next update check in 4m31s Oct 27 08:23:55.160216 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 27 08:23:55.173838 coreos-metadata[1557]: Oct 27 08:23:55.147 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 27 08:23:55.173838 coreos-metadata[1557]: Oct 27 08:23:55.163 INFO Fetch successful Oct 27 08:23:55.125482 dbus-daemon[1558]: [system] SELinux support is enabled Oct 27 08:23:55.174395 extend-filesystems[1594]: resize2fs 1.47.3 (8-Jul-2025) Oct 27 08:23:55.168124 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 27 08:23:55.168678 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 27 08:23:55.190277 systemd[1]: motdgen.service: Deactivated successfully. Oct 27 08:23:55.190928 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 27 08:23:55.199361 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 27 08:23:55.199408 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 27 08:23:55.202451 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 27 08:23:55.202552 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Oct 27 08:23:55.202575 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 27 08:23:55.205408 systemd[1]: Started update-engine.service - Update Engine. Oct 27 08:23:55.217659 jq[1588]: true Oct 27 08:23:55.223277 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 27 08:23:55.237068 (ntainerd)[1602]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Oct 27 08:23:55.253802 tar[1577]: linux-amd64/LICENSE Oct 27 08:23:55.253802 tar[1577]: linux-amd64/helm Oct 27 08:23:55.303542 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Oct 27 08:23:55.306883 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 27 08:23:55.314323 kernel: EXT4-fs (vda9): resized filesystem to 14138363 Oct 27 08:23:55.346776 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:55.362039 extend-filesystems[1594]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Oct 27 08:23:55.362039 extend-filesystems[1594]: old_desc_blocks = 1, new_desc_blocks = 7 Oct 27 08:23:55.362039 extend-filesystems[1594]: The filesystem on /dev/vda9 is now 14138363 (4k) blocks long. Oct 27 08:23:55.363445 extend-filesystems[1561]: Resized filesystem in /dev/vda9 Oct 27 08:23:55.364520 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 27 08:23:55.366110 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 27 08:23:55.455726 bash[1636]: Updated "/home/core/.ssh/authorized_keys" Oct 27 08:23:55.460122 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 27 08:23:55.476162 systemd[1]: Starting sshkeys.service... Oct 27 08:23:55.615380 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Oct 27 08:23:55.623711 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Oct 27 08:23:55.659759 containerd[1602]: time="2025-10-27T08:23:55Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Oct 27 08:23:55.706983 containerd[1602]: time="2025-10-27T08:23:55.699690611Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Oct 27 08:23:55.728898 containerd[1602]: time="2025-10-27T08:23:55.728767926Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="13.63µs" Oct 27 08:23:55.733754 containerd[1602]: time="2025-10-27T08:23:55.733681283Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.733920883Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.734159933Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.734189160Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.734221341Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.734286897Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.734303336Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 27 08:23:55.735378 containerd[1602]: time="2025-10-27T08:23:55.734569370Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 27 08:23:55.755704 containerd[1602]: time="2025-10-27T08:23:55.754765403Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 27 08:23:55.755704 containerd[1602]: time="2025-10-27T08:23:55.754887368Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 27 08:23:55.755704 containerd[1602]: time="2025-10-27T08:23:55.754905781Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Oct 27 08:23:55.755704 containerd[1602]: time="2025-10-27T08:23:55.755165601Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Oct 27 08:23:55.756331 containerd[1602]: time="2025-10-27T08:23:55.756265750Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 27 08:23:55.758729 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 27 08:23:55.759083 containerd[1602]: time="2025-10-27T08:23:55.758793183Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 27 08:23:55.759083 containerd[1602]: time="2025-10-27T08:23:55.758844745Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Oct 27 08:23:55.759038 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:55.759745 containerd[1602]: time="2025-10-27T08:23:55.759468130Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Oct 27 08:23:55.769786 containerd[1602]: time="2025-10-27T08:23:55.764696709Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Oct 27 08:23:55.769786 containerd[1602]: time="2025-10-27T08:23:55.768349544Z" level=info msg="metadata content store policy set" policy=shared Oct 27 08:23:55.772954 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777288532Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777372471Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777389070Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777407291Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777422392Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777440867Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777456747Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777489024Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777510137Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777524064Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777537161Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Oct 27 08:23:55.777614 containerd[1602]: time="2025-10-27T08:23:55.777552723Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.781012465Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.781098098Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.781683533Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.781717017Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.782661962Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.782703499Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.782723250Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.782737092Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.783416190Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Oct 27 08:23:55.783500 containerd[1602]: time="2025-10-27T08:23:55.783452006Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Oct 27 08:23:55.786067 containerd[1602]: time="2025-10-27T08:23:55.783471419Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Oct 27 08:23:55.786067 containerd[1602]: time="2025-10-27T08:23:55.785996820Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Oct 27 08:23:55.786067 containerd[1602]: time="2025-10-27T08:23:55.786036309Z" level=info msg="Start snapshots syncer" Oct 27 08:23:55.792150 containerd[1602]: time="2025-10-27T08:23:55.791626787Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Oct 27 08:23:55.792150 containerd[1602]: time="2025-10-27T08:23:55.792028986Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Oct 27 08:23:55.793496 containerd[1602]: time="2025-10-27T08:23:55.792529603Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Oct 27 08:23:55.797870 containerd[1602]: time="2025-10-27T08:23:55.796100728Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Oct 27 08:23:55.799773 containerd[1602]: time="2025-10-27T08:23:55.798001577Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Oct 27 08:23:55.799773 containerd[1602]: time="2025-10-27T08:23:55.798055713Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Oct 27 08:23:55.799773 containerd[1602]: time="2025-10-27T08:23:55.799652187Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Oct 27 08:23:55.799773 containerd[1602]: time="2025-10-27T08:23:55.799675014Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Oct 27 08:23:55.799773 containerd[1602]: time="2025-10-27T08:23:55.799721255Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Oct 27 08:23:55.799773 containerd[1602]: time="2025-10-27T08:23:55.799737467Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Oct 27 08:23:55.803540 kernel: EDAC MC: Ver: 3.0.0 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805271673Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805379051Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805396129Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805420161Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805484289Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805506824Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805527232Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805539077Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805547956Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805575491Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805618348Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805643665Z" level=info msg="runtime interface created" Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805649737Z" level=info msg="created NRI interface" Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805659751Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Oct 27 08:23:55.805840 containerd[1602]: time="2025-10-27T08:23:55.805690922Z" level=info msg="Connect containerd service" Oct 27 08:23:55.806281 containerd[1602]: time="2025-10-27T08:23:55.805756144Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 27 08:23:55.818333 containerd[1602]: time="2025-10-27T08:23:55.816149309Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 27 08:23:55.850255 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:55.866049 coreos-metadata[1645]: Oct 27 08:23:55.864 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 27 08:23:55.872992 systemd-logind[1571]: New seat seat0. Oct 27 08:23:55.873942 coreos-metadata[1645]: Oct 27 08:23:55.873 INFO Fetch successful Oct 27 08:23:55.879334 systemd-logind[1571]: Watching system buttons on /dev/input/event2 (Power Button) Oct 27 08:23:55.879378 systemd-logind[1571]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Oct 27 08:23:55.880049 systemd[1]: Started systemd-logind.service - User Login Management. Oct 27 08:23:55.896917 unknown[1645]: wrote ssh authorized keys file for user: core Oct 27 08:23:55.922766 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 27 08:23:55.923057 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:55.926886 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:55.933938 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 27 08:23:55.992633 update-ssh-keys[1667]: Updated "/home/core/.ssh/authorized_keys" Oct 27 08:23:55.996884 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Oct 27 08:23:56.000240 locksmithd[1612]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 27 08:23:56.011727 systemd[1]: Finished sshkeys.service. Oct 27 08:23:56.151754 systemd-networkd[1494]: eth0: Gained IPv6LL Oct 27 08:23:56.153659 containerd[1602]: time="2025-10-27T08:23:56.148344516Z" level=info msg="Start subscribing containerd event" Oct 27 08:23:56.153261 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.153937376Z" level=info msg="Start recovering state" Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.153850870Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154224412Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154804500Z" level=info msg="Start event monitor" Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154843478Z" level=info msg="Start cni network conf syncer for default" Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154876116Z" level=info msg="Start streaming server" Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154890627Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154902333Z" level=info msg="runtime interface starting up..." Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154911884Z" level=info msg="starting plugins..." Oct 27 08:23:56.155320 containerd[1602]: time="2025-10-27T08:23:56.154945610Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Oct 27 08:23:56.167173 containerd[1602]: time="2025-10-27T08:23:56.165803286Z" level=info msg="containerd successfully booted in 0.515298s" Oct 27 08:23:56.166028 systemd[1]: Started containerd.service - containerd container runtime. Oct 27 08:23:56.170769 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 27 08:23:56.173355 systemd[1]: Reached target network-online.target - Network is Online. Oct 27 08:23:56.181847 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 27 08:23:56.186706 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 27 08:23:56.264629 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 27 08:23:56.302520 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 27 08:23:56.407788 systemd-networkd[1494]: eth1: Gained IPv6LL Oct 27 08:23:56.409341 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:23:56.570640 sshd_keygen[1596]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 27 08:23:56.647658 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 27 08:23:56.653835 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 27 08:23:56.677018 systemd[1]: issuegen.service: Deactivated successfully. Oct 27 08:23:56.677440 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 27 08:23:56.685909 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 27 08:23:56.713492 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 27 08:23:56.724142 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 27 08:23:56.727480 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Oct 27 08:23:56.729286 systemd[1]: Reached target getty.target - Login Prompts. Oct 27 08:23:56.747231 tar[1577]: linux-amd64/README.md Oct 27 08:23:56.770797 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 27 08:23:57.427003 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:23:57.430883 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 27 08:23:57.435388 systemd[1]: Startup finished in 2.463s (kernel) + 4.925s (initrd) + 5.804s (userspace) = 13.193s. Oct 27 08:23:57.437413 (kubelet)[1728]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 27 08:23:58.099783 kubelet[1728]: E1027 08:23:58.099713 1728 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 27 08:23:58.103178 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 27 08:23:58.103411 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 27 08:23:58.104010 systemd[1]: kubelet.service: Consumed 1.210s CPU time, 262.2M memory peak. Oct 27 08:23:58.581655 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 27 08:23:58.583317 systemd[1]: Started sshd@0-64.23.205.201:22-139.178.89.65:34816.service - OpenSSH per-connection server daemon (139.178.89.65:34816). Oct 27 08:23:58.699011 sshd[1740]: Accepted publickey for core from 139.178.89.65 port 34816 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:23:58.701446 sshd-session[1740]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:23:58.710463 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 27 08:23:58.712187 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 27 08:23:58.721959 systemd-logind[1571]: New session 1 of user core. Oct 27 08:23:58.749804 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 27 08:23:58.754664 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 27 08:23:58.773358 (systemd)[1745]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 27 08:23:58.777217 systemd-logind[1571]: New session c1 of user core. Oct 27 08:23:58.938832 systemd[1745]: Queued start job for default target default.target. Oct 27 08:23:58.962046 systemd[1745]: Created slice app.slice - User Application Slice. Oct 27 08:23:58.962089 systemd[1745]: Reached target paths.target - Paths. Oct 27 08:23:58.962160 systemd[1745]: Reached target timers.target - Timers. Oct 27 08:23:58.964277 systemd[1745]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 27 08:23:59.000012 systemd[1745]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 27 08:23:59.000206 systemd[1745]: Reached target sockets.target - Sockets. Oct 27 08:23:59.000284 systemd[1745]: Reached target basic.target - Basic System. Oct 27 08:23:59.000336 systemd[1745]: Reached target default.target - Main User Target. Oct 27 08:23:59.000383 systemd[1745]: Startup finished in 213ms. Oct 27 08:23:59.000488 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 27 08:23:59.008974 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 27 08:23:59.081978 systemd[1]: Started sshd@1-64.23.205.201:22-139.178.89.65:34830.service - OpenSSH per-connection server daemon (139.178.89.65:34830). Oct 27 08:23:59.145026 sshd[1756]: Accepted publickey for core from 139.178.89.65 port 34830 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:23:59.146972 sshd-session[1756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:23:59.153521 systemd-logind[1571]: New session 2 of user core. Oct 27 08:23:59.170043 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 27 08:23:59.233066 sshd[1759]: Connection closed by 139.178.89.65 port 34830 Oct 27 08:23:59.233885 sshd-session[1756]: pam_unix(sshd:session): session closed for user core Oct 27 08:23:59.243707 systemd[1]: sshd@1-64.23.205.201:22-139.178.89.65:34830.service: Deactivated successfully. Oct 27 08:23:59.245918 systemd[1]: session-2.scope: Deactivated successfully. Oct 27 08:23:59.247546 systemd-logind[1571]: Session 2 logged out. Waiting for processes to exit. Oct 27 08:23:59.249980 systemd[1]: Started sshd@2-64.23.205.201:22-139.178.89.65:34836.service - OpenSSH per-connection server daemon (139.178.89.65:34836). Oct 27 08:23:59.252700 systemd-logind[1571]: Removed session 2. Oct 27 08:23:59.308275 sshd[1765]: Accepted publickey for core from 139.178.89.65 port 34836 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:23:59.309906 sshd-session[1765]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:23:59.315320 systemd-logind[1571]: New session 3 of user core. Oct 27 08:23:59.329961 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 27 08:23:59.385766 sshd[1768]: Connection closed by 139.178.89.65 port 34836 Oct 27 08:23:59.386405 sshd-session[1765]: pam_unix(sshd:session): session closed for user core Oct 27 08:23:59.402071 systemd[1]: sshd@2-64.23.205.201:22-139.178.89.65:34836.service: Deactivated successfully. Oct 27 08:23:59.404228 systemd[1]: session-3.scope: Deactivated successfully. Oct 27 08:23:59.405423 systemd-logind[1571]: Session 3 logged out. Waiting for processes to exit. Oct 27 08:23:59.409457 systemd[1]: Started sshd@3-64.23.205.201:22-139.178.89.65:34844.service - OpenSSH per-connection server daemon (139.178.89.65:34844). Oct 27 08:23:59.410466 systemd-logind[1571]: Removed session 3. Oct 27 08:23:59.492359 sshd[1774]: Accepted publickey for core from 139.178.89.65 port 34844 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:23:59.494457 sshd-session[1774]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:23:59.500971 systemd-logind[1571]: New session 4 of user core. Oct 27 08:23:59.507892 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 27 08:23:59.571044 sshd[1777]: Connection closed by 139.178.89.65 port 34844 Oct 27 08:23:59.571978 sshd-session[1774]: pam_unix(sshd:session): session closed for user core Oct 27 08:23:59.584386 systemd[1]: sshd@3-64.23.205.201:22-139.178.89.65:34844.service: Deactivated successfully. Oct 27 08:23:59.588054 systemd[1]: session-4.scope: Deactivated successfully. Oct 27 08:23:59.589901 systemd-logind[1571]: Session 4 logged out. Waiting for processes to exit. Oct 27 08:23:59.594066 systemd[1]: Started sshd@4-64.23.205.201:22-139.178.89.65:34846.service - OpenSSH per-connection server daemon (139.178.89.65:34846). Oct 27 08:23:59.595815 systemd-logind[1571]: Removed session 4. Oct 27 08:23:59.668659 sshd[1783]: Accepted publickey for core from 139.178.89.65 port 34846 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:23:59.671006 sshd-session[1783]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:23:59.678085 systemd-logind[1571]: New session 5 of user core. Oct 27 08:23:59.696977 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 27 08:23:59.780104 sudo[1787]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 27 08:23:59.780983 sudo[1787]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 27 08:23:59.798486 sudo[1787]: pam_unix(sudo:session): session closed for user root Oct 27 08:23:59.803274 sshd[1786]: Connection closed by 139.178.89.65 port 34846 Oct 27 08:23:59.804257 sshd-session[1783]: pam_unix(sshd:session): session closed for user core Oct 27 08:23:59.819331 systemd[1]: sshd@4-64.23.205.201:22-139.178.89.65:34846.service: Deactivated successfully. Oct 27 08:23:59.822391 systemd[1]: session-5.scope: Deactivated successfully. Oct 27 08:23:59.823762 systemd-logind[1571]: Session 5 logged out. Waiting for processes to exit. Oct 27 08:23:59.829905 systemd[1]: Started sshd@5-64.23.205.201:22-139.178.89.65:34858.service - OpenSSH per-connection server daemon (139.178.89.65:34858). Oct 27 08:23:59.831699 systemd-logind[1571]: Removed session 5. Oct 27 08:23:59.910693 sshd[1793]: Accepted publickey for core from 139.178.89.65 port 34858 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:23:59.913535 sshd-session[1793]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:23:59.922453 systemd-logind[1571]: New session 6 of user core. Oct 27 08:23:59.930032 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 27 08:23:59.998105 sudo[1798]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 27 08:23:59.998489 sudo[1798]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 27 08:24:00.005382 sudo[1798]: pam_unix(sudo:session): session closed for user root Oct 27 08:24:00.016012 sudo[1797]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Oct 27 08:24:00.016411 sudo[1797]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 27 08:24:00.031948 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 27 08:24:00.100328 augenrules[1820]: No rules Oct 27 08:24:00.101994 systemd[1]: audit-rules.service: Deactivated successfully. Oct 27 08:24:00.102289 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 27 08:24:00.104048 sudo[1797]: pam_unix(sudo:session): session closed for user root Oct 27 08:24:00.107753 sshd[1796]: Connection closed by 139.178.89.65 port 34858 Oct 27 08:24:00.108525 sshd-session[1793]: pam_unix(sshd:session): session closed for user core Oct 27 08:24:00.123175 systemd[1]: sshd@5-64.23.205.201:22-139.178.89.65:34858.service: Deactivated successfully. Oct 27 08:24:00.126187 systemd[1]: session-6.scope: Deactivated successfully. Oct 27 08:24:00.127696 systemd-logind[1571]: Session 6 logged out. Waiting for processes to exit. Oct 27 08:24:00.132685 systemd[1]: Started sshd@6-64.23.205.201:22-139.178.89.65:34868.service - OpenSSH per-connection server daemon (139.178.89.65:34868). Oct 27 08:24:00.134117 systemd-logind[1571]: Removed session 6. Oct 27 08:24:00.206436 sshd[1829]: Accepted publickey for core from 139.178.89.65 port 34868 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:24:00.209253 sshd-session[1829]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:24:00.218384 systemd-logind[1571]: New session 7 of user core. Oct 27 08:24:00.225988 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 27 08:24:00.294268 sudo[1833]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 27 08:24:00.296113 sudo[1833]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 27 08:24:00.978214 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 27 08:24:00.998475 (dockerd)[1850]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 27 08:24:01.447622 dockerd[1850]: time="2025-10-27T08:24:01.447262819Z" level=info msg="Starting up" Oct 27 08:24:01.450523 dockerd[1850]: time="2025-10-27T08:24:01.450474294Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Oct 27 08:24:01.474764 dockerd[1850]: time="2025-10-27T08:24:01.474663734Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Oct 27 08:24:01.593721 dockerd[1850]: time="2025-10-27T08:24:01.593551129Z" level=info msg="Loading containers: start." Oct 27 08:24:01.611144 kernel: Initializing XFRM netlink socket Oct 27 08:24:01.960654 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:24:01.971404 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:24:02.024876 systemd-networkd[1494]: docker0: Link UP Oct 27 08:24:02.025253 systemd-timesyncd[1463]: Network configuration changed, trying to establish connection. Oct 27 08:24:02.027979 dockerd[1850]: time="2025-10-27T08:24:02.027855037Z" level=info msg="Loading containers: done." Oct 27 08:24:02.050686 dockerd[1850]: time="2025-10-27T08:24:02.050170051Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 27 08:24:02.051942 dockerd[1850]: time="2025-10-27T08:24:02.051815970Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Oct 27 08:24:02.052064 dockerd[1850]: time="2025-10-27T08:24:02.052019528Z" level=info msg="Initializing buildkit" Oct 27 08:24:02.054452 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3823743213-merged.mount: Deactivated successfully. Oct 27 08:24:02.086907 dockerd[1850]: time="2025-10-27T08:24:02.086830058Z" level=info msg="Completed buildkit initialization" Oct 27 08:24:02.098641 dockerd[1850]: time="2025-10-27T08:24:02.097725310Z" level=info msg="Daemon has completed initialization" Oct 27 08:24:02.098641 dockerd[1850]: time="2025-10-27T08:24:02.097856881Z" level=info msg="API listen on /run/docker.sock" Oct 27 08:24:02.099098 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 27 08:24:03.063323 containerd[1602]: time="2025-10-27T08:24:03.063268644Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Oct 27 08:24:03.680541 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount565160255.mount: Deactivated successfully. Oct 27 08:24:04.893826 containerd[1602]: time="2025-10-27T08:24:04.893752291Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:04.894821 containerd[1602]: time="2025-10-27T08:24:04.894782283Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Oct 27 08:24:04.895697 containerd[1602]: time="2025-10-27T08:24:04.895314117Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:04.897632 containerd[1602]: time="2025-10-27T08:24:04.897569851Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:04.898528 containerd[1602]: time="2025-10-27T08:24:04.898500149Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 1.835181257s" Oct 27 08:24:04.898652 containerd[1602]: time="2025-10-27T08:24:04.898634346Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Oct 27 08:24:04.899334 containerd[1602]: time="2025-10-27T08:24:04.899249924Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Oct 27 08:24:06.523173 containerd[1602]: time="2025-10-27T08:24:06.523065578Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:06.525470 containerd[1602]: time="2025-10-27T08:24:06.525129330Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Oct 27 08:24:06.526120 containerd[1602]: time="2025-10-27T08:24:06.526080081Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:06.529172 containerd[1602]: time="2025-10-27T08:24:06.529067368Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:06.530370 containerd[1602]: time="2025-10-27T08:24:06.530334072Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.63089297s" Oct 27 08:24:06.530484 containerd[1602]: time="2025-10-27T08:24:06.530471247Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Oct 27 08:24:06.531715 containerd[1602]: time="2025-10-27T08:24:06.531670531Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Oct 27 08:24:07.692276 containerd[1602]: time="2025-10-27T08:24:07.692193003Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:07.693843 containerd[1602]: time="2025-10-27T08:24:07.693792367Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Oct 27 08:24:07.695813 containerd[1602]: time="2025-10-27T08:24:07.694956910Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:07.700243 containerd[1602]: time="2025-10-27T08:24:07.700192810Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:07.701490 containerd[1602]: time="2025-10-27T08:24:07.701434368Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 1.169709929s" Oct 27 08:24:07.701681 containerd[1602]: time="2025-10-27T08:24:07.701665002Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Oct 27 08:24:07.702302 containerd[1602]: time="2025-10-27T08:24:07.702265725Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Oct 27 08:24:08.320855 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 27 08:24:08.324850 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 27 08:24:08.518806 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:24:08.529690 (kubelet)[2146]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 27 08:24:08.606732 kubelet[2146]: E1027 08:24:08.605822 2146 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 27 08:24:08.614751 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 27 08:24:08.615790 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 27 08:24:08.616798 systemd[1]: kubelet.service: Consumed 213ms CPU time, 110M memory peak. Oct 27 08:24:08.953711 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount292475432.mount: Deactivated successfully. Oct 27 08:24:09.593756 containerd[1602]: time="2025-10-27T08:24:09.593697924Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:09.595087 containerd[1602]: time="2025-10-27T08:24:09.595044842Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Oct 27 08:24:09.596635 containerd[1602]: time="2025-10-27T08:24:09.595939516Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:09.597399 containerd[1602]: time="2025-10-27T08:24:09.597360265Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:09.598251 containerd[1602]: time="2025-10-27T08:24:09.598211312Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 1.89590582s" Oct 27 08:24:09.598393 containerd[1602]: time="2025-10-27T08:24:09.598369250Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Oct 27 08:24:09.599044 containerd[1602]: time="2025-10-27T08:24:09.599017568Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Oct 27 08:24:09.600413 systemd-resolved[1286]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Oct 27 08:24:10.178423 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2544162020.mount: Deactivated successfully. Oct 27 08:24:10.979630 containerd[1602]: time="2025-10-27T08:24:10.978815766Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:10.979630 containerd[1602]: time="2025-10-27T08:24:10.979499651Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Oct 27 08:24:10.980050 containerd[1602]: time="2025-10-27T08:24:10.979779148Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:10.983085 containerd[1602]: time="2025-10-27T08:24:10.983021606Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:10.984614 containerd[1602]: time="2025-10-27T08:24:10.984447358Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.385289805s" Oct 27 08:24:10.984614 containerd[1602]: time="2025-10-27T08:24:10.984487613Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Oct 27 08:24:10.985310 containerd[1602]: time="2025-10-27T08:24:10.985178452Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Oct 27 08:24:11.438539 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1434783309.mount: Deactivated successfully. Oct 27 08:24:11.444053 containerd[1602]: time="2025-10-27T08:24:11.443989235Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 27 08:24:11.444829 containerd[1602]: time="2025-10-27T08:24:11.444802466Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Oct 27 08:24:11.445390 containerd[1602]: time="2025-10-27T08:24:11.445359220Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 27 08:24:11.449487 containerd[1602]: time="2025-10-27T08:24:11.449430396Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 27 08:24:11.451671 containerd[1602]: time="2025-10-27T08:24:11.451511038Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 466.290489ms" Oct 27 08:24:11.451671 containerd[1602]: time="2025-10-27T08:24:11.451552075Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Oct 27 08:24:11.452310 containerd[1602]: time="2025-10-27T08:24:11.452096573Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Oct 27 08:24:11.903461 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1189654497.mount: Deactivated successfully. Oct 27 08:24:12.662816 systemd-resolved[1286]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Oct 27 08:24:13.674298 containerd[1602]: time="2025-10-27T08:24:13.672966862Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:13.674298 containerd[1602]: time="2025-10-27T08:24:13.674148173Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Oct 27 08:24:13.674298 containerd[1602]: time="2025-10-27T08:24:13.674244355Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:13.677871 containerd[1602]: time="2025-10-27T08:24:13.677807042Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:13.679724 containerd[1602]: time="2025-10-27T08:24:13.679654774Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.227527195s" Oct 27 08:24:13.679724 containerd[1602]: time="2025-10-27T08:24:13.679721708Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Oct 27 08:24:17.590509 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:24:17.591058 systemd[1]: kubelet.service: Consumed 213ms CPU time, 110M memory peak. Oct 27 08:24:17.596164 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 27 08:24:17.630346 systemd[1]: Reload requested from client PID 2294 ('systemctl') (unit session-7.scope)... Oct 27 08:24:17.630534 systemd[1]: Reloading... Oct 27 08:24:17.782620 zram_generator::config[2338]: No configuration found. Oct 27 08:24:18.035017 systemd[1]: Reloading finished in 403 ms. Oct 27 08:24:18.103406 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 27 08:24:18.103729 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 27 08:24:18.104166 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:24:18.104310 systemd[1]: kubelet.service: Consumed 121ms CPU time, 97.9M memory peak. Oct 27 08:24:18.106377 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 27 08:24:18.287889 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:24:18.301367 (kubelet)[2392]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 27 08:24:18.357904 kubelet[2392]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 27 08:24:18.358288 kubelet[2392]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 27 08:24:18.358339 kubelet[2392]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 27 08:24:18.358500 kubelet[2392]: I1027 08:24:18.358471 2392 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 27 08:24:19.014620 kubelet[2392]: I1027 08:24:19.013454 2392 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 27 08:24:19.014620 kubelet[2392]: I1027 08:24:19.013507 2392 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 27 08:24:19.014620 kubelet[2392]: I1027 08:24:19.014034 2392 server.go:954] "Client rotation is on, will bootstrap in background" Oct 27 08:24:19.049923 kubelet[2392]: E1027 08:24:19.049849 2392 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://64.23.205.201:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:19.051274 kubelet[2392]: I1027 08:24:19.051198 2392 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 27 08:24:19.069942 kubelet[2392]: I1027 08:24:19.069903 2392 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 27 08:24:19.080372 kubelet[2392]: I1027 08:24:19.080010 2392 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 27 08:24:19.085788 kubelet[2392]: I1027 08:24:19.085687 2392 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 27 08:24:19.086129 kubelet[2392]: I1027 08:24:19.085937 2392 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-9999.9.9-k-4f7b9e69b1","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 27 08:24:19.088175 kubelet[2392]: I1027 08:24:19.088134 2392 topology_manager.go:138] "Creating topology manager with none policy" Oct 27 08:24:19.088602 kubelet[2392]: I1027 08:24:19.088345 2392 container_manager_linux.go:304] "Creating device plugin manager" Oct 27 08:24:19.089874 kubelet[2392]: I1027 08:24:19.089842 2392 state_mem.go:36] "Initialized new in-memory state store" Oct 27 08:24:19.093625 kubelet[2392]: I1027 08:24:19.093480 2392 kubelet.go:446] "Attempting to sync node with API server" Oct 27 08:24:19.093625 kubelet[2392]: I1027 08:24:19.093524 2392 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 27 08:24:19.093625 kubelet[2392]: I1027 08:24:19.093553 2392 kubelet.go:352] "Adding apiserver pod source" Oct 27 08:24:19.093625 kubelet[2392]: I1027 08:24:19.093566 2392 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 27 08:24:19.100309 kubelet[2392]: W1027 08:24:19.099411 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://64.23.205.201:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-9999.9.9-k-4f7b9e69b1&limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:19.100309 kubelet[2392]: E1027 08:24:19.099481 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://64.23.205.201:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-9999.9.9-k-4f7b9e69b1&limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:19.100309 kubelet[2392]: W1027 08:24:19.099937 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://64.23.205.201:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:19.100309 kubelet[2392]: E1027 08:24:19.099980 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://64.23.205.201:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:19.102078 kubelet[2392]: I1027 08:24:19.102040 2392 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 27 08:24:19.107661 kubelet[2392]: I1027 08:24:19.107629 2392 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 27 08:24:19.108875 kubelet[2392]: W1027 08:24:19.108839 2392 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 27 08:24:19.113387 kubelet[2392]: I1027 08:24:19.113345 2392 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 27 08:24:19.113387 kubelet[2392]: I1027 08:24:19.113392 2392 server.go:1287] "Started kubelet" Oct 27 08:24:19.114615 kubelet[2392]: I1027 08:24:19.113634 2392 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 27 08:24:19.115301 kubelet[2392]: I1027 08:24:19.115275 2392 server.go:479] "Adding debug handlers to kubelet server" Oct 27 08:24:19.121358 kubelet[2392]: I1027 08:24:19.121283 2392 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 27 08:24:19.121617 kubelet[2392]: I1027 08:24:19.121601 2392 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 27 08:24:19.123897 kubelet[2392]: I1027 08:24:19.123699 2392 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 27 08:24:19.126211 kubelet[2392]: E1027 08:24:19.122771 2392 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://64.23.205.201:6443/api/v1/namespaces/default/events\": dial tcp 64.23.205.201:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-9999.9.9-k-4f7b9e69b1.18724b89b750151b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-9999.9.9-k-4f7b9e69b1,UID:ci-9999.9.9-k-4f7b9e69b1,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-9999.9.9-k-4f7b9e69b1,},FirstTimestamp:2025-10-27 08:24:19.113366811 +0000 UTC m=+0.806218334,LastTimestamp:2025-10-27 08:24:19.113366811 +0000 UTC m=+0.806218334,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-9999.9.9-k-4f7b9e69b1,}" Oct 27 08:24:19.127285 kubelet[2392]: I1027 08:24:19.126700 2392 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 27 08:24:19.130413 kubelet[2392]: E1027 08:24:19.128991 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:19.130413 kubelet[2392]: I1027 08:24:19.129074 2392 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 27 08:24:19.130413 kubelet[2392]: I1027 08:24:19.129338 2392 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 27 08:24:19.130413 kubelet[2392]: I1027 08:24:19.129441 2392 reconciler.go:26] "Reconciler: start to sync state" Oct 27 08:24:19.130413 kubelet[2392]: W1027 08:24:19.129844 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://64.23.205.201:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:19.130413 kubelet[2392]: E1027 08:24:19.129888 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://64.23.205.201:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:19.130413 kubelet[2392]: E1027 08:24:19.130117 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.205.201:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-9999.9.9-k-4f7b9e69b1?timeout=10s\": dial tcp 64.23.205.201:6443: connect: connection refused" interval="200ms" Oct 27 08:24:19.138320 kubelet[2392]: I1027 08:24:19.138294 2392 factory.go:221] Registration of the containerd container factory successfully Oct 27 08:24:19.138457 kubelet[2392]: I1027 08:24:19.138446 2392 factory.go:221] Registration of the systemd container factory successfully Oct 27 08:24:19.138688 kubelet[2392]: I1027 08:24:19.138664 2392 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 27 08:24:19.155819 kubelet[2392]: I1027 08:24:19.155768 2392 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 27 08:24:19.157347 kubelet[2392]: I1027 08:24:19.157310 2392 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 27 08:24:19.157507 kubelet[2392]: I1027 08:24:19.157495 2392 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 27 08:24:19.157626 kubelet[2392]: I1027 08:24:19.157612 2392 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 27 08:24:19.157699 kubelet[2392]: I1027 08:24:19.157691 2392 kubelet.go:2382] "Starting kubelet main sync loop" Oct 27 08:24:19.157841 kubelet[2392]: E1027 08:24:19.157818 2392 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 27 08:24:19.162998 kubelet[2392]: W1027 08:24:19.162938 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://64.23.205.201:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:19.165714 kubelet[2392]: E1027 08:24:19.165679 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://64.23.205.201:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:19.169635 kubelet[2392]: I1027 08:24:19.169604 2392 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 27 08:24:19.169635 kubelet[2392]: I1027 08:24:19.169621 2392 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 27 08:24:19.169821 kubelet[2392]: I1027 08:24:19.169654 2392 state_mem.go:36] "Initialized new in-memory state store" Oct 27 08:24:19.170667 kubelet[2392]: I1027 08:24:19.170647 2392 policy_none.go:49] "None policy: Start" Oct 27 08:24:19.170667 kubelet[2392]: I1027 08:24:19.170667 2392 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 27 08:24:19.170784 kubelet[2392]: I1027 08:24:19.170678 2392 state_mem.go:35] "Initializing new in-memory state store" Oct 27 08:24:19.176515 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Oct 27 08:24:19.191130 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Oct 27 08:24:19.195882 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Oct 27 08:24:19.205272 kubelet[2392]: I1027 08:24:19.205065 2392 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 27 08:24:19.205763 kubelet[2392]: I1027 08:24:19.205749 2392 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 27 08:24:19.206240 kubelet[2392]: I1027 08:24:19.206196 2392 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 27 08:24:19.206984 kubelet[2392]: I1027 08:24:19.206890 2392 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 27 08:24:19.208969 kubelet[2392]: E1027 08:24:19.208618 2392 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 27 08:24:19.208969 kubelet[2392]: E1027 08:24:19.208662 2392 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:19.267957 systemd[1]: Created slice kubepods-burstable-pod570fe50a6da5294452f1f27d07f761b8.slice - libcontainer container kubepods-burstable-pod570fe50a6da5294452f1f27d07f761b8.slice. Oct 27 08:24:19.286437 kubelet[2392]: E1027 08:24:19.286406 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.289555 systemd[1]: Created slice kubepods-burstable-pod7f3336d3ffb5159bf5d091da7e24649b.slice - libcontainer container kubepods-burstable-pod7f3336d3ffb5159bf5d091da7e24649b.slice. Oct 27 08:24:19.301432 kubelet[2392]: E1027 08:24:19.301386 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.304504 systemd[1]: Created slice kubepods-burstable-pod0f5c5cae72b141360a951471b3d00ecb.slice - libcontainer container kubepods-burstable-pod0f5c5cae72b141360a951471b3d00ecb.slice. Oct 27 08:24:19.309489 kubelet[2392]: I1027 08:24:19.308475 2392 kubelet_node_status.go:75] "Attempting to register node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.309489 kubelet[2392]: E1027 08:24:19.308870 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://64.23.205.201:6443/api/v1/nodes\": dial tcp 64.23.205.201:6443: connect: connection refused" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.310790 kubelet[2392]: E1027 08:24:19.310757 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.331328 kubelet[2392]: E1027 08:24:19.331263 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.205.201:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-9999.9.9-k-4f7b9e69b1?timeout=10s\": dial tcp 64.23.205.201:6443: connect: connection refused" interval="400ms" Oct 27 08:24:19.430817 kubelet[2392]: I1027 08:24:19.430750 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/570fe50a6da5294452f1f27d07f761b8-usr-share-ca-certificates\") pod \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"570fe50a6da5294452f1f27d07f761b8\") " pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.430817 kubelet[2392]: I1027 08:24:19.430797 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-k8s-certs\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.430817 kubelet[2392]: I1027 08:24:19.430822 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-kubeconfig\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.431500 kubelet[2392]: I1027 08:24:19.430841 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.431500 kubelet[2392]: I1027 08:24:19.430862 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7f3336d3ffb5159bf5d091da7e24649b-kubeconfig\") pod \"kube-scheduler-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"7f3336d3ffb5159bf5d091da7e24649b\") " pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.431500 kubelet[2392]: I1027 08:24:19.430880 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/570fe50a6da5294452f1f27d07f761b8-ca-certs\") pod \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"570fe50a6da5294452f1f27d07f761b8\") " pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.431500 kubelet[2392]: I1027 08:24:19.430898 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/570fe50a6da5294452f1f27d07f761b8-k8s-certs\") pod \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"570fe50a6da5294452f1f27d07f761b8\") " pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.431500 kubelet[2392]: I1027 08:24:19.430912 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-ca-certs\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.431758 kubelet[2392]: I1027 08:24:19.430929 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-flexvolume-dir\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.511706 kubelet[2392]: I1027 08:24:19.511262 2392 kubelet_node_status.go:75] "Attempting to register node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.512061 kubelet[2392]: E1027 08:24:19.512027 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://64.23.205.201:6443/api/v1/nodes\": dial tcp 64.23.205.201:6443: connect: connection refused" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.588704 kubelet[2392]: E1027 08:24:19.588073 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:19.589103 containerd[1602]: time="2025-10-27T08:24:19.588821143Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-9999.9.9-k-4f7b9e69b1,Uid:570fe50a6da5294452f1f27d07f761b8,Namespace:kube-system,Attempt:0,}" Oct 27 08:24:19.602460 kubelet[2392]: E1027 08:24:19.602427 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:19.604176 containerd[1602]: time="2025-10-27T08:24:19.602991748Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-9999.9.9-k-4f7b9e69b1,Uid:7f3336d3ffb5159bf5d091da7e24649b,Namespace:kube-system,Attempt:0,}" Oct 27 08:24:19.612267 kubelet[2392]: E1027 08:24:19.612219 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:19.623255 containerd[1602]: time="2025-10-27T08:24:19.623202570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1,Uid:0f5c5cae72b141360a951471b3d00ecb,Namespace:kube-system,Attempt:0,}" Oct 27 08:24:19.714093 containerd[1602]: time="2025-10-27T08:24:19.714041209Z" level=info msg="connecting to shim a041221c4c810df0bbe3f1cfec57749eec286cbc44f6392c99893acd67588680" address="unix:///run/containerd/s/017677023aefdea1d1bbc18692e483000bee98110d3885dbb34045842a4301a0" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:19.714531 containerd[1602]: time="2025-10-27T08:24:19.714465994Z" level=info msg="connecting to shim d49023a4179c097057e4db6a38d4d4bc47038edde60bc2356687ee21590e06eb" address="unix:///run/containerd/s/9ca95aafa849023232c69a5463090c5e47561f6fb121e85eb80c8b56325bee80" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:19.718259 containerd[1602]: time="2025-10-27T08:24:19.718176791Z" level=info msg="connecting to shim 8232b9768024bf0476c3dafec1924138fd9638bd367e8ff39d8bb09706737508" address="unix:///run/containerd/s/2882f579f1adfd45d2d38865f90555399dd706148c5729cb4dfe5a57bb1b80ea" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:19.734092 kubelet[2392]: E1027 08:24:19.733947 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.205.201:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-9999.9.9-k-4f7b9e69b1?timeout=10s\": dial tcp 64.23.205.201:6443: connect: connection refused" interval="800ms" Oct 27 08:24:19.838905 systemd[1]: Started cri-containerd-8232b9768024bf0476c3dafec1924138fd9638bd367e8ff39d8bb09706737508.scope - libcontainer container 8232b9768024bf0476c3dafec1924138fd9638bd367e8ff39d8bb09706737508. Oct 27 08:24:19.842792 systemd[1]: Started cri-containerd-a041221c4c810df0bbe3f1cfec57749eec286cbc44f6392c99893acd67588680.scope - libcontainer container a041221c4c810df0bbe3f1cfec57749eec286cbc44f6392c99893acd67588680. Oct 27 08:24:19.844562 systemd[1]: Started cri-containerd-d49023a4179c097057e4db6a38d4d4bc47038edde60bc2356687ee21590e06eb.scope - libcontainer container d49023a4179c097057e4db6a38d4d4bc47038edde60bc2356687ee21590e06eb. Oct 27 08:24:19.916027 kubelet[2392]: I1027 08:24:19.915961 2392 kubelet_node_status.go:75] "Attempting to register node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.919772 kubelet[2392]: E1027 08:24:19.919717 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://64.23.205.201:6443/api/v1/nodes\": dial tcp 64.23.205.201:6443: connect: connection refused" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:19.935056 kubelet[2392]: W1027 08:24:19.934227 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://64.23.205.201:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:19.935056 kubelet[2392]: E1027 08:24:19.934315 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://64.23.205.201:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:19.970398 containerd[1602]: time="2025-10-27T08:24:19.970354940Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-9999.9.9-k-4f7b9e69b1,Uid:570fe50a6da5294452f1f27d07f761b8,Namespace:kube-system,Attempt:0,} returns sandbox id \"a041221c4c810df0bbe3f1cfec57749eec286cbc44f6392c99893acd67588680\"" Oct 27 08:24:19.971028 containerd[1602]: time="2025-10-27T08:24:19.970856674Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1,Uid:0f5c5cae72b141360a951471b3d00ecb,Namespace:kube-system,Attempt:0,} returns sandbox id \"8232b9768024bf0476c3dafec1924138fd9638bd367e8ff39d8bb09706737508\"" Oct 27 08:24:19.973089 kubelet[2392]: E1027 08:24:19.973052 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:19.973510 kubelet[2392]: E1027 08:24:19.973432 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:19.978383 containerd[1602]: time="2025-10-27T08:24:19.977974918Z" level=info msg="CreateContainer within sandbox \"a041221c4c810df0bbe3f1cfec57749eec286cbc44f6392c99893acd67588680\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 27 08:24:19.978821 containerd[1602]: time="2025-10-27T08:24:19.978793414Z" level=info msg="CreateContainer within sandbox \"8232b9768024bf0476c3dafec1924138fd9638bd367e8ff39d8bb09706737508\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 27 08:24:19.989861 containerd[1602]: time="2025-10-27T08:24:19.989815752Z" level=info msg="Container 264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:19.998544 containerd[1602]: time="2025-10-27T08:24:19.998471758Z" level=info msg="Container 25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:20.002875 containerd[1602]: time="2025-10-27T08:24:20.002819061Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-9999.9.9-k-4f7b9e69b1,Uid:7f3336d3ffb5159bf5d091da7e24649b,Namespace:kube-system,Attempt:0,} returns sandbox id \"d49023a4179c097057e4db6a38d4d4bc47038edde60bc2356687ee21590e06eb\"" Oct 27 08:24:20.006015 kubelet[2392]: E1027 08:24:20.005763 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:20.011082 containerd[1602]: time="2025-10-27T08:24:20.011020300Z" level=info msg="CreateContainer within sandbox \"d49023a4179c097057e4db6a38d4d4bc47038edde60bc2356687ee21590e06eb\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 27 08:24:20.015524 containerd[1602]: time="2025-10-27T08:24:20.015245095Z" level=info msg="CreateContainer within sandbox \"a041221c4c810df0bbe3f1cfec57749eec286cbc44f6392c99893acd67588680\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d\"" Oct 27 08:24:20.016725 containerd[1602]: time="2025-10-27T08:24:20.016685855Z" level=info msg="StartContainer for \"264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d\"" Oct 27 08:24:20.018291 containerd[1602]: time="2025-10-27T08:24:20.018242364Z" level=info msg="connecting to shim 264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d" address="unix:///run/containerd/s/017677023aefdea1d1bbc18692e483000bee98110d3885dbb34045842a4301a0" protocol=ttrpc version=3 Oct 27 08:24:20.020606 containerd[1602]: time="2025-10-27T08:24:20.020227962Z" level=info msg="CreateContainer within sandbox \"8232b9768024bf0476c3dafec1924138fd9638bd367e8ff39d8bb09706737508\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407\"" Oct 27 08:24:20.022001 containerd[1602]: time="2025-10-27T08:24:20.021957595Z" level=info msg="StartContainer for \"25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407\"" Oct 27 08:24:20.024133 containerd[1602]: time="2025-10-27T08:24:20.023989162Z" level=info msg="connecting to shim 25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407" address="unix:///run/containerd/s/2882f579f1adfd45d2d38865f90555399dd706148c5729cb4dfe5a57bb1b80ea" protocol=ttrpc version=3 Oct 27 08:24:20.026445 kubelet[2392]: W1027 08:24:20.026247 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://64.23.205.201:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-9999.9.9-k-4f7b9e69b1&limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:20.026445 kubelet[2392]: E1027 08:24:20.026390 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://64.23.205.201:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-9999.9.9-k-4f7b9e69b1&limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:20.028472 containerd[1602]: time="2025-10-27T08:24:20.028408720Z" level=info msg="Container 945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:20.043244 containerd[1602]: time="2025-10-27T08:24:20.043116532Z" level=info msg="CreateContainer within sandbox \"d49023a4179c097057e4db6a38d4d4bc47038edde60bc2356687ee21590e06eb\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330\"" Oct 27 08:24:20.044537 containerd[1602]: time="2025-10-27T08:24:20.044501052Z" level=info msg="StartContainer for \"945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330\"" Oct 27 08:24:20.046574 containerd[1602]: time="2025-10-27T08:24:20.046493865Z" level=info msg="connecting to shim 945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330" address="unix:///run/containerd/s/9ca95aafa849023232c69a5463090c5e47561f6fb121e85eb80c8b56325bee80" protocol=ttrpc version=3 Oct 27 08:24:20.055085 systemd[1]: Started cri-containerd-264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d.scope - libcontainer container 264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d. Oct 27 08:24:20.070218 kubelet[2392]: W1027 08:24:20.070081 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://64.23.205.201:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:20.070501 kubelet[2392]: E1027 08:24:20.070413 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://64.23.205.201:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:20.074528 systemd[1]: Started cri-containerd-25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407.scope - libcontainer container 25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407. Oct 27 08:24:20.107944 systemd[1]: Started cri-containerd-945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330.scope - libcontainer container 945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330. Oct 27 08:24:20.159613 containerd[1602]: time="2025-10-27T08:24:20.159422931Z" level=info msg="StartContainer for \"264470290c5e38f7d39c60300d61f738c0bf2a650a531c9efb986b549384a97d\" returns successfully" Oct 27 08:24:20.191317 kubelet[2392]: E1027 08:24:20.191146 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:20.191541 kubelet[2392]: E1027 08:24:20.191486 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:20.199885 kubelet[2392]: W1027 08:24:20.199721 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://64.23.205.201:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.205.201:6443: connect: connection refused Oct 27 08:24:20.200135 kubelet[2392]: E1027 08:24:20.200085 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://64.23.205.201:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 64.23.205.201:6443: connect: connection refused" logger="UnhandledError" Oct 27 08:24:20.202300 containerd[1602]: time="2025-10-27T08:24:20.202176170Z" level=info msg="StartContainer for \"25b10f653733b42b64c4db83e2c9fca9efda1b5749f0c6a92e24ce693225c407\" returns successfully" Oct 27 08:24:20.247742 containerd[1602]: time="2025-10-27T08:24:20.247491838Z" level=info msg="StartContainer for \"945fef7d9e8fb92d8de0bcbf3f330380638d603e0af7711cd1160c8986800330\" returns successfully" Oct 27 08:24:20.723625 kubelet[2392]: I1027 08:24:20.723382 2392 kubelet_node_status.go:75] "Attempting to register node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:21.198352 kubelet[2392]: E1027 08:24:21.198262 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:21.203656 kubelet[2392]: E1027 08:24:21.203140 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:21.204940 kubelet[2392]: E1027 08:24:21.204908 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:21.205143 kubelet[2392]: E1027 08:24:21.205122 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:21.208632 kubelet[2392]: E1027 08:24:21.208565 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:21.208817 kubelet[2392]: E1027 08:24:21.208801 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:22.206181 kubelet[2392]: E1027 08:24:22.206135 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:22.206937 kubelet[2392]: E1027 08:24:22.206311 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:22.206937 kubelet[2392]: E1027 08:24:22.206910 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:22.207635 kubelet[2392]: E1027 08:24:22.207135 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:22.372425 kubelet[2392]: E1027 08:24:22.372384 2392 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:22.413150 kubelet[2392]: I1027 08:24:22.412881 2392 kubelet_node_status.go:78] "Successfully registered node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:22.413150 kubelet[2392]: E1027 08:24:22.412932 2392 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-9999.9.9-k-4f7b9e69b1\": node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:22.426167 kubelet[2392]: E1027 08:24:22.426115 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:22.527292 kubelet[2392]: E1027 08:24:22.526660 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:22.627802 kubelet[2392]: E1027 08:24:22.627741 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:22.728450 kubelet[2392]: E1027 08:24:22.728333 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:22.829263 kubelet[2392]: E1027 08:24:22.829195 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:22.930090 kubelet[2392]: E1027 08:24:22.930033 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:23.030797 kubelet[2392]: E1027 08:24:23.030712 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:23.131569 kubelet[2392]: E1027 08:24:23.131420 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" Oct 27 08:24:23.209980 kubelet[2392]: E1027 08:24:23.209870 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-9999.9.9-k-4f7b9e69b1\" not found" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.210975 kubelet[2392]: E1027 08:24:23.210236 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:23.230972 kubelet[2392]: I1027 08:24:23.230910 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.241618 kubelet[2392]: E1027 08:24:23.241534 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.241618 kubelet[2392]: I1027 08:24:23.241571 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.244547 kubelet[2392]: E1027 08:24:23.244507 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.244547 kubelet[2392]: I1027 08:24:23.244543 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.246752 kubelet[2392]: E1027 08:24:23.246713 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-9999.9.9-k-4f7b9e69b1\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.449717 kubelet[2392]: I1027 08:24:23.449566 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:23.459498 kubelet[2392]: W1027 08:24:23.459450 2392 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 27 08:24:23.459891 kubelet[2392]: E1027 08:24:23.459791 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:24.102014 kubelet[2392]: I1027 08:24:24.101710 2392 apiserver.go:52] "Watching apiserver" Oct 27 08:24:24.129609 kubelet[2392]: I1027 08:24:24.129499 2392 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 27 08:24:24.209161 kubelet[2392]: E1027 08:24:24.208954 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:24.775701 systemd[1]: Reload requested from client PID 2664 ('systemctl') (unit session-7.scope)... Oct 27 08:24:24.775736 systemd[1]: Reloading... Oct 27 08:24:24.915622 zram_generator::config[2711]: No configuration found. Oct 27 08:24:25.219440 systemd[1]: Reloading finished in 443 ms. Oct 27 08:24:25.255268 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 27 08:24:25.255509 kubelet[2392]: I1027 08:24:25.255479 2392 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 27 08:24:25.271207 systemd[1]: kubelet.service: Deactivated successfully. Oct 27 08:24:25.271660 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:24:25.271846 systemd[1]: kubelet.service: Consumed 1.228s CPU time, 126.4M memory peak. Oct 27 08:24:25.275136 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 27 08:24:25.472652 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 27 08:24:25.484903 (kubelet)[2760]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 27 08:24:25.559394 kubelet[2760]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 27 08:24:25.559394 kubelet[2760]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 27 08:24:25.559394 kubelet[2760]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 27 08:24:25.559911 kubelet[2760]: I1027 08:24:25.559482 2760 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 27 08:24:25.568513 kubelet[2760]: I1027 08:24:25.568469 2760 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 27 08:24:25.569184 kubelet[2760]: I1027 08:24:25.568699 2760 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 27 08:24:25.569640 kubelet[2760]: I1027 08:24:25.569623 2760 server.go:954] "Client rotation is on, will bootstrap in background" Oct 27 08:24:25.572834 kubelet[2760]: I1027 08:24:25.572804 2760 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 27 08:24:25.581194 kubelet[2760]: I1027 08:24:25.581134 2760 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 27 08:24:25.586091 kubelet[2760]: I1027 08:24:25.586065 2760 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 27 08:24:25.592910 kubelet[2760]: I1027 08:24:25.592876 2760 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 27 08:24:25.594460 kubelet[2760]: I1027 08:24:25.594382 2760 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 27 08:24:25.594731 kubelet[2760]: I1027 08:24:25.594444 2760 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-9999.9.9-k-4f7b9e69b1","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 27 08:24:25.594863 kubelet[2760]: I1027 08:24:25.594740 2760 topology_manager.go:138] "Creating topology manager with none policy" Oct 27 08:24:25.594863 kubelet[2760]: I1027 08:24:25.594755 2760 container_manager_linux.go:304] "Creating device plugin manager" Oct 27 08:24:25.594863 kubelet[2760]: I1027 08:24:25.594836 2760 state_mem.go:36] "Initialized new in-memory state store" Oct 27 08:24:25.595292 kubelet[2760]: I1027 08:24:25.595251 2760 kubelet.go:446] "Attempting to sync node with API server" Oct 27 08:24:25.596616 kubelet[2760]: I1027 08:24:25.596306 2760 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 27 08:24:25.596616 kubelet[2760]: I1027 08:24:25.596356 2760 kubelet.go:352] "Adding apiserver pod source" Oct 27 08:24:25.596616 kubelet[2760]: I1027 08:24:25.596369 2760 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 27 08:24:25.608620 kubelet[2760]: I1027 08:24:25.606918 2760 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 27 08:24:25.608620 kubelet[2760]: I1027 08:24:25.607496 2760 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 27 08:24:25.608842 kubelet[2760]: I1027 08:24:25.608828 2760 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 27 08:24:25.608915 kubelet[2760]: I1027 08:24:25.608909 2760 server.go:1287] "Started kubelet" Oct 27 08:24:25.613614 kubelet[2760]: I1027 08:24:25.611928 2760 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 27 08:24:25.616183 kubelet[2760]: I1027 08:24:25.616109 2760 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 27 08:24:25.616690 kubelet[2760]: I1027 08:24:25.616672 2760 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 27 08:24:25.616852 kubelet[2760]: I1027 08:24:25.616151 2760 server.go:479] "Adding debug handlers to kubelet server" Oct 27 08:24:25.617982 kubelet[2760]: I1027 08:24:25.617955 2760 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 27 08:24:25.625622 kubelet[2760]: I1027 08:24:25.625530 2760 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 27 08:24:25.628894 kubelet[2760]: I1027 08:24:25.628867 2760 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 27 08:24:25.629477 kubelet[2760]: I1027 08:24:25.629453 2760 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 27 08:24:25.629735 kubelet[2760]: I1027 08:24:25.629724 2760 reconciler.go:26] "Reconciler: start to sync state" Oct 27 08:24:25.632044 kubelet[2760]: I1027 08:24:25.632024 2760 factory.go:221] Registration of the systemd container factory successfully Oct 27 08:24:25.632258 kubelet[2760]: I1027 08:24:25.632240 2760 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 27 08:24:25.635693 kubelet[2760]: I1027 08:24:25.635640 2760 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 27 08:24:25.636431 kubelet[2760]: I1027 08:24:25.636254 2760 factory.go:221] Registration of the containerd container factory successfully Oct 27 08:24:25.637323 kubelet[2760]: I1027 08:24:25.636929 2760 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 27 08:24:25.637323 kubelet[2760]: I1027 08:24:25.636956 2760 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 27 08:24:25.637323 kubelet[2760]: I1027 08:24:25.636993 2760 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 27 08:24:25.637323 kubelet[2760]: I1027 08:24:25.637014 2760 kubelet.go:2382] "Starting kubelet main sync loop" Oct 27 08:24:25.637323 kubelet[2760]: E1027 08:24:25.637075 2760 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 27 08:24:25.643865 kubelet[2760]: E1027 08:24:25.643832 2760 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 27 08:24:25.701445 kubelet[2760]: I1027 08:24:25.701418 2760 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 27 08:24:25.701682 kubelet[2760]: I1027 08:24:25.701654 2760 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 27 08:24:25.701749 kubelet[2760]: I1027 08:24:25.701743 2760 state_mem.go:36] "Initialized new in-memory state store" Oct 27 08:24:25.701962 kubelet[2760]: I1027 08:24:25.701948 2760 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 27 08:24:25.702030 kubelet[2760]: I1027 08:24:25.702011 2760 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 27 08:24:25.702069 kubelet[2760]: I1027 08:24:25.702063 2760 policy_none.go:49] "None policy: Start" Oct 27 08:24:25.702128 kubelet[2760]: I1027 08:24:25.702121 2760 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 27 08:24:25.702170 kubelet[2760]: I1027 08:24:25.702164 2760 state_mem.go:35] "Initializing new in-memory state store" Oct 27 08:24:25.702323 kubelet[2760]: I1027 08:24:25.702313 2760 state_mem.go:75] "Updated machine memory state" Oct 27 08:24:25.709962 kubelet[2760]: I1027 08:24:25.709934 2760 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 27 08:24:25.710252 kubelet[2760]: I1027 08:24:25.710241 2760 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 27 08:24:25.710395 kubelet[2760]: I1027 08:24:25.710365 2760 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 27 08:24:25.710721 kubelet[2760]: I1027 08:24:25.710700 2760 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 27 08:24:25.714573 kubelet[2760]: E1027 08:24:25.714552 2760 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 27 08:24:25.740697 kubelet[2760]: I1027 08:24:25.739876 2760 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.743527 kubelet[2760]: I1027 08:24:25.741828 2760 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.743979 kubelet[2760]: I1027 08:24:25.743952 2760 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.755777 kubelet[2760]: W1027 08:24:25.755740 2760 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 27 08:24:25.757443 kubelet[2760]: W1027 08:24:25.757238 2760 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 27 08:24:25.757443 kubelet[2760]: E1027 08:24:25.757306 2760 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" already exists" pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.759768 kubelet[2760]: W1027 08:24:25.759746 2760 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 27 08:24:25.814707 kubelet[2760]: I1027 08:24:25.813991 2760 kubelet_node_status.go:75] "Attempting to register node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.823648 kubelet[2760]: I1027 08:24:25.823597 2760 kubelet_node_status.go:124] "Node was previously registered" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.823805 kubelet[2760]: I1027 08:24:25.823717 2760 kubelet_node_status.go:78] "Successfully registered node" node="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932277 kubelet[2760]: I1027 08:24:25.932220 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/570fe50a6da5294452f1f27d07f761b8-usr-share-ca-certificates\") pod \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"570fe50a6da5294452f1f27d07f761b8\") " pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932277 kubelet[2760]: I1027 08:24:25.932266 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-ca-certs\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932277 kubelet[2760]: I1027 08:24:25.932289 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-kubeconfig\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932515 kubelet[2760]: I1027 08:24:25.932306 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7f3336d3ffb5159bf5d091da7e24649b-kubeconfig\") pod \"kube-scheduler-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"7f3336d3ffb5159bf5d091da7e24649b\") " pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932515 kubelet[2760]: I1027 08:24:25.932324 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/570fe50a6da5294452f1f27d07f761b8-ca-certs\") pod \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"570fe50a6da5294452f1f27d07f761b8\") " pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932515 kubelet[2760]: I1027 08:24:25.932338 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/570fe50a6da5294452f1f27d07f761b8-k8s-certs\") pod \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"570fe50a6da5294452f1f27d07f761b8\") " pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932515 kubelet[2760]: I1027 08:24:25.932356 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-flexvolume-dir\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932515 kubelet[2760]: I1027 08:24:25.932379 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-k8s-certs\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:25.932754 kubelet[2760]: I1027 08:24:25.932410 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0f5c5cae72b141360a951471b3d00ecb-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1\" (UID: \"0f5c5cae72b141360a951471b3d00ecb\") " pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:26.058518 kubelet[2760]: E1027 08:24:26.058127 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:26.059842 kubelet[2760]: E1027 08:24:26.059431 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:26.062972 kubelet[2760]: E1027 08:24:26.062919 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:26.598940 kubelet[2760]: I1027 08:24:26.598882 2760 apiserver.go:52] "Watching apiserver" Oct 27 08:24:26.630007 kubelet[2760]: I1027 08:24:26.629944 2760 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 27 08:24:26.675955 kubelet[2760]: I1027 08:24:26.675850 2760 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:26.676430 kubelet[2760]: E1027 08:24:26.676390 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:26.676644 kubelet[2760]: I1027 08:24:26.676630 2760 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:26.683424 kubelet[2760]: W1027 08:24:26.683373 2760 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 27 08:24:26.683786 kubelet[2760]: E1027 08:24:26.683668 2760 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-9999.9.9-k-4f7b9e69b1\" already exists" pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:26.684002 kubelet[2760]: E1027 08:24:26.683951 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:26.686095 kubelet[2760]: W1027 08:24:26.686062 2760 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 27 08:24:26.686223 kubelet[2760]: E1027 08:24:26.686130 2760 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-9999.9.9-k-4f7b9e69b1\" already exists" pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:24:26.686358 kubelet[2760]: E1027 08:24:26.686342 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:26.722444 kubelet[2760]: I1027 08:24:26.721978 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-9999.9.9-k-4f7b9e69b1" podStartSLOduration=1.7219573320000001 podStartE2EDuration="1.721957332s" podCreationTimestamp="2025-10-27 08:24:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 08:24:26.708674987 +0000 UTC m=+1.212654481" watchObservedRunningTime="2025-10-27 08:24:26.721957332 +0000 UTC m=+1.225936822" Oct 27 08:24:26.736083 kubelet[2760]: I1027 08:24:26.736024 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-9999.9.9-k-4f7b9e69b1" podStartSLOduration=3.735891288 podStartE2EDuration="3.735891288s" podCreationTimestamp="2025-10-27 08:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 08:24:26.722300426 +0000 UTC m=+1.226279921" watchObservedRunningTime="2025-10-27 08:24:26.735891288 +0000 UTC m=+1.239870782" Oct 27 08:24:26.765606 kubelet[2760]: I1027 08:24:26.765503 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-9999.9.9-k-4f7b9e69b1" podStartSLOduration=1.765482264 podStartE2EDuration="1.765482264s" podCreationTimestamp="2025-10-27 08:24:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 08:24:26.737176934 +0000 UTC m=+1.241156431" watchObservedRunningTime="2025-10-27 08:24:26.765482264 +0000 UTC m=+1.269461758" Oct 27 08:24:27.678406 kubelet[2760]: E1027 08:24:27.678354 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:27.679175 kubelet[2760]: E1027 08:24:27.679097 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:28.680460 kubelet[2760]: E1027 08:24:28.680195 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:28.680460 kubelet[2760]: E1027 08:24:28.680366 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:29.132296 kubelet[2760]: I1027 08:24:29.132111 2760 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 27 08:24:29.133209 containerd[1602]: time="2025-10-27T08:24:29.133104987Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 27 08:24:29.134003 kubelet[2760]: I1027 08:24:29.133981 2760 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 27 08:24:29.990964 systemd[1]: Created slice kubepods-besteffort-podc1b20127_7706_4f49_a29d_ac7e43f562f7.slice - libcontainer container kubepods-besteffort-podc1b20127_7706_4f49_a29d_ac7e43f562f7.slice. Oct 27 08:24:30.076174 kubelet[2760]: I1027 08:24:30.076061 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g2j4\" (UniqueName: \"kubernetes.io/projected/c1b20127-7706-4f49-a29d-ac7e43f562f7-kube-api-access-7g2j4\") pod \"kube-proxy-cjmm5\" (UID: \"c1b20127-7706-4f49-a29d-ac7e43f562f7\") " pod="kube-system/kube-proxy-cjmm5" Oct 27 08:24:30.076174 kubelet[2760]: I1027 08:24:30.076129 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c1b20127-7706-4f49-a29d-ac7e43f562f7-xtables-lock\") pod \"kube-proxy-cjmm5\" (UID: \"c1b20127-7706-4f49-a29d-ac7e43f562f7\") " pod="kube-system/kube-proxy-cjmm5" Oct 27 08:24:30.076174 kubelet[2760]: I1027 08:24:30.076163 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/c1b20127-7706-4f49-a29d-ac7e43f562f7-kube-proxy\") pod \"kube-proxy-cjmm5\" (UID: \"c1b20127-7706-4f49-a29d-ac7e43f562f7\") " pod="kube-system/kube-proxy-cjmm5" Oct 27 08:24:30.076174 kubelet[2760]: I1027 08:24:30.076191 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c1b20127-7706-4f49-a29d-ac7e43f562f7-lib-modules\") pod \"kube-proxy-cjmm5\" (UID: \"c1b20127-7706-4f49-a29d-ac7e43f562f7\") " pod="kube-system/kube-proxy-cjmm5" Oct 27 08:24:30.260199 systemd[1]: Created slice kubepods-besteffort-pode2e3c485_779b_4b42_9409_07f36a194494.slice - libcontainer container kubepods-besteffort-pode2e3c485_779b_4b42_9409_07f36a194494.slice. Oct 27 08:24:30.277828 kubelet[2760]: I1027 08:24:30.277775 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e2e3c485-779b-4b42-9409-07f36a194494-var-lib-calico\") pod \"tigera-operator-7dcd859c48-ptvrb\" (UID: \"e2e3c485-779b-4b42-9409-07f36a194494\") " pod="tigera-operator/tigera-operator-7dcd859c48-ptvrb" Oct 27 08:24:30.277828 kubelet[2760]: I1027 08:24:30.277817 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2b2w\" (UniqueName: \"kubernetes.io/projected/e2e3c485-779b-4b42-9409-07f36a194494-kube-api-access-h2b2w\") pod \"tigera-operator-7dcd859c48-ptvrb\" (UID: \"e2e3c485-779b-4b42-9409-07f36a194494\") " pod="tigera-operator/tigera-operator-7dcd859c48-ptvrb" Oct 27 08:24:30.302116 kubelet[2760]: E1027 08:24:30.302075 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:30.303497 containerd[1602]: time="2025-10-27T08:24:30.303435158Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-cjmm5,Uid:c1b20127-7706-4f49-a29d-ac7e43f562f7,Namespace:kube-system,Attempt:0,}" Oct 27 08:24:30.326427 containerd[1602]: time="2025-10-27T08:24:30.326069854Z" level=info msg="connecting to shim 01a38df0d1f3de2ff7c7834db6d6b459924cd63f46a81ea92a5740c43d7b4be9" address="unix:///run/containerd/s/543b68dfec4f2e8a2187de40ed1160491e5c4d02b21c3229d338ae1f6cde20e2" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:30.357947 systemd[1]: Started cri-containerd-01a38df0d1f3de2ff7c7834db6d6b459924cd63f46a81ea92a5740c43d7b4be9.scope - libcontainer container 01a38df0d1f3de2ff7c7834db6d6b459924cd63f46a81ea92a5740c43d7b4be9. Oct 27 08:24:30.404124 containerd[1602]: time="2025-10-27T08:24:30.404062961Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-cjmm5,Uid:c1b20127-7706-4f49-a29d-ac7e43f562f7,Namespace:kube-system,Attempt:0,} returns sandbox id \"01a38df0d1f3de2ff7c7834db6d6b459924cd63f46a81ea92a5740c43d7b4be9\"" Oct 27 08:24:30.406909 kubelet[2760]: E1027 08:24:30.406853 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:30.415224 containerd[1602]: time="2025-10-27T08:24:30.415175436Z" level=info msg="CreateContainer within sandbox \"01a38df0d1f3de2ff7c7834db6d6b459924cd63f46a81ea92a5740c43d7b4be9\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 27 08:24:30.428085 containerd[1602]: time="2025-10-27T08:24:30.427446346Z" level=info msg="Container a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:30.439251 containerd[1602]: time="2025-10-27T08:24:30.439154297Z" level=info msg="CreateContainer within sandbox \"01a38df0d1f3de2ff7c7834db6d6b459924cd63f46a81ea92a5740c43d7b4be9\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0\"" Oct 27 08:24:30.442371 containerd[1602]: time="2025-10-27T08:24:30.440840329Z" level=info msg="StartContainer for \"a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0\"" Oct 27 08:24:30.444040 containerd[1602]: time="2025-10-27T08:24:30.444000604Z" level=info msg="connecting to shim a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0" address="unix:///run/containerd/s/543b68dfec4f2e8a2187de40ed1160491e5c4d02b21c3229d338ae1f6cde20e2" protocol=ttrpc version=3 Oct 27 08:24:30.465914 systemd[1]: Started cri-containerd-a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0.scope - libcontainer container a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0. Oct 27 08:24:30.488196 kubelet[2760]: E1027 08:24:30.488122 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:30.534182 containerd[1602]: time="2025-10-27T08:24:30.533560510Z" level=info msg="StartContainer for \"a94d312aa1d81fbd536560481d728a75a4e8d5287f648484caa421116a7cabf0\" returns successfully" Oct 27 08:24:30.567907 containerd[1602]: time="2025-10-27T08:24:30.567852422Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-ptvrb,Uid:e2e3c485-779b-4b42-9409-07f36a194494,Namespace:tigera-operator,Attempt:0,}" Oct 27 08:24:30.584914 containerd[1602]: time="2025-10-27T08:24:30.584815053Z" level=info msg="connecting to shim 4f01870b2573226a1de39a6a0545f4a7854dcd5c78af0c13b3c3283fa35aff27" address="unix:///run/containerd/s/c57ff6d6116fb3b855e4479ee7438bc67c7767ff6859347a1579d498572acc74" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:30.621849 systemd[1]: Started cri-containerd-4f01870b2573226a1de39a6a0545f4a7854dcd5c78af0c13b3c3283fa35aff27.scope - libcontainer container 4f01870b2573226a1de39a6a0545f4a7854dcd5c78af0c13b3c3283fa35aff27. Oct 27 08:24:30.697361 kubelet[2760]: E1027 08:24:30.695979 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:30.697361 kubelet[2760]: E1027 08:24:30.697159 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:30.703162 containerd[1602]: time="2025-10-27T08:24:30.702793067Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-ptvrb,Uid:e2e3c485-779b-4b42-9409-07f36a194494,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"4f01870b2573226a1de39a6a0545f4a7854dcd5c78af0c13b3c3283fa35aff27\"" Oct 27 08:24:30.708475 containerd[1602]: time="2025-10-27T08:24:30.708441086Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Oct 27 08:24:30.711299 systemd-resolved[1286]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Oct 27 08:24:30.739863 kubelet[2760]: I1027 08:24:30.738881 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-cjmm5" podStartSLOduration=1.738763027 podStartE2EDuration="1.738763027s" podCreationTimestamp="2025-10-27 08:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 08:24:30.718347923 +0000 UTC m=+5.222327414" watchObservedRunningTime="2025-10-27 08:24:30.738763027 +0000 UTC m=+5.242742519" Oct 27 08:24:31.194633 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1455333625.mount: Deactivated successfully. Oct 27 08:24:31.700293 kubelet[2760]: E1027 08:24:31.700260 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:32.126897 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2575606009.mount: Deactivated successfully. Oct 27 08:24:32.718476 systemd-resolved[1286]: Clock change detected. Flushing caches. Oct 27 08:24:32.719041 systemd-timesyncd[1463]: Contacted time server 72.87.88.202:123 (2.flatcar.pool.ntp.org). Oct 27 08:24:32.719122 systemd-timesyncd[1463]: Initial clock synchronization to Mon 2025-10-27 08:24:32.718242 UTC. Oct 27 08:24:33.781472 containerd[1602]: time="2025-10-27T08:24:33.780842723Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:33.782128 containerd[1602]: time="2025-10-27T08:24:33.782094211Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=25061691" Oct 27 08:24:33.782734 containerd[1602]: time="2025-10-27T08:24:33.782685757Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:33.784439 containerd[1602]: time="2025-10-27T08:24:33.784390058Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:33.785138 containerd[1602]: time="2025-10-27T08:24:33.784998510Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 2.523422594s" Oct 27 08:24:33.785138 containerd[1602]: time="2025-10-27T08:24:33.785031075Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Oct 27 08:24:33.789380 containerd[1602]: time="2025-10-27T08:24:33.789266307Z" level=info msg="CreateContainer within sandbox \"4f01870b2573226a1de39a6a0545f4a7854dcd5c78af0c13b3c3283fa35aff27\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Oct 27 08:24:33.797212 containerd[1602]: time="2025-10-27T08:24:33.797137752Z" level=info msg="Container e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:33.801355 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3845213890.mount: Deactivated successfully. Oct 27 08:24:33.807422 containerd[1602]: time="2025-10-27T08:24:33.807286682Z" level=info msg="CreateContainer within sandbox \"4f01870b2573226a1de39a6a0545f4a7854dcd5c78af0c13b3c3283fa35aff27\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7\"" Oct 27 08:24:33.808891 containerd[1602]: time="2025-10-27T08:24:33.808858786Z" level=info msg="StartContainer for \"e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7\"" Oct 27 08:24:33.810606 containerd[1602]: time="2025-10-27T08:24:33.810525146Z" level=info msg="connecting to shim e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7" address="unix:///run/containerd/s/c57ff6d6116fb3b855e4479ee7438bc67c7767ff6859347a1579d498572acc74" protocol=ttrpc version=3 Oct 27 08:24:33.845812 systemd[1]: Started cri-containerd-e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7.scope - libcontainer container e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7. Oct 27 08:24:33.882445 containerd[1602]: time="2025-10-27T08:24:33.882399673Z" level=info msg="StartContainer for \"e5e4f8926abce01eaff997e8f7b31aa4e9bffdabc2ca78059a6d18619c6962d7\" returns successfully" Oct 27 08:24:35.570588 kubelet[2760]: E1027 08:24:35.570209 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:35.607126 kubelet[2760]: I1027 08:24:35.606892 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-ptvrb" podStartSLOduration=3.080437957 podStartE2EDuration="5.606870298s" podCreationTimestamp="2025-10-27 08:24:30 +0000 UTC" firstStartedPulling="2025-10-27 08:24:30.707020111 +0000 UTC m=+5.210999586" lastFinishedPulling="2025-10-27 08:24:33.786382329 +0000 UTC m=+7.737431927" observedRunningTime="2025-10-27 08:24:34.272848559 +0000 UTC m=+8.223898154" watchObservedRunningTime="2025-10-27 08:24:35.606870298 +0000 UTC m=+9.557919916" Oct 27 08:24:36.264202 kubelet[2760]: E1027 08:24:36.264157 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:39.112130 kubelet[2760]: E1027 08:24:39.111998 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:39.272037 kubelet[2760]: E1027 08:24:39.271995 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:40.802112 update_engine[1572]: I20251027 08:24:40.800747 1572 update_attempter.cc:509] Updating boot flags... Oct 27 08:24:40.953707 sudo[1833]: pam_unix(sudo:session): session closed for user root Oct 27 08:24:40.964049 sshd[1832]: Connection closed by 139.178.89.65 port 34868 Oct 27 08:24:40.962961 sshd-session[1829]: pam_unix(sshd:session): session closed for user core Oct 27 08:24:40.983929 systemd[1]: sshd@6-64.23.205.201:22-139.178.89.65:34868.service: Deactivated successfully. Oct 27 08:24:40.988991 systemd[1]: session-7.scope: Deactivated successfully. Oct 27 08:24:40.990650 systemd[1]: session-7.scope: Consumed 6.301s CPU time, 159.1M memory peak. Oct 27 08:24:40.993023 systemd-logind[1571]: Session 7 logged out. Waiting for processes to exit. Oct 27 08:24:41.012381 systemd-logind[1571]: Removed session 7. Oct 27 08:24:45.868664 systemd[1]: Created slice kubepods-besteffort-pod41443dba_bfd6_4ac7_bf7c_fde0cd8d11cf.slice - libcontainer container kubepods-besteffort-pod41443dba_bfd6_4ac7_bf7c_fde0cd8d11cf.slice. Oct 27 08:24:45.921432 kubelet[2760]: I1027 08:24:45.921388 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf-tigera-ca-bundle\") pod \"calico-typha-685cb5f8dc-lbfhv\" (UID: \"41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf\") " pod="calico-system/calico-typha-685cb5f8dc-lbfhv" Oct 27 08:24:45.921432 kubelet[2760]: I1027 08:24:45.921430 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf-typha-certs\") pod \"calico-typha-685cb5f8dc-lbfhv\" (UID: \"41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf\") " pod="calico-system/calico-typha-685cb5f8dc-lbfhv" Oct 27 08:24:45.921432 kubelet[2760]: I1027 08:24:45.921449 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh84k\" (UniqueName: \"kubernetes.io/projected/41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf-kube-api-access-vh84k\") pod \"calico-typha-685cb5f8dc-lbfhv\" (UID: \"41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf\") " pod="calico-system/calico-typha-685cb5f8dc-lbfhv" Oct 27 08:24:46.076617 systemd[1]: Created slice kubepods-besteffort-pod19d70333_4eb9_4efd_b5e1_4400dba4ecf6.slice - libcontainer container kubepods-besteffort-pod19d70333_4eb9_4efd_b5e1_4400dba4ecf6.slice. Oct 27 08:24:46.122877 kubelet[2760]: I1027 08:24:46.122382 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-cni-net-dir\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.122877 kubelet[2760]: I1027 08:24:46.122447 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-node-certs\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.122877 kubelet[2760]: I1027 08:24:46.122474 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-xtables-lock\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.122877 kubelet[2760]: I1027 08:24:46.122501 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-cni-bin-dir\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.122877 kubelet[2760]: I1027 08:24:46.122545 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-lib-modules\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123111 kubelet[2760]: I1027 08:24:46.122566 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-policysync\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123111 kubelet[2760]: I1027 08:24:46.122587 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-var-lib-calico\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123111 kubelet[2760]: I1027 08:24:46.122608 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-var-run-calico\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123111 kubelet[2760]: I1027 08:24:46.122638 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-flexvol-driver-host\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123111 kubelet[2760]: I1027 08:24:46.122662 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-tigera-ca-bundle\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123237 kubelet[2760]: I1027 08:24:46.122685 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m75f7\" (UniqueName: \"kubernetes.io/projected/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-kube-api-access-m75f7\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.123237 kubelet[2760]: I1027 08:24:46.122707 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/19d70333-4eb9-4efd-b5e1-4400dba4ecf6-cni-log-dir\") pod \"calico-node-txx29\" (UID: \"19d70333-4eb9-4efd-b5e1-4400dba4ecf6\") " pod="calico-system/calico-node-txx29" Oct 27 08:24:46.173670 kubelet[2760]: E1027 08:24:46.173628 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:46.174943 containerd[1602]: time="2025-10-27T08:24:46.174868106Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-685cb5f8dc-lbfhv,Uid:41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf,Namespace:calico-system,Attempt:0,}" Oct 27 08:24:46.232837 kubelet[2760]: E1027 08:24:46.232650 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.232837 kubelet[2760]: W1027 08:24:46.232688 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.237152 kubelet[2760]: E1027 08:24:46.233946 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.240476 kubelet[2760]: E1027 08:24:46.240045 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.240476 kubelet[2760]: W1027 08:24:46.240135 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.241944 kubelet[2760]: E1027 08:24:46.241306 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.241944 kubelet[2760]: W1027 08:24:46.241326 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.241944 kubelet[2760]: E1027 08:24:46.241349 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.243153 kubelet[2760]: E1027 08:24:46.243087 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.243153 kubelet[2760]: E1027 08:24:46.243100 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.243269 kubelet[2760]: W1027 08:24:46.243159 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.243917 kubelet[2760]: E1027 08:24:46.243359 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.243917 kubelet[2760]: E1027 08:24:46.243395 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.243917 kubelet[2760]: W1027 08:24:46.243405 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.243917 kubelet[2760]: E1027 08:24:46.243896 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.244443 kubelet[2760]: E1027 08:24:46.243983 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.244443 kubelet[2760]: W1027 08:24:46.243996 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.244443 kubelet[2760]: E1027 08:24:46.244238 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.244443 kubelet[2760]: E1027 08:24:46.244260 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.244443 kubelet[2760]: W1027 08:24:46.244270 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.244443 kubelet[2760]: E1027 08:24:46.244289 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.244624 kubelet[2760]: E1027 08:24:46.244507 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.244624 kubelet[2760]: W1027 08:24:46.244541 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.244624 kubelet[2760]: E1027 08:24:46.244553 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.245541 kubelet[2760]: E1027 08:24:46.244719 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.245541 kubelet[2760]: W1027 08:24:46.244730 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.245541 kubelet[2760]: E1027 08:24:46.244741 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.245541 kubelet[2760]: E1027 08:24:46.244916 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.245541 kubelet[2760]: W1027 08:24:46.244924 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.245541 kubelet[2760]: E1027 08:24:46.244932 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.246760 kubelet[2760]: E1027 08:24:46.246002 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.246760 kubelet[2760]: W1027 08:24:46.246015 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.246760 kubelet[2760]: E1027 08:24:46.246028 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.247624 kubelet[2760]: E1027 08:24:46.247374 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.247800 kubelet[2760]: W1027 08:24:46.247783 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.247949 kubelet[2760]: E1027 08:24:46.247935 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.249116 kubelet[2760]: E1027 08:24:46.248889 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.249116 kubelet[2760]: W1027 08:24:46.248904 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.249116 kubelet[2760]: E1027 08:24:46.248917 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.251753 containerd[1602]: time="2025-10-27T08:24:46.251647969Z" level=info msg="connecting to shim b146cbbb44210893beeb3730dfc1164b79770e82bfcb4bcf8f10ed225f0cfece" address="unix:///run/containerd/s/76d727b2cac8a9ff8e3ecca19be83bd4b779b228f5bbfbe591ceee46e3c95ed5" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:46.312351 kubelet[2760]: E1027 08:24:46.312317 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.313708 kubelet[2760]: W1027 08:24:46.313468 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.313708 kubelet[2760]: E1027 08:24:46.313649 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.319788 systemd[1]: Started cri-containerd-b146cbbb44210893beeb3730dfc1164b79770e82bfcb4bcf8f10ed225f0cfece.scope - libcontainer container b146cbbb44210893beeb3730dfc1164b79770e82bfcb4bcf8f10ed225f0cfece. Oct 27 08:24:46.335562 kubelet[2760]: E1027 08:24:46.335495 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:24:46.383622 kubelet[2760]: E1027 08:24:46.381336 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:46.384551 containerd[1602]: time="2025-10-27T08:24:46.384325066Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-txx29,Uid:19d70333-4eb9-4efd-b5e1-4400dba4ecf6,Namespace:calico-system,Attempt:0,}" Oct 27 08:24:46.406574 kubelet[2760]: E1027 08:24:46.406508 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.406574 kubelet[2760]: W1027 08:24:46.406565 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.407890 kubelet[2760]: E1027 08:24:46.406593 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.407890 kubelet[2760]: E1027 08:24:46.406971 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.407890 kubelet[2760]: W1027 08:24:46.406982 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.407890 kubelet[2760]: E1027 08:24:46.406994 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.408472 kubelet[2760]: E1027 08:24:46.408444 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.408472 kubelet[2760]: W1027 08:24:46.408461 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.408472 kubelet[2760]: E1027 08:24:46.408474 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.408919 kubelet[2760]: E1027 08:24:46.408904 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.408919 kubelet[2760]: W1027 08:24:46.408917 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.408986 kubelet[2760]: E1027 08:24:46.408929 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.409563 kubelet[2760]: E1027 08:24:46.409289 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.409563 kubelet[2760]: W1027 08:24:46.409302 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.409563 kubelet[2760]: E1027 08:24:46.409313 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.411626 kubelet[2760]: E1027 08:24:46.409567 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.411626 kubelet[2760]: W1027 08:24:46.409576 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.411626 kubelet[2760]: E1027 08:24:46.409586 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.411626 kubelet[2760]: E1027 08:24:46.409863 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.411626 kubelet[2760]: W1027 08:24:46.409876 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.411626 kubelet[2760]: E1027 08:24:46.409890 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.411954 kubelet[2760]: E1027 08:24:46.411938 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.411954 kubelet[2760]: W1027 08:24:46.411953 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.412035 kubelet[2760]: E1027 08:24:46.411966 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.412995 kubelet[2760]: E1027 08:24:46.412393 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.412995 kubelet[2760]: W1027 08:24:46.412406 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.412995 kubelet[2760]: E1027 08:24:46.412423 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.412995 kubelet[2760]: E1027 08:24:46.412729 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.412995 kubelet[2760]: W1027 08:24:46.412739 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.412995 kubelet[2760]: E1027 08:24:46.412749 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.413191 kubelet[2760]: E1027 08:24:46.413109 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.413191 kubelet[2760]: W1027 08:24:46.413120 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.413191 kubelet[2760]: E1027 08:24:46.413131 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.413410 kubelet[2760]: E1027 08:24:46.413321 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.413410 kubelet[2760]: W1027 08:24:46.413333 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.413410 kubelet[2760]: E1027 08:24:46.413343 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.413556 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.415649 kubelet[2760]: W1027 08:24:46.413566 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.413576 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.413766 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.415649 kubelet[2760]: W1027 08:24:46.413774 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.413783 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.414027 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.415649 kubelet[2760]: W1027 08:24:46.414040 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.414050 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.415649 kubelet[2760]: E1027 08:24:46.414410 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.416010 kubelet[2760]: W1027 08:24:46.414420 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.416010 kubelet[2760]: E1027 08:24:46.414430 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.416010 kubelet[2760]: E1027 08:24:46.414736 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.416010 kubelet[2760]: W1027 08:24:46.414746 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.416010 kubelet[2760]: E1027 08:24:46.414756 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.416175 kubelet[2760]: E1027 08:24:46.416158 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.416213 kubelet[2760]: W1027 08:24:46.416174 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.416213 kubelet[2760]: E1027 08:24:46.416186 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.416572 kubelet[2760]: E1027 08:24:46.416492 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.416572 kubelet[2760]: W1027 08:24:46.416503 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.416572 kubelet[2760]: E1027 08:24:46.416541 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.417272 kubelet[2760]: E1027 08:24:46.417177 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.417272 kubelet[2760]: W1027 08:24:46.417188 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.417272 kubelet[2760]: E1027 08:24:46.417199 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.426273 kubelet[2760]: E1027 08:24:46.426230 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.428318 kubelet[2760]: W1027 08:24:46.427377 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.428318 kubelet[2760]: E1027 08:24:46.427414 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.428318 kubelet[2760]: I1027 08:24:46.427452 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw4h6\" (UniqueName: \"kubernetes.io/projected/b9e97248-79da-4981-820e-d3ae986590e1-kube-api-access-sw4h6\") pod \"csi-node-driver-fvjzl\" (UID: \"b9e97248-79da-4981-820e-d3ae986590e1\") " pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:46.429541 kubelet[2760]: E1027 08:24:46.428606 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.429541 kubelet[2760]: W1027 08:24:46.428625 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.429541 kubelet[2760]: E1027 08:24:46.428648 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.429541 kubelet[2760]: I1027 08:24:46.428675 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b9e97248-79da-4981-820e-d3ae986590e1-socket-dir\") pod \"csi-node-driver-fvjzl\" (UID: \"b9e97248-79da-4981-820e-d3ae986590e1\") " pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:46.429946 kubelet[2760]: E1027 08:24:46.429926 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.430055 kubelet[2760]: W1027 08:24:46.430040 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.430537 kubelet[2760]: E1027 08:24:46.430275 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.430537 kubelet[2760]: I1027 08:24:46.430331 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b9e97248-79da-4981-820e-d3ae986590e1-registration-dir\") pod \"csi-node-driver-fvjzl\" (UID: \"b9e97248-79da-4981-820e-d3ae986590e1\") " pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:46.432648 kubelet[2760]: E1027 08:24:46.432586 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.432844 kubelet[2760]: W1027 08:24:46.432731 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.432844 kubelet[2760]: E1027 08:24:46.432777 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.433371 kubelet[2760]: E1027 08:24:46.433273 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.433544 kubelet[2760]: W1027 08:24:46.433428 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.433544 kubelet[2760]: E1027 08:24:46.433466 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.433735 containerd[1602]: time="2025-10-27T08:24:46.433359203Z" level=info msg="connecting to shim 46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b" address="unix:///run/containerd/s/2bc0edaee80744c79d9e57f4cec1b254f62ce3fc76401b8d89d5f7912e030b89" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:24:46.434092 kubelet[2760]: E1027 08:24:46.433899 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.434092 kubelet[2760]: W1027 08:24:46.433911 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.434092 kubelet[2760]: E1027 08:24:46.433943 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.434092 kubelet[2760]: I1027 08:24:46.433996 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/b9e97248-79da-4981-820e-d3ae986590e1-varrun\") pod \"csi-node-driver-fvjzl\" (UID: \"b9e97248-79da-4981-820e-d3ae986590e1\") " pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:46.434623 kubelet[2760]: E1027 08:24:46.434591 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.435724 kubelet[2760]: W1027 08:24:46.435565 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.435724 kubelet[2760]: E1027 08:24:46.435610 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.436035 kubelet[2760]: E1027 08:24:46.435956 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.436035 kubelet[2760]: W1027 08:24:46.435968 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.436035 kubelet[2760]: E1027 08:24:46.435982 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.436288 kubelet[2760]: E1027 08:24:46.436271 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.436288 kubelet[2760]: W1027 08:24:46.436287 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.436761 kubelet[2760]: E1027 08:24:46.436306 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.436761 kubelet[2760]: E1027 08:24:46.436641 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.436761 kubelet[2760]: W1027 08:24:46.436652 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.436761 kubelet[2760]: E1027 08:24:46.436681 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.436890 kubelet[2760]: E1027 08:24:46.436855 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.436890 kubelet[2760]: W1027 08:24:46.436862 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.436890 kubelet[2760]: E1027 08:24:46.436870 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.436961 kubelet[2760]: I1027 08:24:46.436905 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9e97248-79da-4981-820e-d3ae986590e1-kubelet-dir\") pod \"csi-node-driver-fvjzl\" (UID: \"b9e97248-79da-4981-820e-d3ae986590e1\") " pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:46.437696 kubelet[2760]: E1027 08:24:46.437668 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.437696 kubelet[2760]: W1027 08:24:46.437680 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.437696 kubelet[2760]: E1027 08:24:46.437691 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.438194 kubelet[2760]: E1027 08:24:46.437921 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.438194 kubelet[2760]: W1027 08:24:46.437936 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.438194 kubelet[2760]: E1027 08:24:46.437950 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.438194 kubelet[2760]: E1027 08:24:46.438198 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.438603 kubelet[2760]: W1027 08:24:46.438206 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.438603 kubelet[2760]: E1027 08:24:46.438216 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.438603 kubelet[2760]: E1027 08:24:46.438376 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.438603 kubelet[2760]: W1027 08:24:46.438384 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.438603 kubelet[2760]: E1027 08:24:46.438393 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.469229 systemd[1]: Started cri-containerd-46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b.scope - libcontainer container 46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b. Oct 27 08:24:46.495889 containerd[1602]: time="2025-10-27T08:24:46.495826056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-685cb5f8dc-lbfhv,Uid:41443dba-bfd6-4ac7-bf7c-fde0cd8d11cf,Namespace:calico-system,Attempt:0,} returns sandbox id \"b146cbbb44210893beeb3730dfc1164b79770e82bfcb4bcf8f10ed225f0cfece\"" Oct 27 08:24:46.497290 kubelet[2760]: E1027 08:24:46.497263 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:46.500040 containerd[1602]: time="2025-10-27T08:24:46.499801115Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Oct 27 08:24:46.532735 containerd[1602]: time="2025-10-27T08:24:46.532695070Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-txx29,Uid:19d70333-4eb9-4efd-b5e1-4400dba4ecf6,Namespace:calico-system,Attempt:0,} returns sandbox id \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\"" Oct 27 08:24:46.534979 kubelet[2760]: E1027 08:24:46.534949 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:46.540179 kubelet[2760]: E1027 08:24:46.540118 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.540508 kubelet[2760]: W1027 08:24:46.540365 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.540508 kubelet[2760]: E1027 08:24:46.540413 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.541465 kubelet[2760]: E1027 08:24:46.541408 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.542171 kubelet[2760]: W1027 08:24:46.541551 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.542171 kubelet[2760]: E1027 08:24:46.541589 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.542171 kubelet[2760]: E1027 08:24:46.542141 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.542171 kubelet[2760]: W1027 08:24:46.542158 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.542171 kubelet[2760]: E1027 08:24:46.542173 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.542735 kubelet[2760]: E1027 08:24:46.542689 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.542735 kubelet[2760]: W1027 08:24:46.542704 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.542735 kubelet[2760]: E1027 08:24:46.542716 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.543740 kubelet[2760]: E1027 08:24:46.543724 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.543887 kubelet[2760]: W1027 08:24:46.543820 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.543887 kubelet[2760]: E1027 08:24:46.543841 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.544731 kubelet[2760]: E1027 08:24:46.544711 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.544731 kubelet[2760]: W1027 08:24:46.544729 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.544940 kubelet[2760]: E1027 08:24:46.544746 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.545027 kubelet[2760]: E1027 08:24:46.545016 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.545556 kubelet[2760]: W1027 08:24:46.545026 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.545556 kubelet[2760]: E1027 08:24:46.545092 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.545556 kubelet[2760]: E1027 08:24:46.545274 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.545556 kubelet[2760]: W1027 08:24:46.545282 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.545556 kubelet[2760]: E1027 08:24:46.545336 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.545556 kubelet[2760]: E1027 08:24:46.545487 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.545556 kubelet[2760]: W1027 08:24:46.545499 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.545810 kubelet[2760]: E1027 08:24:46.545634 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.545810 kubelet[2760]: E1027 08:24:46.545805 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.545983 kubelet[2760]: W1027 08:24:46.545818 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.545983 kubelet[2760]: E1027 08:24:46.545871 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.545983 kubelet[2760]: E1027 08:24:46.546053 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.545983 kubelet[2760]: W1027 08:24:46.546075 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.545983 kubelet[2760]: E1027 08:24:46.546108 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.546372 kubelet[2760]: E1027 08:24:46.546299 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.546372 kubelet[2760]: W1027 08:24:46.546307 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.546372 kubelet[2760]: E1027 08:24:46.546333 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.546791 kubelet[2760]: E1027 08:24:46.546776 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.546848 kubelet[2760]: W1027 08:24:46.546839 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.546903 kubelet[2760]: E1027 08:24:46.546894 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.548301 kubelet[2760]: E1027 08:24:46.548271 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.548301 kubelet[2760]: W1027 08:24:46.548298 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.548424 kubelet[2760]: E1027 08:24:46.548317 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.549041 kubelet[2760]: E1027 08:24:46.549014 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.549041 kubelet[2760]: W1027 08:24:46.549030 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.549174 kubelet[2760]: E1027 08:24:46.549136 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.549436 kubelet[2760]: E1027 08:24:46.549421 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.549436 kubelet[2760]: W1027 08:24:46.549434 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.549604 kubelet[2760]: E1027 08:24:46.549585 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.553671 kubelet[2760]: E1027 08:24:46.553633 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.554044 kubelet[2760]: W1027 08:24:46.553861 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.554365 kubelet[2760]: E1027 08:24:46.554288 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.554698 kubelet[2760]: W1027 08:24:46.554586 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.555749 kubelet[2760]: E1027 08:24:46.555701 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.555749 kubelet[2760]: W1027 08:24:46.555716 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.556285 kubelet[2760]: E1027 08:24:46.556269 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.556779 kubelet[2760]: E1027 08:24:46.556379 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.556779 kubelet[2760]: E1027 08:24:46.556388 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.556779 kubelet[2760]: E1027 08:24:46.556395 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.557101 kubelet[2760]: W1027 08:24:46.556944 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.557101 kubelet[2760]: E1027 08:24:46.557051 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.557922 kubelet[2760]: E1027 08:24:46.557861 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.557922 kubelet[2760]: W1027 08:24:46.557904 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.558228 kubelet[2760]: E1027 08:24:46.557985 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.558485 kubelet[2760]: E1027 08:24:46.558458 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.558485 kubelet[2760]: W1027 08:24:46.558471 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.558790 kubelet[2760]: E1027 08:24:46.558758 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.559071 kubelet[2760]: E1027 08:24:46.559060 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.559143 kubelet[2760]: W1027 08:24:46.559133 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.559258 kubelet[2760]: E1027 08:24:46.559206 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.559737 kubelet[2760]: E1027 08:24:46.559567 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.559737 kubelet[2760]: W1027 08:24:46.559583 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.559737 kubelet[2760]: E1027 08:24:46.559599 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.560287 kubelet[2760]: E1027 08:24:46.560065 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.560287 kubelet[2760]: W1027 08:24:46.560082 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.560287 kubelet[2760]: E1027 08:24:46.560106 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:46.575441 kubelet[2760]: E1027 08:24:46.575396 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:46.575441 kubelet[2760]: W1027 08:24:46.575429 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:46.575441 kubelet[2760]: E1027 08:24:46.575456 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:47.864149 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3703097981.mount: Deactivated successfully. Oct 27 08:24:48.205173 kubelet[2760]: E1027 08:24:48.205065 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:24:48.968475 containerd[1602]: time="2025-10-27T08:24:48.968418024Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:48.969194 containerd[1602]: time="2025-10-27T08:24:48.969157878Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=35234628" Oct 27 08:24:48.970134 containerd[1602]: time="2025-10-27T08:24:48.969760491Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:48.971843 containerd[1602]: time="2025-10-27T08:24:48.971811427Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:48.972471 containerd[1602]: time="2025-10-27T08:24:48.972443281Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 2.47257942s" Oct 27 08:24:48.972583 containerd[1602]: time="2025-10-27T08:24:48.972568556Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Oct 27 08:24:48.974312 containerd[1602]: time="2025-10-27T08:24:48.974282602Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Oct 27 08:24:48.992150 containerd[1602]: time="2025-10-27T08:24:48.992087450Z" level=info msg="CreateContainer within sandbox \"b146cbbb44210893beeb3730dfc1164b79770e82bfcb4bcf8f10ed225f0cfece\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Oct 27 08:24:48.999798 containerd[1602]: time="2025-10-27T08:24:48.999698372Z" level=info msg="Container 0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:49.020542 containerd[1602]: time="2025-10-27T08:24:49.019647320Z" level=info msg="CreateContainer within sandbox \"b146cbbb44210893beeb3730dfc1164b79770e82bfcb4bcf8f10ed225f0cfece\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda\"" Oct 27 08:24:49.030656 containerd[1602]: time="2025-10-27T08:24:49.030574998Z" level=info msg="StartContainer for \"0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda\"" Oct 27 08:24:49.032095 containerd[1602]: time="2025-10-27T08:24:49.032061189Z" level=info msg="connecting to shim 0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda" address="unix:///run/containerd/s/76d727b2cac8a9ff8e3ecca19be83bd4b779b228f5bbfbe591ceee46e3c95ed5" protocol=ttrpc version=3 Oct 27 08:24:49.059956 systemd[1]: Started cri-containerd-0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda.scope - libcontainer container 0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda. Oct 27 08:24:49.123825 containerd[1602]: time="2025-10-27T08:24:49.123579374Z" level=info msg="StartContainer for \"0476e07648e93efca731edbc934efd99cde3174e78945434409fdf5bab4bbbda\" returns successfully" Oct 27 08:24:49.314668 kubelet[2760]: E1027 08:24:49.313744 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:49.339511 kubelet[2760]: E1027 08:24:49.339453 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.339511 kubelet[2760]: W1027 08:24:49.339547 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.339511 kubelet[2760]: E1027 08:24:49.339582 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.340471 kubelet[2760]: E1027 08:24:49.340368 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.340841 kubelet[2760]: W1027 08:24:49.340611 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.340841 kubelet[2760]: E1027 08:24:49.340639 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.342754 kubelet[2760]: E1027 08:24:49.342730 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.343000 kubelet[2760]: W1027 08:24:49.342937 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.343000 kubelet[2760]: E1027 08:24:49.342969 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.345556 kubelet[2760]: E1027 08:24:49.343441 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.345892 kubelet[2760]: W1027 08:24:49.345686 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.345892 kubelet[2760]: E1027 08:24:49.345720 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.346074 kubelet[2760]: E1027 08:24:49.346056 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.346132 kubelet[2760]: W1027 08:24:49.346122 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.346321 kubelet[2760]: E1027 08:24:49.346306 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.348169 kubelet[2760]: E1027 08:24:49.348097 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.348169 kubelet[2760]: W1027 08:24:49.348118 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.348169 kubelet[2760]: E1027 08:24:49.348137 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.348737 kubelet[2760]: E1027 08:24:49.348648 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.348737 kubelet[2760]: W1027 08:24:49.348672 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.348737 kubelet[2760]: E1027 08:24:49.348684 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.351026 kubelet[2760]: E1027 08:24:49.349636 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.351026 kubelet[2760]: W1027 08:24:49.349650 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.351026 kubelet[2760]: E1027 08:24:49.349663 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.351603 kubelet[2760]: E1027 08:24:49.351537 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.351603 kubelet[2760]: W1027 08:24:49.351550 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.351603 kubelet[2760]: E1027 08:24:49.351564 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.351984 kubelet[2760]: E1027 08:24:49.351952 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.352360 kubelet[2760]: W1027 08:24:49.351966 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.352360 kubelet[2760]: E1027 08:24:49.352233 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.353726 kubelet[2760]: E1027 08:24:49.353644 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.354157 kubelet[2760]: W1027 08:24:49.353965 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.354157 kubelet[2760]: E1027 08:24:49.353984 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.355085 kubelet[2760]: E1027 08:24:49.354964 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.356646 kubelet[2760]: W1027 08:24:49.356547 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.356646 kubelet[2760]: E1027 08:24:49.356576 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.356971 kubelet[2760]: E1027 08:24:49.356902 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.356971 kubelet[2760]: W1027 08:24:49.356913 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.356971 kubelet[2760]: E1027 08:24:49.356926 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.357272 kubelet[2760]: E1027 08:24:49.357208 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.357272 kubelet[2760]: W1027 08:24:49.357218 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.357272 kubelet[2760]: E1027 08:24:49.357237 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.357757 kubelet[2760]: E1027 08:24:49.357659 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.357757 kubelet[2760]: W1027 08:24:49.357671 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.357757 kubelet[2760]: E1027 08:24:49.357682 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.371849 kubelet[2760]: E1027 08:24:49.371710 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.371849 kubelet[2760]: W1027 08:24:49.371736 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.371849 kubelet[2760]: E1027 08:24:49.371760 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.372300 kubelet[2760]: E1027 08:24:49.372213 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.372300 kubelet[2760]: W1027 08:24:49.372225 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.372300 kubelet[2760]: E1027 08:24:49.372246 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.372702 kubelet[2760]: E1027 08:24:49.372565 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.372702 kubelet[2760]: W1027 08:24:49.372585 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.372702 kubelet[2760]: E1027 08:24:49.372603 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.372957 kubelet[2760]: E1027 08:24:49.372945 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.373088 kubelet[2760]: W1027 08:24:49.373006 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.373088 kubelet[2760]: E1027 08:24:49.373028 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.373334 kubelet[2760]: E1027 08:24:49.373283 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.373334 kubelet[2760]: W1027 08:24:49.373292 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.373404 kubelet[2760]: E1027 08:24:49.373323 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.373646 kubelet[2760]: E1027 08:24:49.373597 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.373646 kubelet[2760]: W1027 08:24:49.373607 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.373721 kubelet[2760]: E1027 08:24:49.373654 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.373966 kubelet[2760]: E1027 08:24:49.373907 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.373966 kubelet[2760]: W1027 08:24:49.373925 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.373966 kubelet[2760]: E1027 08:24:49.373954 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.374283 kubelet[2760]: E1027 08:24:49.374182 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.374283 kubelet[2760]: W1027 08:24:49.374191 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.374283 kubelet[2760]: E1027 08:24:49.374211 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.374691 kubelet[2760]: E1027 08:24:49.374533 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.374691 kubelet[2760]: W1027 08:24:49.374544 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.374691 kubelet[2760]: E1027 08:24:49.374561 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.375717 kubelet[2760]: E1027 08:24:49.375691 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.375717 kubelet[2760]: W1027 08:24:49.375713 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.375940 kubelet[2760]: E1027 08:24:49.375740 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.375982 kubelet[2760]: E1027 08:24:49.375939 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.375982 kubelet[2760]: W1027 08:24:49.375952 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.376094 kubelet[2760]: E1027 08:24:49.376058 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.377628 kubelet[2760]: E1027 08:24:49.377607 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.377628 kubelet[2760]: W1027 08:24:49.377624 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.377839 kubelet[2760]: E1027 08:24:49.377693 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.378044 kubelet[2760]: E1027 08:24:49.377839 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.378044 kubelet[2760]: W1027 08:24:49.377847 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.378044 kubelet[2760]: E1027 08:24:49.377875 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.378044 kubelet[2760]: E1027 08:24:49.378036 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.378044 kubelet[2760]: W1027 08:24:49.378043 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.378247 kubelet[2760]: E1027 08:24:49.378066 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.378437 kubelet[2760]: E1027 08:24:49.378417 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.378469 kubelet[2760]: W1027 08:24:49.378436 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.378469 kubelet[2760]: E1027 08:24:49.378452 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.379178 kubelet[2760]: E1027 08:24:49.379161 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.379178 kubelet[2760]: W1027 08:24:49.379174 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.379273 kubelet[2760]: E1027 08:24:49.379191 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.380747 kubelet[2760]: E1027 08:24:49.380727 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.380747 kubelet[2760]: W1027 08:24:49.380743 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.380943 kubelet[2760]: E1027 08:24:49.380763 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:49.381720 kubelet[2760]: E1027 08:24:49.381701 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:49.381807 kubelet[2760]: W1027 08:24:49.381717 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:49.381807 kubelet[2760]: E1027 08:24:49.381761 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.192589 kubelet[2760]: E1027 08:24:50.190305 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:24:50.315469 kubelet[2760]: I1027 08:24:50.315430 2760 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 27 08:24:50.316669 kubelet[2760]: E1027 08:24:50.316641 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:50.363921 kubelet[2760]: E1027 08:24:50.363879 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.363921 kubelet[2760]: W1027 08:24:50.363908 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.363921 kubelet[2760]: E1027 08:24:50.363934 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.364175 kubelet[2760]: E1027 08:24:50.364141 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.364175 kubelet[2760]: W1027 08:24:50.364150 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.364175 kubelet[2760]: E1027 08:24:50.364161 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.364344 kubelet[2760]: E1027 08:24:50.364321 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.364344 kubelet[2760]: W1027 08:24:50.364333 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.364344 kubelet[2760]: E1027 08:24:50.364342 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.364573 kubelet[2760]: E1027 08:24:50.364561 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.364573 kubelet[2760]: W1027 08:24:50.364572 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.364650 kubelet[2760]: E1027 08:24:50.364582 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.364791 kubelet[2760]: E1027 08:24:50.364778 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.364822 kubelet[2760]: W1027 08:24:50.364791 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.364822 kubelet[2760]: E1027 08:24:50.364805 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.364984 kubelet[2760]: E1027 08:24:50.364960 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.364984 kubelet[2760]: W1027 08:24:50.364972 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.364984 kubelet[2760]: E1027 08:24:50.364981 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.365147 kubelet[2760]: E1027 08:24:50.365119 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.365147 kubelet[2760]: W1027 08:24:50.365129 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.365147 kubelet[2760]: E1027 08:24:50.365137 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.365308 kubelet[2760]: E1027 08:24:50.365277 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.365308 kubelet[2760]: W1027 08:24:50.365284 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.365308 kubelet[2760]: E1027 08:24:50.365291 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.365468 kubelet[2760]: E1027 08:24:50.365439 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.365468 kubelet[2760]: W1027 08:24:50.365451 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.365468 kubelet[2760]: E1027 08:24:50.365462 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.365726 kubelet[2760]: E1027 08:24:50.365707 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.365876 kubelet[2760]: W1027 08:24:50.365725 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.365876 kubelet[2760]: E1027 08:24:50.365740 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.366039 kubelet[2760]: E1027 08:24:50.366020 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.366107 kubelet[2760]: W1027 08:24:50.366082 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.366183 kubelet[2760]: E1027 08:24:50.366110 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.366614 kubelet[2760]: E1027 08:24:50.366593 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.366691 kubelet[2760]: W1027 08:24:50.366613 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.366691 kubelet[2760]: E1027 08:24:50.366632 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.366940 kubelet[2760]: E1027 08:24:50.366910 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.366940 kubelet[2760]: W1027 08:24:50.366930 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.367043 kubelet[2760]: E1027 08:24:50.366945 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.367132 kubelet[2760]: E1027 08:24:50.367117 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.367132 kubelet[2760]: W1027 08:24:50.367127 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.367230 kubelet[2760]: E1027 08:24:50.367137 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.367338 kubelet[2760]: E1027 08:24:50.367308 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.367338 kubelet[2760]: W1027 08:24:50.367320 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.367338 kubelet[2760]: E1027 08:24:50.367333 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.379213 kubelet[2760]: E1027 08:24:50.379142 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.379213 kubelet[2760]: W1027 08:24:50.379183 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.379213 kubelet[2760]: E1027 08:24:50.379220 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.379677 kubelet[2760]: E1027 08:24:50.379582 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.379677 kubelet[2760]: W1027 08:24:50.379602 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.379677 kubelet[2760]: E1027 08:24:50.379646 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.379885 kubelet[2760]: E1027 08:24:50.379868 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.379940 kubelet[2760]: W1027 08:24:50.379883 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.379940 kubelet[2760]: E1027 08:24:50.379910 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.380095 kubelet[2760]: E1027 08:24:50.380081 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.380095 kubelet[2760]: W1027 08:24:50.380091 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.380194 kubelet[2760]: E1027 08:24:50.380104 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.380244 kubelet[2760]: E1027 08:24:50.380239 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.380288 kubelet[2760]: W1027 08:24:50.380245 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.380288 kubelet[2760]: E1027 08:24:50.380253 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.380430 kubelet[2760]: E1027 08:24:50.380416 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.380430 kubelet[2760]: W1027 08:24:50.380428 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.381073 kubelet[2760]: E1027 08:24:50.380436 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.381073 kubelet[2760]: E1027 08:24:50.380738 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.381073 kubelet[2760]: W1027 08:24:50.380758 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.381073 kubelet[2760]: E1027 08:24:50.380780 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.381073 kubelet[2760]: E1027 08:24:50.381039 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.381073 kubelet[2760]: W1027 08:24:50.381051 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.381073 kubelet[2760]: E1027 08:24:50.381079 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.381362 kubelet[2760]: E1027 08:24:50.381287 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.381362 kubelet[2760]: W1027 08:24:50.381296 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.381362 kubelet[2760]: E1027 08:24:50.381317 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.381487 kubelet[2760]: E1027 08:24:50.381467 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.381487 kubelet[2760]: W1027 08:24:50.381474 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.381647 kubelet[2760]: E1027 08:24:50.381528 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.381868 kubelet[2760]: E1027 08:24:50.381674 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.381868 kubelet[2760]: W1027 08:24:50.381681 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.381868 kubelet[2760]: E1027 08:24:50.381693 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.382307 kubelet[2760]: E1027 08:24:50.381944 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.382307 kubelet[2760]: W1027 08:24:50.381960 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.382307 kubelet[2760]: E1027 08:24:50.381983 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.382534 kubelet[2760]: E1027 08:24:50.382510 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.382945 kubelet[2760]: W1027 08:24:50.382643 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.382945 kubelet[2760]: E1027 08:24:50.382676 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.383048 kubelet[2760]: E1027 08:24:50.382986 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.383048 kubelet[2760]: W1027 08:24:50.383001 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.383048 kubelet[2760]: E1027 08:24:50.383018 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.383203 kubelet[2760]: E1027 08:24:50.383188 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.383248 kubelet[2760]: W1027 08:24:50.383205 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.383248 kubelet[2760]: E1027 08:24:50.383217 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.383463 kubelet[2760]: E1027 08:24:50.383448 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.383463 kubelet[2760]: W1027 08:24:50.383462 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.383597 kubelet[2760]: E1027 08:24:50.383488 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.383974 kubelet[2760]: E1027 08:24:50.383944 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.383974 kubelet[2760]: W1027 08:24:50.383965 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.384071 kubelet[2760]: E1027 08:24:50.383984 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:50.384234 kubelet[2760]: E1027 08:24:50.384212 2760 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 27 08:24:50.384234 kubelet[2760]: W1027 08:24:50.384229 2760 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 27 08:24:50.384336 kubelet[2760]: E1027 08:24:50.384242 2760 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 27 08:24:51.519582 containerd[1602]: time="2025-10-27T08:24:51.518691221Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:51.519582 containerd[1602]: time="2025-10-27T08:24:51.519389190Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4446754" Oct 27 08:24:51.520085 containerd[1602]: time="2025-10-27T08:24:51.519603299Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:51.522173 containerd[1602]: time="2025-10-27T08:24:51.522108645Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:51.522825 containerd[1602]: time="2025-10-27T08:24:51.522716080Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 2.548403624s" Oct 27 08:24:51.522964 containerd[1602]: time="2025-10-27T08:24:51.522945533Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Oct 27 08:24:51.527143 containerd[1602]: time="2025-10-27T08:24:51.527102934Z" level=info msg="CreateContainer within sandbox \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Oct 27 08:24:51.538716 containerd[1602]: time="2025-10-27T08:24:51.537755168Z" level=info msg="Container 8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:51.555072 containerd[1602]: time="2025-10-27T08:24:51.555012151Z" level=info msg="CreateContainer within sandbox \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\"" Oct 27 08:24:51.556304 containerd[1602]: time="2025-10-27T08:24:51.555921688Z" level=info msg="StartContainer for \"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\"" Oct 27 08:24:51.558434 containerd[1602]: time="2025-10-27T08:24:51.558372685Z" level=info msg="connecting to shim 8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8" address="unix:///run/containerd/s/2bc0edaee80744c79d9e57f4cec1b254f62ce3fc76401b8d89d5f7912e030b89" protocol=ttrpc version=3 Oct 27 08:24:51.593825 systemd[1]: Started cri-containerd-8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8.scope - libcontainer container 8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8. Oct 27 08:24:51.646547 containerd[1602]: time="2025-10-27T08:24:51.645203910Z" level=info msg="StartContainer for \"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\" returns successfully" Oct 27 08:24:51.661431 systemd[1]: cri-containerd-8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8.scope: Deactivated successfully. Oct 27 08:24:51.670433 containerd[1602]: time="2025-10-27T08:24:51.670167130Z" level=info msg="received exit event container_id:\"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\" id:\"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\" pid:3483 exited_at:{seconds:1761553491 nanos:664919778}" Oct 27 08:24:51.690509 containerd[1602]: time="2025-10-27T08:24:51.690451456Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\" id:\"8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8\" pid:3483 exited_at:{seconds:1761553491 nanos:664919778}" Oct 27 08:24:51.717095 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8c392444d17c84fbd363d0f126a622d1594db00395f0b766ce3af1c4d3eb1fe8-rootfs.mount: Deactivated successfully. Oct 27 08:24:52.191665 kubelet[2760]: E1027 08:24:52.191064 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:24:52.323216 kubelet[2760]: E1027 08:24:52.322450 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:52.324943 containerd[1602]: time="2025-10-27T08:24:52.324904637Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Oct 27 08:24:52.347579 kubelet[2760]: I1027 08:24:52.346173 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-685cb5f8dc-lbfhv" podStartSLOduration=4.870936801 podStartE2EDuration="7.345184068s" podCreationTimestamp="2025-10-27 08:24:45 +0000 UTC" firstStartedPulling="2025-10-27 08:24:46.499373543 +0000 UTC m=+20.450423142" lastFinishedPulling="2025-10-27 08:24:48.973620798 +0000 UTC m=+22.924670409" observedRunningTime="2025-10-27 08:24:49.34170246 +0000 UTC m=+23.292752141" watchObservedRunningTime="2025-10-27 08:24:52.345184068 +0000 UTC m=+26.296233707" Oct 27 08:24:54.190595 kubelet[2760]: E1027 08:24:54.190541 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:24:55.908793 containerd[1602]: time="2025-10-27T08:24:55.908718124Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:55.909922 containerd[1602]: time="2025-10-27T08:24:55.909715599Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70446859" Oct 27 08:24:55.910532 containerd[1602]: time="2025-10-27T08:24:55.910479693Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:55.912919 containerd[1602]: time="2025-10-27T08:24:55.912852197Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:24:55.913890 containerd[1602]: time="2025-10-27T08:24:55.913760888Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 3.588624045s" Oct 27 08:24:55.913890 containerd[1602]: time="2025-10-27T08:24:55.913794486Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Oct 27 08:24:55.944318 containerd[1602]: time="2025-10-27T08:24:55.944243203Z" level=info msg="CreateContainer within sandbox \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Oct 27 08:24:55.968540 containerd[1602]: time="2025-10-27T08:24:55.967761134Z" level=info msg="Container 80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:24:55.977221 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3144724079.mount: Deactivated successfully. Oct 27 08:24:55.985804 containerd[1602]: time="2025-10-27T08:24:55.985699737Z" level=info msg="CreateContainer within sandbox \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\"" Oct 27 08:24:55.988153 containerd[1602]: time="2025-10-27T08:24:55.986999999Z" level=info msg="StartContainer for \"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\"" Oct 27 08:24:55.988646 containerd[1602]: time="2025-10-27T08:24:55.988616038Z" level=info msg="connecting to shim 80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20" address="unix:///run/containerd/s/2bc0edaee80744c79d9e57f4cec1b254f62ce3fc76401b8d89d5f7912e030b89" protocol=ttrpc version=3 Oct 27 08:24:56.022975 systemd[1]: Started cri-containerd-80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20.scope - libcontainer container 80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20. Oct 27 08:24:56.073469 containerd[1602]: time="2025-10-27T08:24:56.073429962Z" level=info msg="StartContainer for \"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\" returns successfully" Oct 27 08:24:56.192274 kubelet[2760]: E1027 08:24:56.190913 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:24:56.352855 kubelet[2760]: E1027 08:24:56.349841 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:56.736588 systemd[1]: cri-containerd-80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20.scope: Deactivated successfully. Oct 27 08:24:56.736998 systemd[1]: cri-containerd-80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20.scope: Consumed 585ms CPU time, 165.9M memory peak, 6.6M read from disk, 171.3M written to disk. Oct 27 08:24:56.797859 containerd[1602]: time="2025-10-27T08:24:56.797806408Z" level=info msg="received exit event container_id:\"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\" id:\"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\" pid:3542 exited_at:{seconds:1761553496 nanos:797014615}" Oct 27 08:24:56.807124 containerd[1602]: time="2025-10-27T08:24:56.807047140Z" level=info msg="TaskExit event in podsandbox handler container_id:\"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\" id:\"80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20\" pid:3542 exited_at:{seconds:1761553496 nanos:797014615}" Oct 27 08:24:56.824979 kubelet[2760]: I1027 08:24:56.824713 2760 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Oct 27 08:24:56.881410 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-80809442dd6219cda0a072ba86ad16b16967739aa93f38cb1698b9f6673e4b20-rootfs.mount: Deactivated successfully. Oct 27 08:24:56.916372 systemd[1]: Created slice kubepods-burstable-podc45c241d_abf9_4ea1_a19c_427b39d582a2.slice - libcontainer container kubepods-burstable-podc45c241d_abf9_4ea1_a19c_427b39d582a2.slice. Oct 27 08:24:56.934866 kubelet[2760]: I1027 08:24:56.933310 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-ca-bundle\") pod \"whisker-66595dc58b-fqdkl\" (UID: \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\") " pod="calico-system/whisker-66595dc58b-fqdkl" Oct 27 08:24:56.934866 kubelet[2760]: I1027 08:24:56.933349 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c519ad97-124f-4ebf-ae30-a61647720f7b-calico-apiserver-certs\") pod \"calico-apiserver-55869988d7-jkds2\" (UID: \"c519ad97-124f-4ebf-ae30-a61647720f7b\") " pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" Oct 27 08:24:56.934866 kubelet[2760]: I1027 08:24:56.933368 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q2gc\" (UniqueName: \"kubernetes.io/projected/c519ad97-124f-4ebf-ae30-a61647720f7b-kube-api-access-4q2gc\") pod \"calico-apiserver-55869988d7-jkds2\" (UID: \"c519ad97-124f-4ebf-ae30-a61647720f7b\") " pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" Oct 27 08:24:56.934866 kubelet[2760]: I1027 08:24:56.933389 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99mhv\" (UniqueName: \"kubernetes.io/projected/b6db6e46-66a7-435c-a200-9c21f50724bf-kube-api-access-99mhv\") pod \"calico-apiserver-55869988d7-t5jfb\" (UID: \"b6db6e46-66a7-435c-a200-9c21f50724bf\") " pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" Oct 27 08:24:56.934866 kubelet[2760]: I1027 08:24:56.933404 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c45c241d-abf9-4ea1-a19c-427b39d582a2-config-volume\") pod \"coredns-668d6bf9bc-qdwvh\" (UID: \"c45c241d-abf9-4ea1-a19c-427b39d582a2\") " pod="kube-system/coredns-668d6bf9bc-qdwvh" Oct 27 08:24:56.933992 systemd[1]: Created slice kubepods-besteffort-podffef7bc7_dd37_414d_945d_436bb2e2b656.slice - libcontainer container kubepods-besteffort-podffef7bc7_dd37_414d_945d_436bb2e2b656.slice. Oct 27 08:24:56.935252 kubelet[2760]: I1027 08:24:56.933421 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkstz\" (UniqueName: \"kubernetes.io/projected/ce8b823a-2ab2-4338-a4da-ee011d8881cc-kube-api-access-lkstz\") pod \"whisker-66595dc58b-fqdkl\" (UID: \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\") " pod="calico-system/whisker-66595dc58b-fqdkl" Oct 27 08:24:56.935252 kubelet[2760]: I1027 08:24:56.933439 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6hcw\" (UniqueName: \"kubernetes.io/projected/ffef7bc7-dd37-414d-945d-436bb2e2b656-kube-api-access-t6hcw\") pod \"calico-kube-controllers-9b7544668-z2ztz\" (UID: \"ffef7bc7-dd37-414d-945d-436bb2e2b656\") " pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" Oct 27 08:24:56.935252 kubelet[2760]: I1027 08:24:56.933459 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/b6db6e46-66a7-435c-a200-9c21f50724bf-calico-apiserver-certs\") pod \"calico-apiserver-55869988d7-t5jfb\" (UID: \"b6db6e46-66a7-435c-a200-9c21f50724bf\") " pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" Oct 27 08:24:56.935252 kubelet[2760]: I1027 08:24:56.933476 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51805d3a-45e5-4447-9079-e21004ab5665-config\") pod \"goldmane-666569f655-qlx6r\" (UID: \"51805d3a-45e5-4447-9079-e21004ab5665\") " pod="calico-system/goldmane-666569f655-qlx6r" Oct 27 08:24:56.935252 kubelet[2760]: I1027 08:24:56.933491 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51805d3a-45e5-4447-9079-e21004ab5665-goldmane-ca-bundle\") pod \"goldmane-666569f655-qlx6r\" (UID: \"51805d3a-45e5-4447-9079-e21004ab5665\") " pod="calico-system/goldmane-666569f655-qlx6r" Oct 27 08:24:56.935385 kubelet[2760]: I1027 08:24:56.933507 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/51805d3a-45e5-4447-9079-e21004ab5665-goldmane-key-pair\") pod \"goldmane-666569f655-qlx6r\" (UID: \"51805d3a-45e5-4447-9079-e21004ab5665\") " pod="calico-system/goldmane-666569f655-qlx6r" Oct 27 08:24:56.935750 kubelet[2760]: I1027 08:24:56.935491 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x79st\" (UniqueName: \"kubernetes.io/projected/51805d3a-45e5-4447-9079-e21004ab5665-kube-api-access-x79st\") pod \"goldmane-666569f655-qlx6r\" (UID: \"51805d3a-45e5-4447-9079-e21004ab5665\") " pod="calico-system/goldmane-666569f655-qlx6r" Oct 27 08:24:56.935750 kubelet[2760]: I1027 08:24:56.935554 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5fgm\" (UniqueName: \"kubernetes.io/projected/ce579d70-472c-4f85-ab9b-4d5886747e0d-kube-api-access-b5fgm\") pod \"coredns-668d6bf9bc-mrp2n\" (UID: \"ce579d70-472c-4f85-ab9b-4d5886747e0d\") " pod="kube-system/coredns-668d6bf9bc-mrp2n" Oct 27 08:24:56.935750 kubelet[2760]: I1027 08:24:56.935573 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffef7bc7-dd37-414d-945d-436bb2e2b656-tigera-ca-bundle\") pod \"calico-kube-controllers-9b7544668-z2ztz\" (UID: \"ffef7bc7-dd37-414d-945d-436bb2e2b656\") " pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" Oct 27 08:24:56.935750 kubelet[2760]: I1027 08:24:56.935607 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr94x\" (UniqueName: \"kubernetes.io/projected/c45c241d-abf9-4ea1-a19c-427b39d582a2-kube-api-access-cr94x\") pod \"coredns-668d6bf9bc-qdwvh\" (UID: \"c45c241d-abf9-4ea1-a19c-427b39d582a2\") " pod="kube-system/coredns-668d6bf9bc-qdwvh" Oct 27 08:24:56.935750 kubelet[2760]: I1027 08:24:56.935625 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce579d70-472c-4f85-ab9b-4d5886747e0d-config-volume\") pod \"coredns-668d6bf9bc-mrp2n\" (UID: \"ce579d70-472c-4f85-ab9b-4d5886747e0d\") " pod="kube-system/coredns-668d6bf9bc-mrp2n" Oct 27 08:24:56.935922 kubelet[2760]: I1027 08:24:56.935644 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-backend-key-pair\") pod \"whisker-66595dc58b-fqdkl\" (UID: \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\") " pod="calico-system/whisker-66595dc58b-fqdkl" Oct 27 08:24:56.948712 systemd[1]: Created slice kubepods-burstable-podce579d70_472c_4f85_ab9b_4d5886747e0d.slice - libcontainer container kubepods-burstable-podce579d70_472c_4f85_ab9b_4d5886747e0d.slice. Oct 27 08:24:56.957789 systemd[1]: Created slice kubepods-besteffort-podb6db6e46_66a7_435c_a200_9c21f50724bf.slice - libcontainer container kubepods-besteffort-podb6db6e46_66a7_435c_a200_9c21f50724bf.slice. Oct 27 08:24:56.968570 systemd[1]: Created slice kubepods-besteffort-pod51805d3a_45e5_4447_9079_e21004ab5665.slice - libcontainer container kubepods-besteffort-pod51805d3a_45e5_4447_9079_e21004ab5665.slice. Oct 27 08:24:56.981623 systemd[1]: Created slice kubepods-besteffort-podce8b823a_2ab2_4338_a4da_ee011d8881cc.slice - libcontainer container kubepods-besteffort-podce8b823a_2ab2_4338_a4da_ee011d8881cc.slice. Oct 27 08:24:56.987422 systemd[1]: Created slice kubepods-besteffort-podc519ad97_124f_4ebf_ae30_a61647720f7b.slice - libcontainer container kubepods-besteffort-podc519ad97_124f_4ebf_ae30_a61647720f7b.slice. Oct 27 08:24:57.229968 kubelet[2760]: E1027 08:24:57.229906 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:57.231841 containerd[1602]: time="2025-10-27T08:24:57.231793276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qdwvh,Uid:c45c241d-abf9-4ea1-a19c-427b39d582a2,Namespace:kube-system,Attempt:0,}" Oct 27 08:24:57.245224 containerd[1602]: time="2025-10-27T08:24:57.244867797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-9b7544668-z2ztz,Uid:ffef7bc7-dd37-414d-945d-436bb2e2b656,Namespace:calico-system,Attempt:0,}" Oct 27 08:24:57.254814 kubelet[2760]: E1027 08:24:57.254760 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:57.268717 containerd[1602]: time="2025-10-27T08:24:57.268660973Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-mrp2n,Uid:ce579d70-472c-4f85-ab9b-4d5886747e0d,Namespace:kube-system,Attempt:0,}" Oct 27 08:24:57.274413 containerd[1602]: time="2025-10-27T08:24:57.274042576Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-t5jfb,Uid:b6db6e46-66a7-435c-a200-9c21f50724bf,Namespace:calico-apiserver,Attempt:0,}" Oct 27 08:24:57.277649 containerd[1602]: time="2025-10-27T08:24:57.277411992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-qlx6r,Uid:51805d3a-45e5-4447-9079-e21004ab5665,Namespace:calico-system,Attempt:0,}" Oct 27 08:24:57.341148 containerd[1602]: time="2025-10-27T08:24:57.340786371Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-jkds2,Uid:c519ad97-124f-4ebf-ae30-a61647720f7b,Namespace:calico-apiserver,Attempt:0,}" Oct 27 08:24:57.354497 containerd[1602]: time="2025-10-27T08:24:57.354437036Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-66595dc58b-fqdkl,Uid:ce8b823a-2ab2-4338-a4da-ee011d8881cc,Namespace:calico-system,Attempt:0,}" Oct 27 08:24:57.401382 kubelet[2760]: E1027 08:24:57.400885 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:24:57.419701 containerd[1602]: time="2025-10-27T08:24:57.419653958Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Oct 27 08:24:57.621426 containerd[1602]: time="2025-10-27T08:24:57.620797384Z" level=error msg="Failed to destroy network for sandbox \"beedb060712951f10a4530691024153ed1a5767258330b8e5697de452d2a816f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.625054 containerd[1602]: time="2025-10-27T08:24:57.624974178Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-9b7544668-z2ztz,Uid:ffef7bc7-dd37-414d-945d-436bb2e2b656,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"beedb060712951f10a4530691024153ed1a5767258330b8e5697de452d2a816f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.625978 kubelet[2760]: E1027 08:24:57.625718 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"beedb060712951f10a4530691024153ed1a5767258330b8e5697de452d2a816f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.626264 kubelet[2760]: E1027 08:24:57.626230 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"beedb060712951f10a4530691024153ed1a5767258330b8e5697de452d2a816f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" Oct 27 08:24:57.626619 kubelet[2760]: E1027 08:24:57.626381 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"beedb060712951f10a4530691024153ed1a5767258330b8e5697de452d2a816f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" Oct 27 08:24:57.626797 kubelet[2760]: E1027 08:24:57.626744 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-9b7544668-z2ztz_calico-system(ffef7bc7-dd37-414d-945d-436bb2e2b656)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-9b7544668-z2ztz_calico-system(ffef7bc7-dd37-414d-945d-436bb2e2b656)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"beedb060712951f10a4530691024153ed1a5767258330b8e5697de452d2a816f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:24:57.690574 containerd[1602]: time="2025-10-27T08:24:57.690499808Z" level=error msg="Failed to destroy network for sandbox \"c218af4eb683d5a4cfecb1eb36585e587b5b06a1852d97cb6c8423431b8a4775\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.692357 containerd[1602]: time="2025-10-27T08:24:57.692301481Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-mrp2n,Uid:ce579d70-472c-4f85-ab9b-4d5886747e0d,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c218af4eb683d5a4cfecb1eb36585e587b5b06a1852d97cb6c8423431b8a4775\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.693414 kubelet[2760]: E1027 08:24:57.692761 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c218af4eb683d5a4cfecb1eb36585e587b5b06a1852d97cb6c8423431b8a4775\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.693414 kubelet[2760]: E1027 08:24:57.693222 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c218af4eb683d5a4cfecb1eb36585e587b5b06a1852d97cb6c8423431b8a4775\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-mrp2n" Oct 27 08:24:57.693414 kubelet[2760]: E1027 08:24:57.693255 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c218af4eb683d5a4cfecb1eb36585e587b5b06a1852d97cb6c8423431b8a4775\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-mrp2n" Oct 27 08:24:57.694130 kubelet[2760]: E1027 08:24:57.693324 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-mrp2n_kube-system(ce579d70-472c-4f85-ab9b-4d5886747e0d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-mrp2n_kube-system(ce579d70-472c-4f85-ab9b-4d5886747e0d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c218af4eb683d5a4cfecb1eb36585e587b5b06a1852d97cb6c8423431b8a4775\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-mrp2n" podUID="ce579d70-472c-4f85-ab9b-4d5886747e0d" Oct 27 08:24:57.704781 containerd[1602]: time="2025-10-27T08:24:57.704718977Z" level=error msg="Failed to destroy network for sandbox \"e96769b1aaa85e97dbc8dfdabd1d7eaf1df982074b892bcb3a3b26d13fd5f933\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.706026 containerd[1602]: time="2025-10-27T08:24:57.705640740Z" level=error msg="Failed to destroy network for sandbox \"f795c4a7e2de828e8eb46032736ff610050399c61bda98103fe6516c91dda3b1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.706026 containerd[1602]: time="2025-10-27T08:24:57.705927883Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qdwvh,Uid:c45c241d-abf9-4ea1-a19c-427b39d582a2,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e96769b1aaa85e97dbc8dfdabd1d7eaf1df982074b892bcb3a3b26d13fd5f933\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.706558 containerd[1602]: time="2025-10-27T08:24:57.706346311Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-66595dc58b-fqdkl,Uid:ce8b823a-2ab2-4338-a4da-ee011d8881cc,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f795c4a7e2de828e8eb46032736ff610050399c61bda98103fe6516c91dda3b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.706660 kubelet[2760]: E1027 08:24:57.706460 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e96769b1aaa85e97dbc8dfdabd1d7eaf1df982074b892bcb3a3b26d13fd5f933\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.706660 kubelet[2760]: E1027 08:24:57.706538 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f795c4a7e2de828e8eb46032736ff610050399c61bda98103fe6516c91dda3b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.706660 kubelet[2760]: E1027 08:24:57.706589 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f795c4a7e2de828e8eb46032736ff610050399c61bda98103fe6516c91dda3b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-66595dc58b-fqdkl" Oct 27 08:24:57.706660 kubelet[2760]: E1027 08:24:57.706610 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f795c4a7e2de828e8eb46032736ff610050399c61bda98103fe6516c91dda3b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-66595dc58b-fqdkl" Oct 27 08:24:57.707506 kubelet[2760]: E1027 08:24:57.706671 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-66595dc58b-fqdkl_calico-system(ce8b823a-2ab2-4338-a4da-ee011d8881cc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-66595dc58b-fqdkl_calico-system(ce8b823a-2ab2-4338-a4da-ee011d8881cc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f795c4a7e2de828e8eb46032736ff610050399c61bda98103fe6516c91dda3b1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-66595dc58b-fqdkl" podUID="ce8b823a-2ab2-4338-a4da-ee011d8881cc" Oct 27 08:24:57.707506 kubelet[2760]: E1027 08:24:57.706736 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e96769b1aaa85e97dbc8dfdabd1d7eaf1df982074b892bcb3a3b26d13fd5f933\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-qdwvh" Oct 27 08:24:57.707506 kubelet[2760]: E1027 08:24:57.706758 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e96769b1aaa85e97dbc8dfdabd1d7eaf1df982074b892bcb3a3b26d13fd5f933\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-qdwvh" Oct 27 08:24:57.708023 kubelet[2760]: E1027 08:24:57.707030 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-qdwvh_kube-system(c45c241d-abf9-4ea1-a19c-427b39d582a2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-qdwvh_kube-system(c45c241d-abf9-4ea1-a19c-427b39d582a2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e96769b1aaa85e97dbc8dfdabd1d7eaf1df982074b892bcb3a3b26d13fd5f933\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-qdwvh" podUID="c45c241d-abf9-4ea1-a19c-427b39d582a2" Oct 27 08:24:57.711896 containerd[1602]: time="2025-10-27T08:24:57.711814912Z" level=error msg="Failed to destroy network for sandbox \"9296aabce26d4d4b6d045a880ed37a4e57995a51132106b9d13a0a89ea95464b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.712431 containerd[1602]: time="2025-10-27T08:24:57.712322958Z" level=error msg="Failed to destroy network for sandbox \"961e8936cf7a384eb3cffe03d68107e6270b6052fcce02e73607ff6506d45515\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.713463 containerd[1602]: time="2025-10-27T08:24:57.713405644Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-qlx6r,Uid:51805d3a-45e5-4447-9079-e21004ab5665,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9296aabce26d4d4b6d045a880ed37a4e57995a51132106b9d13a0a89ea95464b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.714399 kubelet[2760]: E1027 08:24:57.714348 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9296aabce26d4d4b6d045a880ed37a4e57995a51132106b9d13a0a89ea95464b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.715180 kubelet[2760]: E1027 08:24:57.714933 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"961e8936cf7a384eb3cffe03d68107e6270b6052fcce02e73607ff6506d45515\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.715180 kubelet[2760]: E1027 08:24:57.715069 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"961e8936cf7a384eb3cffe03d68107e6270b6052fcce02e73607ff6506d45515\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" Oct 27 08:24:57.715180 kubelet[2760]: E1027 08:24:57.715092 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"961e8936cf7a384eb3cffe03d68107e6270b6052fcce02e73607ff6506d45515\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" Oct 27 08:24:57.715702 containerd[1602]: time="2025-10-27T08:24:57.714773064Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-t5jfb,Uid:b6db6e46-66a7-435c-a200-9c21f50724bf,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"961e8936cf7a384eb3cffe03d68107e6270b6052fcce02e73607ff6506d45515\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.715850 kubelet[2760]: E1027 08:24:57.715133 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-55869988d7-t5jfb_calico-apiserver(b6db6e46-66a7-435c-a200-9c21f50724bf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-55869988d7-t5jfb_calico-apiserver(b6db6e46-66a7-435c-a200-9c21f50724bf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"961e8936cf7a384eb3cffe03d68107e6270b6052fcce02e73607ff6506d45515\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:24:57.715850 kubelet[2760]: E1027 08:24:57.715012 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9296aabce26d4d4b6d045a880ed37a4e57995a51132106b9d13a0a89ea95464b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-qlx6r" Oct 27 08:24:57.715850 kubelet[2760]: E1027 08:24:57.715182 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9296aabce26d4d4b6d045a880ed37a4e57995a51132106b9d13a0a89ea95464b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-qlx6r" Oct 27 08:24:57.716128 kubelet[2760]: E1027 08:24:57.715204 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-qlx6r_calico-system(51805d3a-45e5-4447-9079-e21004ab5665)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-qlx6r_calico-system(51805d3a-45e5-4447-9079-e21004ab5665)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9296aabce26d4d4b6d045a880ed37a4e57995a51132106b9d13a0a89ea95464b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:24:57.737007 containerd[1602]: time="2025-10-27T08:24:57.736914637Z" level=error msg="Failed to destroy network for sandbox \"02a2e1230775e06f0da347fa6743647fc420cebd3b117c56f2b8cc05f89eac28\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.738063 containerd[1602]: time="2025-10-27T08:24:57.737983548Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-jkds2,Uid:c519ad97-124f-4ebf-ae30-a61647720f7b,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"02a2e1230775e06f0da347fa6743647fc420cebd3b117c56f2b8cc05f89eac28\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.738452 kubelet[2760]: E1027 08:24:57.738395 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"02a2e1230775e06f0da347fa6743647fc420cebd3b117c56f2b8cc05f89eac28\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:57.738598 kubelet[2760]: E1027 08:24:57.738470 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"02a2e1230775e06f0da347fa6743647fc420cebd3b117c56f2b8cc05f89eac28\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" Oct 27 08:24:57.738598 kubelet[2760]: E1027 08:24:57.738491 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"02a2e1230775e06f0da347fa6743647fc420cebd3b117c56f2b8cc05f89eac28\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" Oct 27 08:24:57.738598 kubelet[2760]: E1027 08:24:57.738556 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-55869988d7-jkds2_calico-apiserver(c519ad97-124f-4ebf-ae30-a61647720f7b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-55869988d7-jkds2_calico-apiserver(c519ad97-124f-4ebf-ae30-a61647720f7b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"02a2e1230775e06f0da347fa6743647fc420cebd3b117c56f2b8cc05f89eac28\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:24:58.197708 systemd[1]: Created slice kubepods-besteffort-podb9e97248_79da_4981_820e_d3ae986590e1.slice - libcontainer container kubepods-besteffort-podb9e97248_79da_4981_820e_d3ae986590e1.slice. Oct 27 08:24:58.201463 containerd[1602]: time="2025-10-27T08:24:58.201417461Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fvjzl,Uid:b9e97248-79da-4981-820e-d3ae986590e1,Namespace:calico-system,Attempt:0,}" Oct 27 08:24:58.273145 containerd[1602]: time="2025-10-27T08:24:58.273091294Z" level=error msg="Failed to destroy network for sandbox \"30c2162001db7f2315a144391cf365e97832dae8adc02b5cb490a7ae103156a8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:58.275657 systemd[1]: run-netns-cni\x2d559178ab\x2d11fe\x2dd6fb\x2d0df4\x2d61db18e8281c.mount: Deactivated successfully. Oct 27 08:24:58.277456 containerd[1602]: time="2025-10-27T08:24:58.277412348Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fvjzl,Uid:b9e97248-79da-4981-820e-d3ae986590e1,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"30c2162001db7f2315a144391cf365e97832dae8adc02b5cb490a7ae103156a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:58.278081 kubelet[2760]: E1027 08:24:58.278033 2760 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"30c2162001db7f2315a144391cf365e97832dae8adc02b5cb490a7ae103156a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 27 08:24:58.278442 kubelet[2760]: E1027 08:24:58.278096 2760 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"30c2162001db7f2315a144391cf365e97832dae8adc02b5cb490a7ae103156a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:58.278442 kubelet[2760]: E1027 08:24:58.278123 2760 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"30c2162001db7f2315a144391cf365e97832dae8adc02b5cb490a7ae103156a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-fvjzl" Oct 27 08:24:58.278442 kubelet[2760]: E1027 08:24:58.278195 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"30c2162001db7f2315a144391cf365e97832dae8adc02b5cb490a7ae103156a8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:25:04.103338 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3390950430.mount: Deactivated successfully. Oct 27 08:25:04.231780 containerd[1602]: time="2025-10-27T08:25:04.231699590Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156883675" Oct 27 08:25:04.234550 containerd[1602]: time="2025-10-27T08:25:04.233779823Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:25:04.247320 containerd[1602]: time="2025-10-27T08:25:04.247197389Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:25:04.249107 containerd[1602]: time="2025-10-27T08:25:04.248996893Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 27 08:25:04.257569 containerd[1602]: time="2025-10-27T08:25:04.257300914Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 6.829738221s" Oct 27 08:25:04.257569 containerd[1602]: time="2025-10-27T08:25:04.257475977Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Oct 27 08:25:04.292580 containerd[1602]: time="2025-10-27T08:25:04.292216057Z" level=info msg="CreateContainer within sandbox \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Oct 27 08:25:04.425043 containerd[1602]: time="2025-10-27T08:25:04.424754478Z" level=info msg="Container 1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:25:04.426882 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2330295176.mount: Deactivated successfully. Oct 27 08:25:04.493260 containerd[1602]: time="2025-10-27T08:25:04.493174652Z" level=info msg="CreateContainer within sandbox \"46822bcbf2adfb104419e1aa25674d120ac8eecc861fd1ffb439382387ff866b\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\"" Oct 27 08:25:04.494978 containerd[1602]: time="2025-10-27T08:25:04.494716757Z" level=info msg="StartContainer for \"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\"" Oct 27 08:25:04.501143 containerd[1602]: time="2025-10-27T08:25:04.501080787Z" level=info msg="connecting to shim 1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c" address="unix:///run/containerd/s/2bc0edaee80744c79d9e57f4cec1b254f62ce3fc76401b8d89d5f7912e030b89" protocol=ttrpc version=3 Oct 27 08:25:04.578884 systemd[1]: Started cri-containerd-1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c.scope - libcontainer container 1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c. Oct 27 08:25:04.715647 containerd[1602]: time="2025-10-27T08:25:04.714157114Z" level=info msg="StartContainer for \"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\" returns successfully" Oct 27 08:25:04.837953 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Oct 27 08:25:04.840099 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Oct 27 08:25:05.124695 kubelet[2760]: I1027 08:25:05.124643 2760 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkstz\" (UniqueName: \"kubernetes.io/projected/ce8b823a-2ab2-4338-a4da-ee011d8881cc-kube-api-access-lkstz\") pod \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\" (UID: \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\") " Oct 27 08:25:05.127113 kubelet[2760]: I1027 08:25:05.125594 2760 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-ca-bundle\") pod \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\" (UID: \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\") " Oct 27 08:25:05.127113 kubelet[2760]: I1027 08:25:05.125663 2760 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-backend-key-pair\") pod \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\" (UID: \"ce8b823a-2ab2-4338-a4da-ee011d8881cc\") " Oct 27 08:25:05.128078 kubelet[2760]: I1027 08:25:05.128019 2760 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "ce8b823a-2ab2-4338-a4da-ee011d8881cc" (UID: "ce8b823a-2ab2-4338-a4da-ee011d8881cc"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Oct 27 08:25:05.133923 systemd[1]: var-lib-kubelet-pods-ce8b823a\x2d2ab2\x2d4338\x2da4da\x2dee011d8881cc-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dlkstz.mount: Deactivated successfully. Oct 27 08:25:05.140721 kubelet[2760]: I1027 08:25:05.135974 2760 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce8b823a-2ab2-4338-a4da-ee011d8881cc-kube-api-access-lkstz" (OuterVolumeSpecName: "kube-api-access-lkstz") pod "ce8b823a-2ab2-4338-a4da-ee011d8881cc" (UID: "ce8b823a-2ab2-4338-a4da-ee011d8881cc"). InnerVolumeSpecName "kube-api-access-lkstz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Oct 27 08:25:05.144780 kubelet[2760]: I1027 08:25:05.144708 2760 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "ce8b823a-2ab2-4338-a4da-ee011d8881cc" (UID: "ce8b823a-2ab2-4338-a4da-ee011d8881cc"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Oct 27 08:25:05.146162 systemd[1]: var-lib-kubelet-pods-ce8b823a\x2d2ab2\x2d4338\x2da4da\x2dee011d8881cc-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Oct 27 08:25:05.226656 kubelet[2760]: I1027 08:25:05.226593 2760 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lkstz\" (UniqueName: \"kubernetes.io/projected/ce8b823a-2ab2-4338-a4da-ee011d8881cc-kube-api-access-lkstz\") on node \"ci-9999.9.9-k-4f7b9e69b1\" DevicePath \"\"" Oct 27 08:25:05.226656 kubelet[2760]: I1027 08:25:05.226646 2760 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-ca-bundle\") on node \"ci-9999.9.9-k-4f7b9e69b1\" DevicePath \"\"" Oct 27 08:25:05.226656 kubelet[2760]: I1027 08:25:05.226661 2760 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ce8b823a-2ab2-4338-a4da-ee011d8881cc-whisker-backend-key-pair\") on node \"ci-9999.9.9-k-4f7b9e69b1\" DevicePath \"\"" Oct 27 08:25:05.486272 kubelet[2760]: E1027 08:25:05.485834 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:05.498972 systemd[1]: Removed slice kubepods-besteffort-podce8b823a_2ab2_4338_a4da_ee011d8881cc.slice - libcontainer container kubepods-besteffort-podce8b823a_2ab2_4338_a4da_ee011d8881cc.slice. Oct 27 08:25:05.540631 kubelet[2760]: I1027 08:25:05.540254 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-txx29" podStartSLOduration=1.8158023920000002 podStartE2EDuration="19.540234065s" podCreationTimestamp="2025-10-27 08:24:46 +0000 UTC" firstStartedPulling="2025-10-27 08:24:46.536653963 +0000 UTC m=+20.487703561" lastFinishedPulling="2025-10-27 08:25:04.261085623 +0000 UTC m=+38.212135234" observedRunningTime="2025-10-27 08:25:05.522349357 +0000 UTC m=+39.473398975" watchObservedRunningTime="2025-10-27 08:25:05.540234065 +0000 UTC m=+39.491283711" Oct 27 08:25:05.637726 systemd[1]: Created slice kubepods-besteffort-pod50c0829b_498d_4ffb_8c88_828cd743d3a0.slice - libcontainer container kubepods-besteffort-pod50c0829b_498d_4ffb_8c88_828cd743d3a0.slice. Oct 27 08:25:05.731121 kubelet[2760]: I1027 08:25:05.731044 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50c0829b-498d-4ffb-8c88-828cd743d3a0-whisker-ca-bundle\") pod \"whisker-84f885dcfb-4c4ln\" (UID: \"50c0829b-498d-4ffb-8c88-828cd743d3a0\") " pod="calico-system/whisker-84f885dcfb-4c4ln" Oct 27 08:25:05.731429 kubelet[2760]: I1027 08:25:05.731408 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw5t4\" (UniqueName: \"kubernetes.io/projected/50c0829b-498d-4ffb-8c88-828cd743d3a0-kube-api-access-fw5t4\") pod \"whisker-84f885dcfb-4c4ln\" (UID: \"50c0829b-498d-4ffb-8c88-828cd743d3a0\") " pod="calico-system/whisker-84f885dcfb-4c4ln" Oct 27 08:25:05.731606 kubelet[2760]: I1027 08:25:05.731505 2760 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/50c0829b-498d-4ffb-8c88-828cd743d3a0-whisker-backend-key-pair\") pod \"whisker-84f885dcfb-4c4ln\" (UID: \"50c0829b-498d-4ffb-8c88-828cd743d3a0\") " pod="calico-system/whisker-84f885dcfb-4c4ln" Oct 27 08:25:05.943287 containerd[1602]: time="2025-10-27T08:25:05.943214374Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-84f885dcfb-4c4ln,Uid:50c0829b-498d-4ffb-8c88-828cd743d3a0,Namespace:calico-system,Attempt:0,}" Oct 27 08:25:06.195607 kubelet[2760]: I1027 08:25:06.195099 2760 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce8b823a-2ab2-4338-a4da-ee011d8881cc" path="/var/lib/kubelet/pods/ce8b823a-2ab2-4338-a4da-ee011d8881cc/volumes" Oct 27 08:25:06.315872 systemd-networkd[1494]: calif19535d5367: Link UP Oct 27 08:25:06.316320 systemd-networkd[1494]: calif19535d5367: Gained carrier Oct 27 08:25:06.340434 containerd[1602]: 2025-10-27 08:25:05.986 [INFO][3873] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:06.340434 containerd[1602]: 2025-10-27 08:25:06.027 [INFO][3873] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0 whisker-84f885dcfb- calico-system 50c0829b-498d-4ffb-8c88-828cd743d3a0 895 0 2025-10-27 08:25:05 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:84f885dcfb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 whisker-84f885dcfb-4c4ln eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calif19535d5367 [] [] }} ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-" Oct 27 08:25:06.340434 containerd[1602]: 2025-10-27 08:25:06.027 [INFO][3873] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.340434 containerd[1602]: 2025-10-27 08:25:06.208 [INFO][3885] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" HandleID="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.210 [INFO][3885] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" HandleID="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fcb0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"whisker-84f885dcfb-4c4ln", "timestamp":"2025-10-27 08:25:06.2089239 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.210 [INFO][3885] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.211 [INFO][3885] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.212 [INFO][3885] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.231 [INFO][3885] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.245 [INFO][3885] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.256 [INFO][3885] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.260 [INFO][3885] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341114 containerd[1602]: 2025-10-27 08:25:06.265 [INFO][3885] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.265 [INFO][3885] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.268 [INFO][3885] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.277 [INFO][3885] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.289 [INFO][3885] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.1/26] block=192.168.14.0/26 handle="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.289 [INFO][3885] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.1/26] handle="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.289 [INFO][3885] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:06.341440 containerd[1602]: 2025-10-27 08:25:06.289 [INFO][3885] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.1/26] IPv6=[] ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" HandleID="k8s-pod-network.38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.342998 containerd[1602]: 2025-10-27 08:25:06.294 [INFO][3873] cni-plugin/k8s.go 418: Populated endpoint ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0", GenerateName:"whisker-84f885dcfb-", Namespace:"calico-system", SelfLink:"", UID:"50c0829b-498d-4ffb-8c88-828cd743d3a0", ResourceVersion:"895", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 25, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"84f885dcfb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"whisker-84f885dcfb-4c4ln", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.14.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif19535d5367", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:06.342998 containerd[1602]: 2025-10-27 08:25:06.295 [INFO][3873] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.1/32] ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.343161 containerd[1602]: 2025-10-27 08:25:06.295 [INFO][3873] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif19535d5367 ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.343161 containerd[1602]: 2025-10-27 08:25:06.314 [INFO][3873] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.343329 containerd[1602]: 2025-10-27 08:25:06.316 [INFO][3873] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0", GenerateName:"whisker-84f885dcfb-", Namespace:"calico-system", SelfLink:"", UID:"50c0829b-498d-4ffb-8c88-828cd743d3a0", ResourceVersion:"895", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 25, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"84f885dcfb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e", Pod:"whisker-84f885dcfb-4c4ln", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.14.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif19535d5367", MAC:"16:58:62:83:b8:28", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:06.343445 containerd[1602]: 2025-10-27 08:25:06.334 [INFO][3873] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" Namespace="calico-system" Pod="whisker-84f885dcfb-4c4ln" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-whisker--84f885dcfb--4c4ln-eth0" Oct 27 08:25:06.435759 containerd[1602]: time="2025-10-27T08:25:06.435698884Z" level=info msg="connecting to shim 38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e" address="unix:///run/containerd/s/fb74219043c53ab049971d44b12391c5b9858758e15982fa8b413b1996e6a57a" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:06.501822 kubelet[2760]: E1027 08:25:06.501211 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:06.506908 systemd[1]: Started cri-containerd-38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e.scope - libcontainer container 38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e. Oct 27 08:25:06.799035 containerd[1602]: time="2025-10-27T08:25:06.798764660Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-84f885dcfb-4c4ln,Uid:50c0829b-498d-4ffb-8c88-828cd743d3a0,Namespace:calico-system,Attempt:0,} returns sandbox id \"38b8c8c0ac80b0347d224bf862639d2b3210d333f4b1bf5b4990ec6fa978ad3e\"" Oct 27 08:25:06.813883 containerd[1602]: time="2025-10-27T08:25:06.813820808Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 27 08:25:07.203740 containerd[1602]: time="2025-10-27T08:25:07.203683222Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:07.212621 containerd[1602]: time="2025-10-27T08:25:07.205903858Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 27 08:25:07.213566 containerd[1602]: time="2025-10-27T08:25:07.205910581Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 27 08:25:07.213701 kubelet[2760]: E1027 08:25:07.213097 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 27 08:25:07.213701 kubelet[2760]: E1027 08:25:07.213168 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 27 08:25:07.237511 containerd[1602]: time="2025-10-27T08:25:07.237038398Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\" id:\"452a8b2fae52a2d8b3cccf56769c1c598119eb7ab85e6606e07559c76df2df0a\" pid:4019 exit_status:1 exited_at:{seconds:1761553507 nanos:221686617}" Oct 27 08:25:07.245196 kubelet[2760]: E1027 08:25:07.244662 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d7eec9ae1b0242a8b0b738edf8f38b0b,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fw5t4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-84f885dcfb-4c4ln_calico-system(50c0829b-498d-4ffb-8c88-828cd743d3a0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:07.250785 containerd[1602]: time="2025-10-27T08:25:07.250728981Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 27 08:25:07.513945 kubelet[2760]: E1027 08:25:07.510341 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:07.633406 containerd[1602]: time="2025-10-27T08:25:07.633360862Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\" id:\"50ea419ece6f24be029c069e89d7e7a2dcbb2e56fb07bef04e7f6ae51f2fb819\" pid:4074 exit_status:1 exited_at:{seconds:1761553507 nanos:632865694}" Oct 27 08:25:07.634341 containerd[1602]: time="2025-10-27T08:25:07.634310432Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:07.635766 containerd[1602]: time="2025-10-27T08:25:07.635716423Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 27 08:25:07.635966 containerd[1602]: time="2025-10-27T08:25:07.635926362Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 27 08:25:07.637337 kubelet[2760]: E1027 08:25:07.637160 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 27 08:25:07.637337 kubelet[2760]: E1027 08:25:07.637222 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 27 08:25:07.637646 kubelet[2760]: E1027 08:25:07.637374 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fw5t4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-84f885dcfb-4c4ln_calico-system(50c0829b-498d-4ffb-8c88-828cd743d3a0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:07.638725 kubelet[2760]: E1027 08:25:07.638646 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-84f885dcfb-4c4ln" podUID="50c0829b-498d-4ffb-8c88-828cd743d3a0" Oct 27 08:25:08.192155 containerd[1602]: time="2025-10-27T08:25:08.192030262Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-9b7544668-z2ztz,Uid:ffef7bc7-dd37-414d-945d-436bb2e2b656,Namespace:calico-system,Attempt:0,}" Oct 27 08:25:08.327696 systemd-networkd[1494]: calif19535d5367: Gained IPv6LL Oct 27 08:25:08.411646 systemd-networkd[1494]: cali348d9b4faf5: Link UP Oct 27 08:25:08.415401 systemd-networkd[1494]: cali348d9b4faf5: Gained carrier Oct 27 08:25:08.442593 containerd[1602]: 2025-10-27 08:25:08.243 [INFO][4100] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:08.442593 containerd[1602]: 2025-10-27 08:25:08.263 [INFO][4100] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0 calico-kube-controllers-9b7544668- calico-system ffef7bc7-dd37-414d-945d-436bb2e2b656 822 0 2025-10-27 08:24:46 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:9b7544668 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 calico-kube-controllers-9b7544668-z2ztz eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali348d9b4faf5 [] [] }} ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-" Oct 27 08:25:08.442593 containerd[1602]: 2025-10-27 08:25:08.263 [INFO][4100] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.442593 containerd[1602]: 2025-10-27 08:25:08.331 [INFO][4119] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" HandleID="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.331 [INFO][4119] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" HandleID="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5800), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"calico-kube-controllers-9b7544668-z2ztz", "timestamp":"2025-10-27 08:25:08.331015472 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.331 [INFO][4119] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.331 [INFO][4119] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.331 [INFO][4119] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.346 [INFO][4119] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.355 [INFO][4119] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.363 [INFO][4119] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.366 [INFO][4119] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.443288 containerd[1602]: 2025-10-27 08:25:08.371 [INFO][4119] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.371 [INFO][4119] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.374 [INFO][4119] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047 Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.381 [INFO][4119] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.394 [INFO][4119] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.2/26] block=192.168.14.0/26 handle="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.395 [INFO][4119] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.2/26] handle="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.396 [INFO][4119] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:08.444833 containerd[1602]: 2025-10-27 08:25:08.396 [INFO][4119] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.2/26] IPv6=[] ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" HandleID="k8s-pod-network.9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.445135 containerd[1602]: 2025-10-27 08:25:08.401 [INFO][4100] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0", GenerateName:"calico-kube-controllers-9b7544668-", Namespace:"calico-system", SelfLink:"", UID:"ffef7bc7-dd37-414d-945d-436bb2e2b656", ResourceVersion:"822", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"9b7544668", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"calico-kube-controllers-9b7544668-z2ztz", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.14.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali348d9b4faf5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:08.445264 containerd[1602]: 2025-10-27 08:25:08.402 [INFO][4100] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.2/32] ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.445264 containerd[1602]: 2025-10-27 08:25:08.402 [INFO][4100] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali348d9b4faf5 ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.445264 containerd[1602]: 2025-10-27 08:25:08.416 [INFO][4100] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.445400 containerd[1602]: 2025-10-27 08:25:08.417 [INFO][4100] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0", GenerateName:"calico-kube-controllers-9b7544668-", Namespace:"calico-system", SelfLink:"", UID:"ffef7bc7-dd37-414d-945d-436bb2e2b656", ResourceVersion:"822", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"9b7544668", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047", Pod:"calico-kube-controllers-9b7544668-z2ztz", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.14.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali348d9b4faf5", MAC:"3e:bd:6d:32:2e:49", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:08.445494 containerd[1602]: 2025-10-27 08:25:08.436 [INFO][4100] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" Namespace="calico-system" Pod="calico-kube-controllers-9b7544668-z2ztz" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--kube--controllers--9b7544668--z2ztz-eth0" Oct 27 08:25:08.472802 containerd[1602]: time="2025-10-27T08:25:08.472745876Z" level=info msg="connecting to shim 9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047" address="unix:///run/containerd/s/bc8b0141ba7bb056778b6b6bc3f8cd7edf83c2bcd9274378ef3887cf73b789c0" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:08.512888 systemd[1]: Started cri-containerd-9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047.scope - libcontainer container 9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047. Oct 27 08:25:08.518750 kubelet[2760]: E1027 08:25:08.518693 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-84f885dcfb-4c4ln" podUID="50c0829b-498d-4ffb-8c88-828cd743d3a0" Oct 27 08:25:08.613660 containerd[1602]: time="2025-10-27T08:25:08.613613112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-9b7544668-z2ztz,Uid:ffef7bc7-dd37-414d-945d-436bb2e2b656,Namespace:calico-system,Attempt:0,} returns sandbox id \"9ae873811c20493936b633d177b20c5703aff16745e52e1ac2437352355db047\"" Oct 27 08:25:08.617352 containerd[1602]: time="2025-10-27T08:25:08.617304603Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 27 08:25:08.980269 containerd[1602]: time="2025-10-27T08:25:08.980047189Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:08.981138 containerd[1602]: time="2025-10-27T08:25:08.980996320Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 27 08:25:08.981138 containerd[1602]: time="2025-10-27T08:25:08.981076460Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 27 08:25:08.981558 kubelet[2760]: E1027 08:25:08.981419 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 27 08:25:08.981738 kubelet[2760]: E1027 08:25:08.981710 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 27 08:25:08.982019 kubelet[2760]: E1027 08:25:08.981966 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6hcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-9b7544668-z2ztz_calico-system(ffef7bc7-dd37-414d-945d-436bb2e2b656): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:08.984309 kubelet[2760]: E1027 08:25:08.983501 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:25:09.191981 kubelet[2760]: E1027 08:25:09.191705 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:09.192789 containerd[1602]: time="2025-10-27T08:25:09.192719153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qdwvh,Uid:c45c241d-abf9-4ea1-a19c-427b39d582a2,Namespace:kube-system,Attempt:0,}" Oct 27 08:25:09.396458 systemd-networkd[1494]: cali305ea7d8e53: Link UP Oct 27 08:25:09.397869 systemd-networkd[1494]: cali305ea7d8e53: Gained carrier Oct 27 08:25:09.417652 containerd[1602]: 2025-10-27 08:25:09.232 [INFO][4180] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:09.417652 containerd[1602]: 2025-10-27 08:25:09.250 [INFO][4180] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0 coredns-668d6bf9bc- kube-system c45c241d-abf9-4ea1-a19c-427b39d582a2 816 0 2025-10-27 08:24:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 coredns-668d6bf9bc-qdwvh eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali305ea7d8e53 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-" Oct 27 08:25:09.417652 containerd[1602]: 2025-10-27 08:25:09.250 [INFO][4180] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.417652 containerd[1602]: 2025-10-27 08:25:09.313 [INFO][4197] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" HandleID="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.313 [INFO][4197] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" HandleID="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000ef760), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"coredns-668d6bf9bc-qdwvh", "timestamp":"2025-10-27 08:25:09.313000429 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.313 [INFO][4197] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.313 [INFO][4197] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.313 [INFO][4197] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.331 [INFO][4197] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.339 [INFO][4197] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.351 [INFO][4197] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.357 [INFO][4197] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.417995 containerd[1602]: 2025-10-27 08:25:09.364 [INFO][4197] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.364 [INFO][4197] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.368 [INFO][4197] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8 Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.375 [INFO][4197] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.387 [INFO][4197] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.3/26] block=192.168.14.0/26 handle="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.387 [INFO][4197] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.3/26] handle="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.387 [INFO][4197] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:09.418398 containerd[1602]: 2025-10-27 08:25:09.387 [INFO][4197] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.3/26] IPv6=[] ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" HandleID="k8s-pod-network.c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.418773 containerd[1602]: 2025-10-27 08:25:09.391 [INFO][4180] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"c45c241d-abf9-4ea1-a19c-427b39d582a2", ResourceVersion:"816", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"coredns-668d6bf9bc-qdwvh", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.14.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali305ea7d8e53", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:09.418773 containerd[1602]: 2025-10-27 08:25:09.391 [INFO][4180] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.3/32] ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.418773 containerd[1602]: 2025-10-27 08:25:09.391 [INFO][4180] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali305ea7d8e53 ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.418773 containerd[1602]: 2025-10-27 08:25:09.399 [INFO][4180] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.418773 containerd[1602]: 2025-10-27 08:25:09.400 [INFO][4180] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"c45c241d-abf9-4ea1-a19c-427b39d582a2", ResourceVersion:"816", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8", Pod:"coredns-668d6bf9bc-qdwvh", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.14.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali305ea7d8e53", MAC:"aa:60:db:d9:e2:6f", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:09.418773 containerd[1602]: 2025-10-27 08:25:09.411 [INFO][4180] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-qdwvh" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--qdwvh-eth0" Oct 27 08:25:09.470161 containerd[1602]: time="2025-10-27T08:25:09.470090669Z" level=info msg="connecting to shim c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8" address="unix:///run/containerd/s/edfca5255ddf956b8194a85f5fb87217696425277d34deeb8260441be54ff3e1" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:09.527104 systemd[1]: Started cri-containerd-c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8.scope - libcontainer container c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8. Oct 27 08:25:09.532644 kubelet[2760]: E1027 08:25:09.532211 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:25:09.665447 containerd[1602]: time="2025-10-27T08:25:09.665184924Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qdwvh,Uid:c45c241d-abf9-4ea1-a19c-427b39d582a2,Namespace:kube-system,Attempt:0,} returns sandbox id \"c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8\"" Oct 27 08:25:09.668278 kubelet[2760]: E1027 08:25:09.667858 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:09.672717 containerd[1602]: time="2025-10-27T08:25:09.672577324Z" level=info msg="CreateContainer within sandbox \"c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 27 08:25:09.690727 containerd[1602]: time="2025-10-27T08:25:09.688626073Z" level=info msg="Container aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:25:09.697101 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3281036127.mount: Deactivated successfully. Oct 27 08:25:09.700728 containerd[1602]: time="2025-10-27T08:25:09.700659021Z" level=info msg="CreateContainer within sandbox \"c7ea19c5447d12c1f944656446a63f73eb761fed9feaa90bc791b2560ef683a8\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e\"" Oct 27 08:25:09.701491 containerd[1602]: time="2025-10-27T08:25:09.701410559Z" level=info msg="StartContainer for \"aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e\"" Oct 27 08:25:09.705853 containerd[1602]: time="2025-10-27T08:25:09.705591137Z" level=info msg="connecting to shim aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e" address="unix:///run/containerd/s/edfca5255ddf956b8194a85f5fb87217696425277d34deeb8260441be54ff3e1" protocol=ttrpc version=3 Oct 27 08:25:09.737864 systemd[1]: Started cri-containerd-aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e.scope - libcontainer container aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e. Oct 27 08:25:09.783042 containerd[1602]: time="2025-10-27T08:25:09.782889803Z" level=info msg="StartContainer for \"aed7248e84f6fca72e4dbdd49a1558cc38a909839781e68f8f587bc1d305da2e\" returns successfully" Oct 27 08:25:10.495865 systemd-networkd[1494]: cali305ea7d8e53: Gained IPv6LL Oct 27 08:25:10.497295 systemd-networkd[1494]: cali348d9b4faf5: Gained IPv6LL Oct 27 08:25:10.533925 kubelet[2760]: E1027 08:25:10.533752 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:10.535398 kubelet[2760]: E1027 08:25:10.535333 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:25:10.577896 kubelet[2760]: I1027 08:25:10.577646 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-qdwvh" podStartSLOduration=40.577584686 podStartE2EDuration="40.577584686s" podCreationTimestamp="2025-10-27 08:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 08:25:10.576388323 +0000 UTC m=+44.527437942" watchObservedRunningTime="2025-10-27 08:25:10.577584686 +0000 UTC m=+44.528634353" Oct 27 08:25:11.191509 kubelet[2760]: E1027 08:25:11.191282 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:11.192593 containerd[1602]: time="2025-10-27T08:25:11.192112046Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-mrp2n,Uid:ce579d70-472c-4f85-ab9b-4d5886747e0d,Namespace:kube-system,Attempt:0,}" Oct 27 08:25:11.193799 containerd[1602]: time="2025-10-27T08:25:11.192773556Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-t5jfb,Uid:b6db6e46-66a7-435c-a200-9c21f50724bf,Namespace:calico-apiserver,Attempt:0,}" Oct 27 08:25:11.415625 systemd-networkd[1494]: cali50170c0c9f8: Link UP Oct 27 08:25:11.415923 systemd-networkd[1494]: cali50170c0c9f8: Gained carrier Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.249 [INFO][4332] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.269 [INFO][4332] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0 calico-apiserver-55869988d7- calico-apiserver b6db6e46-66a7-435c-a200-9c21f50724bf 826 0 2025-10-27 08:24:41 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:55869988d7 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 calico-apiserver-55869988d7-t5jfb eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali50170c0c9f8 [] [] }} ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.270 [INFO][4332] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.343 [INFO][4355] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" HandleID="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.344 [INFO][4355] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" HandleID="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fe00), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"calico-apiserver-55869988d7-t5jfb", "timestamp":"2025-10-27 08:25:11.343928997 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.344 [INFO][4355] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.344 [INFO][4355] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.344 [INFO][4355] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.353 [INFO][4355] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.360 [INFO][4355] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.367 [INFO][4355] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.369 [INFO][4355] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.372 [INFO][4355] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.372 [INFO][4355] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.375 [INFO][4355] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68 Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.381 [INFO][4355] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.402 [INFO][4355] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.4/26] block=192.168.14.0/26 handle="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.402 [INFO][4355] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.4/26] handle="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.403 [INFO][4355] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:11.454584 containerd[1602]: 2025-10-27 08:25:11.403 [INFO][4355] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.4/26] IPv6=[] ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" HandleID="k8s-pod-network.dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.457330 containerd[1602]: 2025-10-27 08:25:11.409 [INFO][4332] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0", GenerateName:"calico-apiserver-55869988d7-", Namespace:"calico-apiserver", SelfLink:"", UID:"b6db6e46-66a7-435c-a200-9c21f50724bf", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"55869988d7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"calico-apiserver-55869988d7-t5jfb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.14.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali50170c0c9f8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:11.457330 containerd[1602]: 2025-10-27 08:25:11.410 [INFO][4332] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.4/32] ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.457330 containerd[1602]: 2025-10-27 08:25:11.410 [INFO][4332] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali50170c0c9f8 ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.457330 containerd[1602]: 2025-10-27 08:25:11.415 [INFO][4332] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.457330 containerd[1602]: 2025-10-27 08:25:11.416 [INFO][4332] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0", GenerateName:"calico-apiserver-55869988d7-", Namespace:"calico-apiserver", SelfLink:"", UID:"b6db6e46-66a7-435c-a200-9c21f50724bf", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"55869988d7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68", Pod:"calico-apiserver-55869988d7-t5jfb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.14.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali50170c0c9f8", MAC:"46:e3:99:93:dd:df", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:11.457330 containerd[1602]: 2025-10-27 08:25:11.447 [INFO][4332] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-t5jfb" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--t5jfb-eth0" Oct 27 08:25:11.502409 containerd[1602]: time="2025-10-27T08:25:11.502348401Z" level=info msg="connecting to shim dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68" address="unix:///run/containerd/s/7fcb37676b9bebb1facac4a4eb4bbc728e425b56f51099ff978d09d538464b1d" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:11.551836 kubelet[2760]: E1027 08:25:11.551118 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:11.573939 systemd[1]: Started cri-containerd-dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68.scope - libcontainer container dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68. Oct 27 08:25:11.592981 systemd-networkd[1494]: caliedfd9abdf2b: Link UP Oct 27 08:25:11.594235 systemd-networkd[1494]: caliedfd9abdf2b: Gained carrier Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.260 [INFO][4330] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.282 [INFO][4330] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0 coredns-668d6bf9bc- kube-system ce579d70-472c-4f85-ab9b-4d5886747e0d 823 0 2025-10-27 08:24:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 coredns-668d6bf9bc-mrp2n eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] caliedfd9abdf2b [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.282 [INFO][4330] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.347 [INFO][4360] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" HandleID="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.347 [INFO][4360] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" HandleID="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5cd0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"coredns-668d6bf9bc-mrp2n", "timestamp":"2025-10-27 08:25:11.347220481 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.347 [INFO][4360] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.403 [INFO][4360] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.403 [INFO][4360] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.462 [INFO][4360] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.473 [INFO][4360] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.493 [INFO][4360] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.506 [INFO][4360] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.520 [INFO][4360] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.520 [INFO][4360] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.532 [INFO][4360] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4 Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.540 [INFO][4360] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.564 [INFO][4360] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.5/26] block=192.168.14.0/26 handle="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.564 [INFO][4360] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.5/26] handle="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.564 [INFO][4360] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:11.632569 containerd[1602]: 2025-10-27 08:25:11.564 [INFO][4360] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.5/26] IPv6=[] ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" HandleID="k8s-pod-network.2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.633285 containerd[1602]: 2025-10-27 08:25:11.576 [INFO][4330] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"ce579d70-472c-4f85-ab9b-4d5886747e0d", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"coredns-668d6bf9bc-mrp2n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.14.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliedfd9abdf2b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:11.633285 containerd[1602]: 2025-10-27 08:25:11.579 [INFO][4330] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.5/32] ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.633285 containerd[1602]: 2025-10-27 08:25:11.579 [INFO][4330] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliedfd9abdf2b ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.633285 containerd[1602]: 2025-10-27 08:25:11.596 [INFO][4330] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.633285 containerd[1602]: 2025-10-27 08:25:11.598 [INFO][4330] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"ce579d70-472c-4f85-ab9b-4d5886747e0d", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4", Pod:"coredns-668d6bf9bc-mrp2n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.14.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliedfd9abdf2b", MAC:"f2:e8:aa:0c:5b:91", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:11.633285 containerd[1602]: 2025-10-27 08:25:11.626 [INFO][4330] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" Namespace="kube-system" Pod="coredns-668d6bf9bc-mrp2n" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-coredns--668d6bf9bc--mrp2n-eth0" Oct 27 08:25:11.683190 containerd[1602]: time="2025-10-27T08:25:11.683143010Z" level=info msg="connecting to shim 2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4" address="unix:///run/containerd/s/6cfa4fac93f6c890c31a960746afa5edc64326e7dd55682305f0c84301e45e5c" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:11.737780 systemd[1]: Started cri-containerd-2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4.scope - libcontainer container 2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4. Oct 27 08:25:11.803802 containerd[1602]: time="2025-10-27T08:25:11.803106959Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-mrp2n,Uid:ce579d70-472c-4f85-ab9b-4d5886747e0d,Namespace:kube-system,Attempt:0,} returns sandbox id \"2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4\"" Oct 27 08:25:11.805473 kubelet[2760]: E1027 08:25:11.805445 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:11.812189 containerd[1602]: time="2025-10-27T08:25:11.810908535Z" level=info msg="CreateContainer within sandbox \"2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 27 08:25:11.877380 containerd[1602]: time="2025-10-27T08:25:11.877252486Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-t5jfb,Uid:b6db6e46-66a7-435c-a200-9c21f50724bf,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"dd342e2adeba478315ade42e887ad7dea1a9d297a48d893e8d47ad13fb85cb68\"" Oct 27 08:25:11.881186 containerd[1602]: time="2025-10-27T08:25:11.881149931Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 27 08:25:11.883034 containerd[1602]: time="2025-10-27T08:25:11.882983302Z" level=info msg="Container 43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4: CDI devices from CRI Config.CDIDevices: []" Oct 27 08:25:11.890746 containerd[1602]: time="2025-10-27T08:25:11.890692267Z" level=info msg="CreateContainer within sandbox \"2150b8538fdb0117cbcc2c0c729b811d6edd8d4979336c3ae5b9479a0b72b1b4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4\"" Oct 27 08:25:11.891736 containerd[1602]: time="2025-10-27T08:25:11.891639355Z" level=info msg="StartContainer for \"43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4\"" Oct 27 08:25:11.894083 containerd[1602]: time="2025-10-27T08:25:11.894050659Z" level=info msg="connecting to shim 43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4" address="unix:///run/containerd/s/6cfa4fac93f6c890c31a960746afa5edc64326e7dd55682305f0c84301e45e5c" protocol=ttrpc version=3 Oct 27 08:25:11.932721 systemd[1]: Started cri-containerd-43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4.scope - libcontainer container 43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4. Oct 27 08:25:11.992792 containerd[1602]: time="2025-10-27T08:25:11.992666916Z" level=info msg="StartContainer for \"43ac4970ae4f655a634971be114ff139f8cf83d76785263d3bbfa608218f58a4\" returns successfully" Oct 27 08:25:12.191550 containerd[1602]: time="2025-10-27T08:25:12.191387750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-qlx6r,Uid:51805d3a-45e5-4447-9079-e21004ab5665,Namespace:calico-system,Attempt:0,}" Oct 27 08:25:12.362690 systemd-networkd[1494]: calic5c7a264924: Link UP Oct 27 08:25:12.364246 systemd-networkd[1494]: calic5c7a264924: Gained carrier Oct 27 08:25:12.367428 containerd[1602]: time="2025-10-27T08:25:12.366993252Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:12.371552 containerd[1602]: time="2025-10-27T08:25:12.370651725Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:12.372119 containerd[1602]: time="2025-10-27T08:25:12.371668223Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 27 08:25:12.375634 kubelet[2760]: E1027 08:25:12.375580 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:12.375804 kubelet[2760]: E1027 08:25:12.375642 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:12.375866 kubelet[2760]: E1027 08:25:12.375811 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99mhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-55869988d7-t5jfb_calico-apiserver(b6db6e46-66a7-435c-a200-9c21f50724bf): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:12.382067 kubelet[2760]: E1027 08:25:12.380765 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.234 [INFO][4530] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.249 [INFO][4530] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0 goldmane-666569f655- calico-system 51805d3a-45e5-4447-9079-e21004ab5665 824 0 2025-10-27 08:24:43 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 goldmane-666569f655-qlx6r eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calic5c7a264924 [] [] }} ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.250 [INFO][4530] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.290 [INFO][4538] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" HandleID="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.291 [INFO][4538] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" HandleID="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f0e0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"goldmane-666569f655-qlx6r", "timestamp":"2025-10-27 08:25:12.290892791 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.291 [INFO][4538] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.291 [INFO][4538] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.291 [INFO][4538] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.301 [INFO][4538] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.310 [INFO][4538] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.317 [INFO][4538] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.320 [INFO][4538] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.325 [INFO][4538] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.326 [INFO][4538] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.328 [INFO][4538] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88 Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.336 [INFO][4538] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.350 [INFO][4538] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.6/26] block=192.168.14.0/26 handle="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.351 [INFO][4538] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.6/26] handle="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.351 [INFO][4538] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:12.403216 containerd[1602]: 2025-10-27 08:25:12.351 [INFO][4538] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.6/26] IPv6=[] ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" HandleID="k8s-pod-network.4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.404848 containerd[1602]: 2025-10-27 08:25:12.356 [INFO][4530] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"51805d3a-45e5-4447-9079-e21004ab5665", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 43, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"goldmane-666569f655-qlx6r", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.14.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calic5c7a264924", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:12.404848 containerd[1602]: 2025-10-27 08:25:12.357 [INFO][4530] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.6/32] ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.404848 containerd[1602]: 2025-10-27 08:25:12.357 [INFO][4530] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic5c7a264924 ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.404848 containerd[1602]: 2025-10-27 08:25:12.369 [INFO][4530] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.404848 containerd[1602]: 2025-10-27 08:25:12.372 [INFO][4530] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"51805d3a-45e5-4447-9079-e21004ab5665", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 43, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88", Pod:"goldmane-666569f655-qlx6r", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.14.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calic5c7a264924", MAC:"b6:09:5d:ee:7f:36", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:12.404848 containerd[1602]: 2025-10-27 08:25:12.397 [INFO][4530] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" Namespace="calico-system" Pod="goldmane-666569f655-qlx6r" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-goldmane--666569f655--qlx6r-eth0" Oct 27 08:25:12.447484 containerd[1602]: time="2025-10-27T08:25:12.447419330Z" level=info msg="connecting to shim 4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88" address="unix:///run/containerd/s/b126c676a64282fe8ef337a7d1aedb0dfb9a1ce22465aed1ce5c8ce6f79f18e1" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:12.499797 systemd[1]: Started cri-containerd-4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88.scope - libcontainer container 4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88. Oct 27 08:25:12.561418 kubelet[2760]: E1027 08:25:12.561373 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:12.567122 kubelet[2760]: E1027 08:25:12.567084 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:12.569552 kubelet[2760]: E1027 08:25:12.569489 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:25:12.589804 kubelet[2760]: I1027 08:25:12.589726 2760 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-mrp2n" podStartSLOduration=42.589704374 podStartE2EDuration="42.589704374s" podCreationTimestamp="2025-10-27 08:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 08:25:12.586988896 +0000 UTC m=+46.538038505" watchObservedRunningTime="2025-10-27 08:25:12.589704374 +0000 UTC m=+46.540753988" Oct 27 08:25:12.636428 containerd[1602]: time="2025-10-27T08:25:12.636338422Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-qlx6r,Uid:51805d3a-45e5-4447-9079-e21004ab5665,Namespace:calico-system,Attempt:0,} returns sandbox id \"4fc63d68fa68d7906ebdf18e5aa77ff3c5d5d462ba054758209086b2e76e3d88\"" Oct 27 08:25:12.639061 containerd[1602]: time="2025-10-27T08:25:12.638954472Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 27 08:25:13.000806 containerd[1602]: time="2025-10-27T08:25:13.000504055Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:13.002217 containerd[1602]: time="2025-10-27T08:25:13.002143404Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 27 08:25:13.002557 containerd[1602]: time="2025-10-27T08:25:13.002461611Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:13.002845 kubelet[2760]: E1027 08:25:13.002788 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 27 08:25:13.003225 kubelet[2760]: E1027 08:25:13.003147 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 27 08:25:13.003889 kubelet[2760]: E1027 08:25:13.003766 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x79st,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-qlx6r_calico-system(51805d3a-45e5-4447-9079-e21004ab5665): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:13.005035 kubelet[2760]: E1027 08:25:13.004993 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:25:13.191837 containerd[1602]: time="2025-10-27T08:25:13.191782492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fvjzl,Uid:b9e97248-79da-4981-820e-d3ae986590e1,Namespace:calico-system,Attempt:0,}" Oct 27 08:25:13.192215 containerd[1602]: time="2025-10-27T08:25:13.191785088Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-jkds2,Uid:c519ad97-124f-4ebf-ae30-a61647720f7b,Namespace:calico-apiserver,Attempt:0,}" Oct 27 08:25:13.249083 systemd-networkd[1494]: cali50170c0c9f8: Gained IPv6LL Oct 27 08:25:13.313315 systemd-networkd[1494]: caliedfd9abdf2b: Gained IPv6LL Oct 27 08:25:13.466044 systemd-networkd[1494]: cali241483e0a7f: Link UP Oct 27 08:25:13.466221 systemd-networkd[1494]: cali241483e0a7f: Gained carrier Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.275 [INFO][4629] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.304 [INFO][4629] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0 calico-apiserver-55869988d7- calico-apiserver c519ad97-124f-4ebf-ae30-a61647720f7b 821 0 2025-10-27 08:24:41 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:55869988d7 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 calico-apiserver-55869988d7-jkds2 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali241483e0a7f [] [] }} ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.305 [INFO][4629] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.368 [INFO][4648] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" HandleID="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.368 [INFO][4648] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" HandleID="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5230), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"calico-apiserver-55869988d7-jkds2", "timestamp":"2025-10-27 08:25:13.368169869 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.368 [INFO][4648] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.368 [INFO][4648] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.368 [INFO][4648] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.391 [INFO][4648] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.401 [INFO][4648] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.422 [INFO][4648] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.427 [INFO][4648] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.431 [INFO][4648] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.432 [INFO][4648] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.434 [INFO][4648] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550 Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.442 [INFO][4648] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.455 [INFO][4648] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.7/26] block=192.168.14.0/26 handle="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.456 [INFO][4648] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.7/26] handle="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.456 [INFO][4648] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:13.486047 containerd[1602]: 2025-10-27 08:25:13.456 [INFO][4648] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.7/26] IPv6=[] ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" HandleID="k8s-pod-network.9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.490304 containerd[1602]: 2025-10-27 08:25:13.461 [INFO][4629] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0", GenerateName:"calico-apiserver-55869988d7-", Namespace:"calico-apiserver", SelfLink:"", UID:"c519ad97-124f-4ebf-ae30-a61647720f7b", ResourceVersion:"821", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"55869988d7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"calico-apiserver-55869988d7-jkds2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.14.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali241483e0a7f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:13.490304 containerd[1602]: 2025-10-27 08:25:13.462 [INFO][4629] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.7/32] ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.490304 containerd[1602]: 2025-10-27 08:25:13.462 [INFO][4629] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali241483e0a7f ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.490304 containerd[1602]: 2025-10-27 08:25:13.465 [INFO][4629] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.490304 containerd[1602]: 2025-10-27 08:25:13.465 [INFO][4629] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0", GenerateName:"calico-apiserver-55869988d7-", Namespace:"calico-apiserver", SelfLink:"", UID:"c519ad97-124f-4ebf-ae30-a61647720f7b", ResourceVersion:"821", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"55869988d7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550", Pod:"calico-apiserver-55869988d7-jkds2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.14.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali241483e0a7f", MAC:"3a:49:a0:b9:f6:47", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:13.490304 containerd[1602]: 2025-10-27 08:25:13.480 [INFO][4629] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" Namespace="calico-apiserver" Pod="calico-apiserver-55869988d7-jkds2" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-calico--apiserver--55869988d7--jkds2-eth0" Oct 27 08:25:13.534937 containerd[1602]: time="2025-10-27T08:25:13.534731764Z" level=info msg="connecting to shim 9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550" address="unix:///run/containerd/s/96836eb156441b84ae4fd62dee5df8467b7243ff81e2eb458382aedabf8e04db" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:13.578820 kubelet[2760]: E1027 08:25:13.578590 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:13.582380 kubelet[2760]: E1027 08:25:13.581965 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:13.585220 kubelet[2760]: E1027 08:25:13.585183 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:25:13.585395 kubelet[2760]: E1027 08:25:13.585277 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:25:13.590113 systemd-networkd[1494]: calid178337ed70: Link UP Oct 27 08:25:13.590270 systemd-networkd[1494]: calid178337ed70: Gained carrier Oct 27 08:25:13.590867 systemd[1]: Started cri-containerd-9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550.scope - libcontainer container 9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550. Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.309 [INFO][4622] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.338 [INFO][4622] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0 csi-node-driver- calico-system b9e97248-79da-4981-820e-d3ae986590e1 708 0 2025-10-27 08:24:46 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-9999.9.9-k-4f7b9e69b1 csi-node-driver-fvjzl eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calid178337ed70 [] [] }} ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.338 [INFO][4622] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.401 [INFO][4653] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" HandleID="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.403 [INFO][4653] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" HandleID="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5800), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-9999.9.9-k-4f7b9e69b1", "pod":"csi-node-driver-fvjzl", "timestamp":"2025-10-27 08:25:13.401944682 +0000 UTC"}, Hostname:"ci-9999.9.9-k-4f7b9e69b1", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.403 [INFO][4653] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.456 [INFO][4653] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.457 [INFO][4653] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-9999.9.9-k-4f7b9e69b1' Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.493 [INFO][4653] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.502 [INFO][4653] ipam/ipam.go 394: Looking up existing affinities for host host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.518 [INFO][4653] ipam/ipam.go 511: Trying affinity for 192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.522 [INFO][4653] ipam/ipam.go 158: Attempting to load block cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.527 [INFO][4653] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.14.0/26 host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.528 [INFO][4653] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.14.0/26 handle="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.533 [INFO][4653] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38 Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.547 [INFO][4653] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.14.0/26 handle="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.574 [INFO][4653] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.14.8/26] block=192.168.14.0/26 handle="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.574 [INFO][4653] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.14.8/26] handle="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" host="ci-9999.9.9-k-4f7b9e69b1" Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.574 [INFO][4653] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Oct 27 08:25:13.621397 containerd[1602]: 2025-10-27 08:25:13.574 [INFO][4653] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.14.8/26] IPv6=[] ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" HandleID="k8s-pod-network.7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Workload="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.623419 containerd[1602]: 2025-10-27 08:25:13.581 [INFO][4622] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b9e97248-79da-4981-820e-d3ae986590e1", ResourceVersion:"708", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"", Pod:"csi-node-driver-fvjzl", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.14.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calid178337ed70", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:13.623419 containerd[1602]: 2025-10-27 08:25:13.582 [INFO][4622] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.14.8/32] ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.623419 containerd[1602]: 2025-10-27 08:25:13.582 [INFO][4622] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid178337ed70 ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.623419 containerd[1602]: 2025-10-27 08:25:13.589 [INFO][4622] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.623419 containerd[1602]: 2025-10-27 08:25:13.589 [INFO][4622] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b9e97248-79da-4981-820e-d3ae986590e1", ResourceVersion:"708", Generation:0, CreationTimestamp:time.Date(2025, time.October, 27, 8, 24, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-9999.9.9-k-4f7b9e69b1", ContainerID:"7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38", Pod:"csi-node-driver-fvjzl", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.14.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calid178337ed70", MAC:"0e:3e:67:1d:e2:4e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 27 08:25:13.623419 containerd[1602]: 2025-10-27 08:25:13.614 [INFO][4622] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" Namespace="calico-system" Pod="csi-node-driver-fvjzl" WorkloadEndpoint="ci--9999.9.9--k--4f7b9e69b1-k8s-csi--node--driver--fvjzl-eth0" Oct 27 08:25:13.673079 containerd[1602]: time="2025-10-27T08:25:13.673015422Z" level=info msg="connecting to shim 7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38" address="unix:///run/containerd/s/761f109a5352a1f27012c2c8ac2e3b2d06556d4ec77366f8932c8a18e52ba096" namespace=k8s.io protocol=ttrpc version=3 Oct 27 08:25:13.747796 systemd[1]: Started cri-containerd-7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38.scope - libcontainer container 7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38. Oct 27 08:25:13.806422 containerd[1602]: time="2025-10-27T08:25:13.806352898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-55869988d7-jkds2,Uid:c519ad97-124f-4ebf-ae30-a61647720f7b,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"9b55d2bf4dabb91f7794e96d3beed6b6120c6828b1d2ec4b6b6a60c6c76b3550\"" Oct 27 08:25:13.811822 containerd[1602]: time="2025-10-27T08:25:13.811784425Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 27 08:25:13.818009 containerd[1602]: time="2025-10-27T08:25:13.817854872Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-fvjzl,Uid:b9e97248-79da-4981-820e-d3ae986590e1,Namespace:calico-system,Attempt:0,} returns sandbox id \"7b8db831af30f186c340dde4a2eafca574197aecadaab6f1384e6950bc078d38\"" Oct 27 08:25:14.399680 systemd-networkd[1494]: calic5c7a264924: Gained IPv6LL Oct 27 08:25:14.520949 containerd[1602]: time="2025-10-27T08:25:14.520897080Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:14.522766 containerd[1602]: time="2025-10-27T08:25:14.521890275Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 27 08:25:14.522766 containerd[1602]: time="2025-10-27T08:25:14.521926591Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:14.523631 kubelet[2760]: E1027 08:25:14.522953 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:14.523631 kubelet[2760]: E1027 08:25:14.523008 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:14.523631 kubelet[2760]: E1027 08:25:14.523242 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4q2gc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-55869988d7-jkds2_calico-apiserver(c519ad97-124f-4ebf-ae30-a61647720f7b): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:14.524234 containerd[1602]: time="2025-10-27T08:25:14.524092681Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 27 08:25:14.525651 kubelet[2760]: E1027 08:25:14.524868 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:25:14.582744 kubelet[2760]: E1027 08:25:14.582426 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:25:14.586245 kubelet[2760]: E1027 08:25:14.585931 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:14.587261 kubelet[2760]: E1027 08:25:14.587175 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:25:14.773594 kubelet[2760]: I1027 08:25:14.773439 2760 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 27 08:25:14.774347 kubelet[2760]: E1027 08:25:14.774153 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:14.872951 containerd[1602]: time="2025-10-27T08:25:14.872886869Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:14.874006 containerd[1602]: time="2025-10-27T08:25:14.873870737Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 27 08:25:14.874122 containerd[1602]: time="2025-10-27T08:25:14.873949313Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 27 08:25:14.874409 kubelet[2760]: E1027 08:25:14.874305 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 27 08:25:14.874409 kubelet[2760]: E1027 08:25:14.874370 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 27 08:25:14.874873 kubelet[2760]: E1027 08:25:14.874808 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw4h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:14.878437 containerd[1602]: time="2025-10-27T08:25:14.878378143Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 27 08:25:15.224222 containerd[1602]: time="2025-10-27T08:25:15.224011164Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:15.225226 containerd[1602]: time="2025-10-27T08:25:15.225013505Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 27 08:25:15.225226 containerd[1602]: time="2025-10-27T08:25:15.225150647Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 27 08:25:15.225786 kubelet[2760]: E1027 08:25:15.225740 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 27 08:25:15.226135 kubelet[2760]: E1027 08:25:15.225913 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 27 08:25:15.226135 kubelet[2760]: E1027 08:25:15.226077 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw4h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:15.227637 kubelet[2760]: E1027 08:25:15.227543 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:25:15.359786 systemd-networkd[1494]: cali241483e0a7f: Gained IPv6LL Oct 27 08:25:15.552652 systemd-networkd[1494]: calid178337ed70: Gained IPv6LL Oct 27 08:25:15.590741 kubelet[2760]: E1027 08:25:15.590282 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:15.593675 kubelet[2760]: E1027 08:25:15.592807 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:25:15.594581 kubelet[2760]: E1027 08:25:15.594304 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:25:16.383961 systemd-networkd[1494]: vxlan.calico: Link UP Oct 27 08:25:16.383971 systemd-networkd[1494]: vxlan.calico: Gained carrier Oct 27 08:25:17.677308 systemd[1]: Started sshd@7-64.23.205.201:22-139.178.89.65:49824.service - OpenSSH per-connection server daemon (139.178.89.65:49824). Oct 27 08:25:17.808976 sshd[4934]: Accepted publickey for core from 139.178.89.65 port 49824 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:17.812770 sshd-session[4934]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:17.819999 systemd-logind[1571]: New session 8 of user core. Oct 27 08:25:17.825821 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 27 08:25:18.243438 systemd-networkd[1494]: vxlan.calico: Gained IPv6LL Oct 27 08:25:18.530257 sshd[4939]: Connection closed by 139.178.89.65 port 49824 Oct 27 08:25:18.531149 sshd-session[4934]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:18.540846 systemd[1]: sshd@7-64.23.205.201:22-139.178.89.65:49824.service: Deactivated successfully. Oct 27 08:25:18.547615 systemd[1]: session-8.scope: Deactivated successfully. Oct 27 08:25:18.549692 systemd-logind[1571]: Session 8 logged out. Waiting for processes to exit. Oct 27 08:25:18.553352 systemd-logind[1571]: Removed session 8. Oct 27 08:25:19.195302 containerd[1602]: time="2025-10-27T08:25:19.195024739Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 27 08:25:19.513698 containerd[1602]: time="2025-10-27T08:25:19.513324305Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:19.514306 containerd[1602]: time="2025-10-27T08:25:19.514191372Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 27 08:25:19.514474 containerd[1602]: time="2025-10-27T08:25:19.514440031Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 27 08:25:19.514777 kubelet[2760]: E1027 08:25:19.514733 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 27 08:25:19.515820 kubelet[2760]: E1027 08:25:19.515414 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 27 08:25:19.515820 kubelet[2760]: E1027 08:25:19.515625 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d7eec9ae1b0242a8b0b738edf8f38b0b,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fw5t4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-84f885dcfb-4c4ln_calico-system(50c0829b-498d-4ffb-8c88-828cd743d3a0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:19.523311 containerd[1602]: time="2025-10-27T08:25:19.523264654Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 27 08:25:19.876412 containerd[1602]: time="2025-10-27T08:25:19.876219759Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:19.877119 containerd[1602]: time="2025-10-27T08:25:19.876989472Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 27 08:25:19.877119 containerd[1602]: time="2025-10-27T08:25:19.876995354Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 27 08:25:19.877370 kubelet[2760]: E1027 08:25:19.877304 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 27 08:25:19.877446 kubelet[2760]: E1027 08:25:19.877381 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 27 08:25:19.879043 kubelet[2760]: E1027 08:25:19.877555 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fw5t4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-84f885dcfb-4c4ln_calico-system(50c0829b-498d-4ffb-8c88-828cd743d3a0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:19.879359 kubelet[2760]: E1027 08:25:19.879319 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-84f885dcfb-4c4ln" podUID="50c0829b-498d-4ffb-8c88-828cd743d3a0" Oct 27 08:25:22.215348 containerd[1602]: time="2025-10-27T08:25:22.215304153Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 27 08:25:22.732871 containerd[1602]: time="2025-10-27T08:25:22.732739298Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:22.733491 containerd[1602]: time="2025-10-27T08:25:22.733438723Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 27 08:25:22.733751 containerd[1602]: time="2025-10-27T08:25:22.733465838Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 27 08:25:22.733819 kubelet[2760]: E1027 08:25:22.733711 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 27 08:25:22.734404 kubelet[2760]: E1027 08:25:22.733843 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 27 08:25:22.734404 kubelet[2760]: E1027 08:25:22.734005 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6hcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-9b7544668-z2ztz_calico-system(ffef7bc7-dd37-414d-945d-436bb2e2b656): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:22.735213 kubelet[2760]: E1027 08:25:22.735131 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:25:23.548472 systemd[1]: Started sshd@8-64.23.205.201:22-139.178.89.65:49832.service - OpenSSH per-connection server daemon (139.178.89.65:49832). Oct 27 08:25:23.629905 sshd[4960]: Accepted publickey for core from 139.178.89.65 port 49832 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:23.631726 sshd-session[4960]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:23.638310 systemd-logind[1571]: New session 9 of user core. Oct 27 08:25:23.643989 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 27 08:25:23.813329 sshd[4963]: Connection closed by 139.178.89.65 port 49832 Oct 27 08:25:23.814237 sshd-session[4960]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:23.821337 systemd[1]: sshd@8-64.23.205.201:22-139.178.89.65:49832.service: Deactivated successfully. Oct 27 08:25:23.824698 systemd[1]: session-9.scope: Deactivated successfully. Oct 27 08:25:23.825915 systemd-logind[1571]: Session 9 logged out. Waiting for processes to exit. Oct 27 08:25:23.828279 systemd-logind[1571]: Removed session 9. Oct 27 08:25:26.237607 containerd[1602]: time="2025-10-27T08:25:26.237271525Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 27 08:25:26.608460 containerd[1602]: time="2025-10-27T08:25:26.608271812Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:26.609216 containerd[1602]: time="2025-10-27T08:25:26.609127512Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 27 08:25:26.609444 containerd[1602]: time="2025-10-27T08:25:26.609184591Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:26.609634 kubelet[2760]: E1027 08:25:26.609431 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:26.609634 kubelet[2760]: E1027 08:25:26.609552 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:26.610905 kubelet[2760]: E1027 08:25:26.609749 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99mhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-55869988d7-t5jfb_calico-apiserver(b6db6e46-66a7-435c-a200-9c21f50724bf): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:26.611245 kubelet[2760]: E1027 08:25:26.610963 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:25:28.192773 containerd[1602]: time="2025-10-27T08:25:28.192427165Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 27 08:25:28.532611 containerd[1602]: time="2025-10-27T08:25:28.532425286Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:28.533334 containerd[1602]: time="2025-10-27T08:25:28.533248862Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 27 08:25:28.533334 containerd[1602]: time="2025-10-27T08:25:28.533298733Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 27 08:25:28.534041 kubelet[2760]: E1027 08:25:28.533538 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 27 08:25:28.534041 kubelet[2760]: E1027 08:25:28.533602 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 27 08:25:28.534868 kubelet[2760]: E1027 08:25:28.534470 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw4h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:28.537334 containerd[1602]: time="2025-10-27T08:25:28.537289923Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 27 08:25:28.829512 systemd[1]: Started sshd@9-64.23.205.201:22-139.178.89.65:60762.service - OpenSSH per-connection server daemon (139.178.89.65:60762). Oct 27 08:25:28.872076 containerd[1602]: time="2025-10-27T08:25:28.872028546Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:28.872844 containerd[1602]: time="2025-10-27T08:25:28.872806576Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 27 08:25:28.873025 containerd[1602]: time="2025-10-27T08:25:28.872925702Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 27 08:25:28.874785 kubelet[2760]: E1027 08:25:28.874742 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 27 08:25:28.874898 kubelet[2760]: E1027 08:25:28.874799 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 27 08:25:28.874971 kubelet[2760]: E1027 08:25:28.874934 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw4h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:28.876346 kubelet[2760]: E1027 08:25:28.876151 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:25:28.902312 sshd[4986]: Accepted publickey for core from 139.178.89.65 port 60762 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:28.904030 sshd-session[4986]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:28.909791 systemd-logind[1571]: New session 10 of user core. Oct 27 08:25:28.915855 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 27 08:25:29.100695 sshd[4989]: Connection closed by 139.178.89.65 port 60762 Oct 27 08:25:29.103212 sshd-session[4986]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:29.115824 systemd[1]: sshd@9-64.23.205.201:22-139.178.89.65:60762.service: Deactivated successfully. Oct 27 08:25:29.121273 systemd[1]: session-10.scope: Deactivated successfully. Oct 27 08:25:29.123024 systemd-logind[1571]: Session 10 logged out. Waiting for processes to exit. Oct 27 08:25:29.127648 systemd[1]: Started sshd@10-64.23.205.201:22-139.178.89.65:60778.service - OpenSSH per-connection server daemon (139.178.89.65:60778). Oct 27 08:25:29.128771 systemd-logind[1571]: Removed session 10. Oct 27 08:25:29.194125 containerd[1602]: time="2025-10-27T08:25:29.192791997Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 27 08:25:29.254766 sshd[5001]: Accepted publickey for core from 139.178.89.65 port 60778 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:29.259017 sshd-session[5001]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:29.267630 systemd-logind[1571]: New session 11 of user core. Oct 27 08:25:29.273937 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 27 08:25:29.501570 sshd[5004]: Connection closed by 139.178.89.65 port 60778 Oct 27 08:25:29.501233 sshd-session[5001]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:29.516906 systemd[1]: sshd@10-64.23.205.201:22-139.178.89.65:60778.service: Deactivated successfully. Oct 27 08:25:29.521216 systemd[1]: session-11.scope: Deactivated successfully. Oct 27 08:25:29.528740 systemd-logind[1571]: Session 11 logged out. Waiting for processes to exit. Oct 27 08:25:29.530965 systemd[1]: Started sshd@11-64.23.205.201:22-139.178.89.65:60780.service - OpenSSH per-connection server daemon (139.178.89.65:60780). Oct 27 08:25:29.534201 systemd-logind[1571]: Removed session 11. Oct 27 08:25:29.538540 containerd[1602]: time="2025-10-27T08:25:29.537917839Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:29.540904 containerd[1602]: time="2025-10-27T08:25:29.540767746Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 27 08:25:29.541320 kubelet[2760]: E1027 08:25:29.541101 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:29.541320 kubelet[2760]: E1027 08:25:29.541162 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:29.543183 kubelet[2760]: E1027 08:25:29.541898 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4q2gc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-55869988d7-jkds2_calico-apiserver(c519ad97-124f-4ebf-ae30-a61647720f7b): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:29.543309 containerd[1602]: time="2025-10-27T08:25:29.540857794Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:29.544049 kubelet[2760]: E1027 08:25:29.543544 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:25:29.610666 sshd[5014]: Accepted publickey for core from 139.178.89.65 port 60780 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:29.612979 sshd-session[5014]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:29.621949 systemd-logind[1571]: New session 12 of user core. Oct 27 08:25:29.628965 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 27 08:25:29.810304 sshd[5017]: Connection closed by 139.178.89.65 port 60780 Oct 27 08:25:29.812940 sshd-session[5014]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:29.820798 systemd-logind[1571]: Session 12 logged out. Waiting for processes to exit. Oct 27 08:25:29.821138 systemd[1]: sshd@11-64.23.205.201:22-139.178.89.65:60780.service: Deactivated successfully. Oct 27 08:25:29.825238 systemd[1]: session-12.scope: Deactivated successfully. Oct 27 08:25:29.828502 systemd-logind[1571]: Removed session 12. Oct 27 08:25:30.196706 containerd[1602]: time="2025-10-27T08:25:30.196578297Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 27 08:25:30.626802 containerd[1602]: time="2025-10-27T08:25:30.626746804Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:30.627784 containerd[1602]: time="2025-10-27T08:25:30.627692424Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 27 08:25:30.627978 containerd[1602]: time="2025-10-27T08:25:30.627953416Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:30.628239 kubelet[2760]: E1027 08:25:30.628196 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 27 08:25:30.628826 kubelet[2760]: E1027 08:25:30.628255 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 27 08:25:30.628826 kubelet[2760]: E1027 08:25:30.628404 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x79st,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-qlx6r_calico-system(51805d3a-45e5-4447-9079-e21004ab5665): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:30.629753 kubelet[2760]: E1027 08:25:30.629705 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:25:31.196262 kubelet[2760]: E1027 08:25:31.196202 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-84f885dcfb-4c4ln" podUID="50c0829b-498d-4ffb-8c88-828cd743d3a0" Oct 27 08:25:34.831985 systemd[1]: Started sshd@12-64.23.205.201:22-139.178.89.65:60790.service - OpenSSH per-connection server daemon (139.178.89.65:60790). Oct 27 08:25:34.902644 sshd[5035]: Accepted publickey for core from 139.178.89.65 port 60790 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:34.904695 sshd-session[5035]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:34.911678 systemd-logind[1571]: New session 13 of user core. Oct 27 08:25:34.915775 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 27 08:25:35.070411 sshd[5038]: Connection closed by 139.178.89.65 port 60790 Oct 27 08:25:35.072990 sshd-session[5035]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:35.080945 systemd[1]: sshd@12-64.23.205.201:22-139.178.89.65:60790.service: Deactivated successfully. Oct 27 08:25:35.083212 systemd[1]: session-13.scope: Deactivated successfully. Oct 27 08:25:35.085743 systemd-logind[1571]: Session 13 logged out. Waiting for processes to exit. Oct 27 08:25:35.087977 systemd-logind[1571]: Removed session 13. Oct 27 08:25:35.193224 kubelet[2760]: E1027 08:25:35.192844 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:25:37.192586 kubelet[2760]: E1027 08:25:37.192170 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:25:37.658403 containerd[1602]: time="2025-10-27T08:25:37.658342758Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\" id:\"75464b967eab928a13c94e76172917c1d862798501441d5303bbea647ae53d58\" pid:5074 exit_status:1 exited_at:{seconds:1761553537 nanos:657973453}" Oct 27 08:25:40.086479 systemd[1]: Started sshd@13-64.23.205.201:22-139.178.89.65:40596.service - OpenSSH per-connection server daemon (139.178.89.65:40596). Oct 27 08:25:40.170189 sshd[5088]: Accepted publickey for core from 139.178.89.65 port 40596 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:40.171857 sshd-session[5088]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:40.177955 systemd-logind[1571]: New session 14 of user core. Oct 27 08:25:40.193989 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 27 08:25:40.386207 sshd[5091]: Connection closed by 139.178.89.65 port 40596 Oct 27 08:25:40.386984 sshd-session[5088]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:40.394437 systemd[1]: sshd@13-64.23.205.201:22-139.178.89.65:40596.service: Deactivated successfully. Oct 27 08:25:40.399625 systemd[1]: session-14.scope: Deactivated successfully. Oct 27 08:25:40.401761 systemd-logind[1571]: Session 14 logged out. Waiting for processes to exit. Oct 27 08:25:40.404453 systemd-logind[1571]: Removed session 14. Oct 27 08:25:42.196629 kubelet[2760]: E1027 08:25:42.195623 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:25:42.196629 kubelet[2760]: E1027 08:25:42.196105 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:25:43.194567 kubelet[2760]: E1027 08:25:43.194493 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:25:45.405908 systemd[1]: Started sshd@14-64.23.205.201:22-139.178.89.65:40610.service - OpenSSH per-connection server daemon (139.178.89.65:40610). Oct 27 08:25:45.491012 sshd[5108]: Accepted publickey for core from 139.178.89.65 port 40610 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:45.494545 sshd-session[5108]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:45.504316 systemd-logind[1571]: New session 15 of user core. Oct 27 08:25:45.509812 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 27 08:25:45.699578 sshd[5111]: Connection closed by 139.178.89.65 port 40610 Oct 27 08:25:45.700432 sshd-session[5108]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:45.706447 systemd[1]: sshd@14-64.23.205.201:22-139.178.89.65:40610.service: Deactivated successfully. Oct 27 08:25:45.710443 systemd[1]: session-15.scope: Deactivated successfully. Oct 27 08:25:45.712406 systemd-logind[1571]: Session 15 logged out. Waiting for processes to exit. Oct 27 08:25:45.715677 systemd-logind[1571]: Removed session 15. Oct 27 08:25:46.194350 containerd[1602]: time="2025-10-27T08:25:46.194093372Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Oct 27 08:25:46.514884 containerd[1602]: time="2025-10-27T08:25:46.514584816Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:46.515725 containerd[1602]: time="2025-10-27T08:25:46.515604629Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Oct 27 08:25:46.515725 containerd[1602]: time="2025-10-27T08:25:46.515618726Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=73" Oct 27 08:25:46.516007 kubelet[2760]: E1027 08:25:46.515952 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 27 08:25:46.516412 kubelet[2760]: E1027 08:25:46.516014 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Oct 27 08:25:46.516412 kubelet[2760]: E1027 08:25:46.516150 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d7eec9ae1b0242a8b0b738edf8f38b0b,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fw5t4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-84f885dcfb-4c4ln_calico-system(50c0829b-498d-4ffb-8c88-828cd743d3a0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.4\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:46.518563 containerd[1602]: time="2025-10-27T08:25:46.518487301Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Oct 27 08:25:46.892181 containerd[1602]: time="2025-10-27T08:25:46.892131410Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:46.893051 containerd[1602]: time="2025-10-27T08:25:46.892992403Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Oct 27 08:25:46.893470 containerd[1602]: time="2025-10-27T08:25:46.893116277Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=85" Oct 27 08:25:46.893550 kubelet[2760]: E1027 08:25:46.893435 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 27 08:25:46.893550 kubelet[2760]: E1027 08:25:46.893498 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Oct 27 08:25:46.896173 kubelet[2760]: E1027 08:25:46.896107 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fw5t4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-84f885dcfb-4c4ln_calico-system(50c0829b-498d-4ffb-8c88-828cd743d3a0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:46.897614 kubelet[2760]: E1027 08:25:46.897559 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-84f885dcfb-4c4ln" podUID="50c0829b-498d-4ffb-8c88-828cd743d3a0" Oct 27 08:25:48.193229 containerd[1602]: time="2025-10-27T08:25:48.192502650Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 27 08:25:48.546022 containerd[1602]: time="2025-10-27T08:25:48.545427701Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:48.546629 containerd[1602]: time="2025-10-27T08:25:48.546502855Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 27 08:25:48.546629 containerd[1602]: time="2025-10-27T08:25:48.546570975Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:48.546921 kubelet[2760]: E1027 08:25:48.546884 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:48.548583 kubelet[2760]: E1027 08:25:48.547428 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:48.548836 kubelet[2760]: E1027 08:25:48.548790 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99mhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-55869988d7-t5jfb_calico-apiserver(b6db6e46-66a7-435c-a200-9c21f50724bf): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:48.550177 kubelet[2760]: E1027 08:25:48.550131 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:25:49.194409 containerd[1602]: time="2025-10-27T08:25:49.193885324Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Oct 27 08:25:49.543600 containerd[1602]: time="2025-10-27T08:25:49.543421949Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:49.544533 containerd[1602]: time="2025-10-27T08:25:49.544422000Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Oct 27 08:25:49.544533 containerd[1602]: time="2025-10-27T08:25:49.544489876Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=85" Oct 27 08:25:49.544843 kubelet[2760]: E1027 08:25:49.544786 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 27 08:25:49.544919 kubelet[2760]: E1027 08:25:49.544843 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Oct 27 08:25:49.545170 kubelet[2760]: E1027 08:25:49.545000 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6hcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-9b7544668-z2ztz_calico-system(ffef7bc7-dd37-414d-945d-436bb2e2b656): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:49.546682 kubelet[2760]: E1027 08:25:49.546640 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:25:50.718034 systemd[1]: Started sshd@15-64.23.205.201:22-139.178.89.65:44934.service - OpenSSH per-connection server daemon (139.178.89.65:44934). Oct 27 08:25:50.831275 sshd[5124]: Accepted publickey for core from 139.178.89.65 port 44934 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:50.836330 sshd-session[5124]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:50.844048 systemd-logind[1571]: New session 16 of user core. Oct 27 08:25:50.849861 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 27 08:25:51.103983 sshd[5127]: Connection closed by 139.178.89.65 port 44934 Oct 27 08:25:51.104905 sshd-session[5124]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:51.120005 systemd[1]: sshd@15-64.23.205.201:22-139.178.89.65:44934.service: Deactivated successfully. Oct 27 08:25:51.124168 systemd[1]: session-16.scope: Deactivated successfully. Oct 27 08:25:51.125509 systemd-logind[1571]: Session 16 logged out. Waiting for processes to exit. Oct 27 08:25:51.129053 systemd[1]: Started sshd@16-64.23.205.201:22-139.178.89.65:44944.service - OpenSSH per-connection server daemon (139.178.89.65:44944). Oct 27 08:25:51.131218 systemd-logind[1571]: Removed session 16. Oct 27 08:25:51.199663 kubelet[2760]: E1027 08:25:51.198005 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:51.254111 sshd[5139]: Accepted publickey for core from 139.178.89.65 port 44944 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:51.257925 sshd-session[5139]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:51.267936 systemd-logind[1571]: New session 17 of user core. Oct 27 08:25:51.276953 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 27 08:25:51.602222 sshd[5142]: Connection closed by 139.178.89.65 port 44944 Oct 27 08:25:51.603947 sshd-session[5139]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:51.616304 systemd[1]: sshd@16-64.23.205.201:22-139.178.89.65:44944.service: Deactivated successfully. Oct 27 08:25:51.619409 systemd[1]: session-17.scope: Deactivated successfully. Oct 27 08:25:51.622947 systemd-logind[1571]: Session 17 logged out. Waiting for processes to exit. Oct 27 08:25:51.625494 systemd[1]: Started sshd@17-64.23.205.201:22-139.178.89.65:44948.service - OpenSSH per-connection server daemon (139.178.89.65:44948). Oct 27 08:25:51.627469 systemd-logind[1571]: Removed session 17. Oct 27 08:25:51.714374 sshd[5153]: Accepted publickey for core from 139.178.89.65 port 44948 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:51.717046 sshd-session[5153]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:51.722775 systemd-logind[1571]: New session 18 of user core. Oct 27 08:25:51.727953 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 27 08:25:52.492940 sshd[5156]: Connection closed by 139.178.89.65 port 44948 Oct 27 08:25:52.496586 sshd-session[5153]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:52.507919 systemd[1]: sshd@17-64.23.205.201:22-139.178.89.65:44948.service: Deactivated successfully. Oct 27 08:25:52.514243 systemd[1]: session-18.scope: Deactivated successfully. Oct 27 08:25:52.516244 systemd-logind[1571]: Session 18 logged out. Waiting for processes to exit. Oct 27 08:25:52.520088 systemd-logind[1571]: Removed session 18. Oct 27 08:25:52.523147 systemd[1]: Started sshd@18-64.23.205.201:22-139.178.89.65:44958.service - OpenSSH per-connection server daemon (139.178.89.65:44958). Oct 27 08:25:52.625562 sshd[5171]: Accepted publickey for core from 139.178.89.65 port 44958 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:52.627219 sshd-session[5171]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:52.633749 systemd-logind[1571]: New session 19 of user core. Oct 27 08:25:52.637780 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 27 08:25:53.050353 sshd[5176]: Connection closed by 139.178.89.65 port 44958 Oct 27 08:25:53.056956 sshd-session[5171]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:53.074276 systemd[1]: sshd@18-64.23.205.201:22-139.178.89.65:44958.service: Deactivated successfully. Oct 27 08:25:53.080683 systemd[1]: session-19.scope: Deactivated successfully. Oct 27 08:25:53.082786 systemd-logind[1571]: Session 19 logged out. Waiting for processes to exit. Oct 27 08:25:53.089311 systemd[1]: Started sshd@19-64.23.205.201:22-139.178.89.65:44964.service - OpenSSH per-connection server daemon (139.178.89.65:44964). Oct 27 08:25:53.090399 systemd-logind[1571]: Removed session 19. Oct 27 08:25:53.192645 containerd[1602]: time="2025-10-27T08:25:53.192591358Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Oct 27 08:25:53.203411 sshd[5186]: Accepted publickey for core from 139.178.89.65 port 44964 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:53.206329 sshd-session[5186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:53.217907 systemd-logind[1571]: New session 20 of user core. Oct 27 08:25:53.225891 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 27 08:25:53.381779 sshd[5189]: Connection closed by 139.178.89.65 port 44964 Oct 27 08:25:53.382678 sshd-session[5186]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:53.389686 systemd[1]: sshd@19-64.23.205.201:22-139.178.89.65:44964.service: Deactivated successfully. Oct 27 08:25:53.397084 systemd[1]: session-20.scope: Deactivated successfully. Oct 27 08:25:53.399725 systemd-logind[1571]: Session 20 logged out. Waiting for processes to exit. Oct 27 08:25:53.403199 systemd-logind[1571]: Removed session 20. Oct 27 08:25:53.656981 containerd[1602]: time="2025-10-27T08:25:53.656820416Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:53.658091 containerd[1602]: time="2025-10-27T08:25:53.658026107Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Oct 27 08:25:53.658471 containerd[1602]: time="2025-10-27T08:25:53.658096825Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:53.658601 kubelet[2760]: E1027 08:25:53.658565 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:53.659195 kubelet[2760]: E1027 08:25:53.658637 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Oct 27 08:25:53.659195 kubelet[2760]: E1027 08:25:53.658957 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4q2gc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-55869988d7-jkds2_calico-apiserver(c519ad97-124f-4ebf-ae30-a61647720f7b): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:53.660212 kubelet[2760]: E1027 08:25:53.660150 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:25:54.191151 kubelet[2760]: E1027 08:25:54.191029 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:25:56.195340 containerd[1602]: time="2025-10-27T08:25:56.194960709Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Oct 27 08:25:56.575681 containerd[1602]: time="2025-10-27T08:25:56.575438209Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:56.576570 containerd[1602]: time="2025-10-27T08:25:56.576505563Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" Oct 27 08:25:56.576710 containerd[1602]: time="2025-10-27T08:25:56.576537290Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=69" Oct 27 08:25:56.577565 kubelet[2760]: E1027 08:25:56.576902 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 27 08:25:56.577565 kubelet[2760]: E1027 08:25:56.576968 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Oct 27 08:25:56.577565 kubelet[2760]: E1027 08:25:56.577109 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw4h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/csi:v3.30.4\": ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:56.580636 containerd[1602]: time="2025-10-27T08:25:56.580271811Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Oct 27 08:25:56.907544 containerd[1602]: time="2025-10-27T08:25:56.907253781Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:56.908675 containerd[1602]: time="2025-10-27T08:25:56.908543460Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Oct 27 08:25:56.908675 containerd[1602]: time="2025-10-27T08:25:56.908556506Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=93" Oct 27 08:25:56.910366 kubelet[2760]: E1027 08:25:56.909006 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 27 08:25:56.910366 kubelet[2760]: E1027 08:25:56.909062 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Oct 27 08:25:56.910366 kubelet[2760]: E1027 08:25:56.909199 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sw4h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-fvjzl_calico-system(b9e97248-79da-4981-820e-d3ae986590e1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:56.910777 kubelet[2760]: E1027 08:25:56.910737 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-fvjzl" podUID="b9e97248-79da-4981-820e-d3ae986590e1" Oct 27 08:25:57.192453 containerd[1602]: time="2025-10-27T08:25:57.192085350Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Oct 27 08:25:57.567355 containerd[1602]: time="2025-10-27T08:25:57.567156043Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Oct 27 08:25:57.568463 containerd[1602]: time="2025-10-27T08:25:57.568402753Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Oct 27 08:25:57.568647 containerd[1602]: time="2025-10-27T08:25:57.568606417Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=77" Oct 27 08:25:57.568932 kubelet[2760]: E1027 08:25:57.568885 2760 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 27 08:25:57.569057 kubelet[2760]: E1027 08:25:57.569040 2760 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Oct 27 08:25:57.571573 kubelet[2760]: E1027 08:25:57.569270 2760 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x79st,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-qlx6r_calico-system(51805d3a-45e5-4447-9079-e21004ab5665): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Oct 27 08:25:57.573113 kubelet[2760]: E1027 08:25:57.573043 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:25:58.217373 kubelet[2760]: E1027 08:25:58.217196 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-84f885dcfb-4c4ln" podUID="50c0829b-498d-4ffb-8c88-828cd743d3a0" Oct 27 08:25:58.402992 systemd[1]: Started sshd@20-64.23.205.201:22-139.178.89.65:55028.service - OpenSSH per-connection server daemon (139.178.89.65:55028). Oct 27 08:25:58.486277 sshd[5211]: Accepted publickey for core from 139.178.89.65 port 55028 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:25:58.488309 sshd-session[5211]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:25:58.495637 systemd-logind[1571]: New session 21 of user core. Oct 27 08:25:58.502907 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 27 08:25:58.640967 sshd[5214]: Connection closed by 139.178.89.65 port 55028 Oct 27 08:25:58.641718 sshd-session[5211]: pam_unix(sshd:session): session closed for user core Oct 27 08:25:58.648303 systemd[1]: sshd@20-64.23.205.201:22-139.178.89.65:55028.service: Deactivated successfully. Oct 27 08:25:58.651666 systemd[1]: session-21.scope: Deactivated successfully. Oct 27 08:25:58.653838 systemd-logind[1571]: Session 21 logged out. Waiting for processes to exit. Oct 27 08:25:58.656113 systemd-logind[1571]: Removed session 21. Oct 27 08:26:00.192290 kubelet[2760]: E1027 08:26:00.192160 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-9b7544668-z2ztz" podUID="ffef7bc7-dd37-414d-945d-436bb2e2b656" Oct 27 08:26:02.192058 kubelet[2760]: E1027 08:26:02.191034 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:26:02.192058 kubelet[2760]: E1027 08:26:02.191900 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:26:03.192899 kubelet[2760]: E1027 08:26:03.192823 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-t5jfb" podUID="b6db6e46-66a7-435c-a200-9c21f50724bf" Oct 27 08:26:03.660629 systemd[1]: Started sshd@21-64.23.205.201:22-139.178.89.65:55032.service - OpenSSH per-connection server daemon (139.178.89.65:55032). Oct 27 08:26:03.786613 sshd[5229]: Accepted publickey for core from 139.178.89.65 port 55032 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:26:03.789079 sshd-session[5229]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:26:03.795614 systemd-logind[1571]: New session 22 of user core. Oct 27 08:26:03.801843 systemd[1]: Started session-22.scope - Session 22 of User core. Oct 27 08:26:04.073474 sshd[5232]: Connection closed by 139.178.89.65 port 55032 Oct 27 08:26:04.074793 sshd-session[5229]: pam_unix(sshd:session): session closed for user core Oct 27 08:26:04.079093 systemd[1]: sshd@21-64.23.205.201:22-139.178.89.65:55032.service: Deactivated successfully. Oct 27 08:26:04.081947 systemd[1]: session-22.scope: Deactivated successfully. Oct 27 08:26:04.083404 systemd-logind[1571]: Session 22 logged out. Waiting for processes to exit. Oct 27 08:26:04.086129 systemd-logind[1571]: Removed session 22. Oct 27 08:26:06.194237 kubelet[2760]: E1027 08:26:06.194100 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-55869988d7-jkds2" podUID="c519ad97-124f-4ebf-ae30-a61647720f7b" Oct 27 08:26:07.636638 containerd[1602]: time="2025-10-27T08:26:07.636502324Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1c5554e76a0305ec5029c50652e156c595a2ca46789f158a695c9c350f06ce8c\" id:\"4bf95ffc4a3ff12e33971efebbbf19c7ea4e301a2742fdbabf745c831df648e6\" pid:5255 exited_at:{seconds:1761553567 nanos:635789869}" Oct 27 08:26:07.639976 kubelet[2760]: E1027 08:26:07.639933 2760 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Oct 27 08:26:09.096681 systemd[1]: Started sshd@22-64.23.205.201:22-139.178.89.65:50018.service - OpenSSH per-connection server daemon (139.178.89.65:50018). Oct 27 08:26:09.198823 kubelet[2760]: E1027 08:26:09.198776 2760 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-qlx6r" podUID="51805d3a-45e5-4447-9079-e21004ab5665" Oct 27 08:26:09.206659 sshd[5266]: Accepted publickey for core from 139.178.89.65 port 50018 ssh2: RSA SHA256:rxa87oi8ZZqMD8URaMdjWEem69/UDQnMWUTPMulZcos Oct 27 08:26:09.213132 sshd-session[5266]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 27 08:26:09.224154 systemd-logind[1571]: New session 23 of user core. Oct 27 08:26:09.231838 systemd[1]: Started session-23.scope - Session 23 of User core. Oct 27 08:26:09.487208 sshd[5269]: Connection closed by 139.178.89.65 port 50018 Oct 27 08:26:09.489056 sshd-session[5266]: pam_unix(sshd:session): session closed for user core Oct 27 08:26:09.500281 systemd[1]: sshd@22-64.23.205.201:22-139.178.89.65:50018.service: Deactivated successfully. Oct 27 08:26:09.504376 systemd[1]: session-23.scope: Deactivated successfully. Oct 27 08:26:09.507169 systemd-logind[1571]: Session 23 logged out. Waiting for processes to exit. Oct 27 08:26:09.508459 systemd-logind[1571]: Removed session 23.