Sep 12 18:07:28.877752 kernel: Linux version 6.12.47-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Fri Sep 12 15:34:39 -00 2025 Sep 12 18:07:28.877785 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=271a44cc8ea1639cfb6fdf777202a5f025fda0b3ce9b293cc4e0e7047aecb858 Sep 12 18:07:28.877795 kernel: BIOS-provided physical RAM map: Sep 12 18:07:28.877802 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 12 18:07:28.877808 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 12 18:07:28.877815 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 12 18:07:28.877823 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Sep 12 18:07:28.877838 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Sep 12 18:07:28.877848 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 12 18:07:28.877854 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 12 18:07:28.877861 kernel: NX (Execute Disable) protection: active Sep 12 18:07:28.877868 kernel: APIC: Static calls initialized Sep 12 18:07:28.877874 kernel: SMBIOS 2.8 present. Sep 12 18:07:28.877881 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Sep 12 18:07:28.877892 kernel: DMI: Memory slots populated: 1/1 Sep 12 18:07:28.877900 kernel: Hypervisor detected: KVM Sep 12 18:07:28.877910 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 12 18:07:28.877918 kernel: kvm-clock: using sched offset of 4973076456 cycles Sep 12 18:07:28.877926 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 12 18:07:28.877934 kernel: tsc: Detected 2494.140 MHz processor Sep 12 18:07:28.877942 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 12 18:07:28.877950 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 12 18:07:28.877958 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Sep 12 18:07:28.877969 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Sep 12 18:07:28.877977 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 12 18:07:28.877984 kernel: ACPI: Early table checksum verification disabled Sep 12 18:07:28.877992 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Sep 12 18:07:28.877999 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878007 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878015 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878022 kernel: ACPI: FACS 0x000000007FFE0000 000040 Sep 12 18:07:28.878030 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878041 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878048 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878056 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 18:07:28.878064 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Sep 12 18:07:28.878071 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Sep 12 18:07:28.878079 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Sep 12 18:07:28.878086 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Sep 12 18:07:28.878094 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Sep 12 18:07:28.878108 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Sep 12 18:07:28.878116 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Sep 12 18:07:28.878124 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Sep 12 18:07:28.878132 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Sep 12 18:07:28.878140 kernel: NUMA: Node 0 [mem 0x00001000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00001000-0x7ffdafff] Sep 12 18:07:28.878151 kernel: NODE_DATA(0) allocated [mem 0x7ffd3dc0-0x7ffdafff] Sep 12 18:07:28.878159 kernel: Zone ranges: Sep 12 18:07:28.878167 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 12 18:07:28.878175 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Sep 12 18:07:28.878183 kernel: Normal empty Sep 12 18:07:28.878191 kernel: Device empty Sep 12 18:07:28.878221 kernel: Movable zone start for each node Sep 12 18:07:28.878230 kernel: Early memory node ranges Sep 12 18:07:28.878238 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 12 18:07:28.878246 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Sep 12 18:07:28.878257 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Sep 12 18:07:28.878265 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 12 18:07:28.878273 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 12 18:07:28.878281 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Sep 12 18:07:28.878289 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 12 18:07:28.878297 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 12 18:07:28.878308 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 12 18:07:28.878317 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 12 18:07:28.878327 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 12 18:07:28.878338 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 12 18:07:28.878346 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 12 18:07:28.878356 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 12 18:07:28.878364 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 12 18:07:28.878373 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 12 18:07:28.878381 kernel: TSC deadline timer available Sep 12 18:07:28.878389 kernel: CPU topo: Max. logical packages: 1 Sep 12 18:07:28.878397 kernel: CPU topo: Max. logical dies: 1 Sep 12 18:07:28.878405 kernel: CPU topo: Max. dies per package: 1 Sep 12 18:07:28.878415 kernel: CPU topo: Max. threads per core: 1 Sep 12 18:07:28.878424 kernel: CPU topo: Num. cores per package: 2 Sep 12 18:07:28.878432 kernel: CPU topo: Num. threads per package: 2 Sep 12 18:07:28.878439 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Sep 12 18:07:28.878447 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 12 18:07:28.878455 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Sep 12 18:07:28.878463 kernel: Booting paravirtualized kernel on KVM Sep 12 18:07:28.878472 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 12 18:07:28.880159 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Sep 12 18:07:28.880179 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Sep 12 18:07:28.880201 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Sep 12 18:07:28.880214 kernel: pcpu-alloc: [0] 0 1 Sep 12 18:07:28.880250 kernel: kvm-guest: PV spinlocks disabled, no host support Sep 12 18:07:28.880282 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=271a44cc8ea1639cfb6fdf777202a5f025fda0b3ce9b293cc4e0e7047aecb858 Sep 12 18:07:28.880297 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 12 18:07:28.880311 kernel: random: crng init done Sep 12 18:07:28.880324 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 12 18:07:28.880338 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 12 18:07:28.880355 kernel: Fallback order for Node 0: 0 Sep 12 18:07:28.880368 kernel: Built 1 zonelists, mobility grouping on. Total pages: 524153 Sep 12 18:07:28.880381 kernel: Policy zone: DMA32 Sep 12 18:07:28.880395 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 12 18:07:28.880413 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 12 18:07:28.880427 kernel: Kernel/User page tables isolation: enabled Sep 12 18:07:28.880441 kernel: ftrace: allocating 40125 entries in 157 pages Sep 12 18:07:28.880455 kernel: ftrace: allocated 157 pages with 5 groups Sep 12 18:07:28.880468 kernel: Dynamic Preempt: voluntary Sep 12 18:07:28.880505 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 12 18:07:28.880525 kernel: rcu: RCU event tracing is enabled. Sep 12 18:07:28.880539 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 12 18:07:28.880552 kernel: Trampoline variant of Tasks RCU enabled. Sep 12 18:07:28.880566 kernel: Rude variant of Tasks RCU enabled. Sep 12 18:07:28.880579 kernel: Tracing variant of Tasks RCU enabled. Sep 12 18:07:28.880592 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 12 18:07:28.880606 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 12 18:07:28.880619 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 18:07:28.880643 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 18:07:28.880660 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 18:07:28.880675 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 12 18:07:28.880688 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 12 18:07:28.880701 kernel: Console: colour VGA+ 80x25 Sep 12 18:07:28.880715 kernel: printk: legacy console [tty0] enabled Sep 12 18:07:28.880728 kernel: printk: legacy console [ttyS0] enabled Sep 12 18:07:28.880740 kernel: ACPI: Core revision 20240827 Sep 12 18:07:28.880755 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 12 18:07:28.880783 kernel: APIC: Switch to symmetric I/O mode setup Sep 12 18:07:28.880798 kernel: x2apic enabled Sep 12 18:07:28.880812 kernel: APIC: Switched APIC routing to: physical x2apic Sep 12 18:07:28.880830 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 12 18:07:28.880848 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Sep 12 18:07:28.880862 kernel: Calibrating delay loop (skipped) preset value.. 4988.28 BogoMIPS (lpj=2494140) Sep 12 18:07:28.880876 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Sep 12 18:07:28.880890 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Sep 12 18:07:28.880904 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 12 18:07:28.880922 kernel: Spectre V2 : Mitigation: Retpolines Sep 12 18:07:28.880937 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 12 18:07:28.880951 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Sep 12 18:07:28.880966 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 12 18:07:28.880979 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 12 18:07:28.880993 kernel: MDS: Mitigation: Clear CPU buffers Sep 12 18:07:28.881008 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 12 18:07:28.881026 kernel: active return thunk: its_return_thunk Sep 12 18:07:28.881040 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 12 18:07:28.881055 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 12 18:07:28.881069 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 12 18:07:28.881083 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 12 18:07:28.881097 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 12 18:07:28.881111 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Sep 12 18:07:28.881126 kernel: Freeing SMP alternatives memory: 32K Sep 12 18:07:28.881140 kernel: pid_max: default: 32768 minimum: 301 Sep 12 18:07:28.881175 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 12 18:07:28.881193 kernel: landlock: Up and running. Sep 12 18:07:28.881207 kernel: SELinux: Initializing. Sep 12 18:07:28.881220 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 12 18:07:28.881232 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 12 18:07:28.881246 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Sep 12 18:07:28.881260 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Sep 12 18:07:28.881274 kernel: signal: max sigframe size: 1776 Sep 12 18:07:28.881289 kernel: rcu: Hierarchical SRCU implementation. Sep 12 18:07:28.881307 kernel: rcu: Max phase no-delay instances is 400. Sep 12 18:07:28.881321 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 12 18:07:28.881336 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 12 18:07:28.881350 kernel: smp: Bringing up secondary CPUs ... Sep 12 18:07:28.881368 kernel: smpboot: x86: Booting SMP configuration: Sep 12 18:07:28.881383 kernel: .... node #0, CPUs: #1 Sep 12 18:07:28.881397 kernel: smp: Brought up 1 node, 2 CPUs Sep 12 18:07:28.881411 kernel: smpboot: Total of 2 processors activated (9976.56 BogoMIPS) Sep 12 18:07:28.881426 kernel: Memory: 1966916K/2096612K available (14336K kernel code, 2432K rwdata, 9960K rodata, 54040K init, 2924K bss, 125140K reserved, 0K cma-reserved) Sep 12 18:07:28.881444 kernel: devtmpfs: initialized Sep 12 18:07:28.881458 kernel: x86/mm: Memory block size: 128MB Sep 12 18:07:28.881485 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 12 18:07:28.882006 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 12 18:07:28.882027 kernel: pinctrl core: initialized pinctrl subsystem Sep 12 18:07:28.882042 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 12 18:07:28.882056 kernel: audit: initializing netlink subsys (disabled) Sep 12 18:07:28.882071 kernel: audit: type=2000 audit(1757700445.415:1): state=initialized audit_enabled=0 res=1 Sep 12 18:07:28.882085 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 12 18:07:28.882106 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 12 18:07:28.882118 kernel: cpuidle: using governor menu Sep 12 18:07:28.882132 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 12 18:07:28.882147 kernel: dca service started, version 1.12.1 Sep 12 18:07:28.882162 kernel: PCI: Using configuration type 1 for base access Sep 12 18:07:28.882177 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 12 18:07:28.882191 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 12 18:07:28.882205 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 12 18:07:28.882224 kernel: ACPI: Added _OSI(Module Device) Sep 12 18:07:28.882242 kernel: ACPI: Added _OSI(Processor Device) Sep 12 18:07:28.882256 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 12 18:07:28.882270 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 12 18:07:28.882285 kernel: ACPI: Interpreter enabled Sep 12 18:07:28.882298 kernel: ACPI: PM: (supports S0 S5) Sep 12 18:07:28.882312 kernel: ACPI: Using IOAPIC for interrupt routing Sep 12 18:07:28.882326 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 12 18:07:28.882340 kernel: PCI: Using E820 reservations for host bridge windows Sep 12 18:07:28.882354 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Sep 12 18:07:28.882372 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 12 18:07:28.883465 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 12 18:07:28.883590 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Sep 12 18:07:28.883684 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Sep 12 18:07:28.883696 kernel: acpiphp: Slot [3] registered Sep 12 18:07:28.883706 kernel: acpiphp: Slot [4] registered Sep 12 18:07:28.883715 kernel: acpiphp: Slot [5] registered Sep 12 18:07:28.883728 kernel: acpiphp: Slot [6] registered Sep 12 18:07:28.883737 kernel: acpiphp: Slot [7] registered Sep 12 18:07:28.883746 kernel: acpiphp: Slot [8] registered Sep 12 18:07:28.883755 kernel: acpiphp: Slot [9] registered Sep 12 18:07:28.883764 kernel: acpiphp: Slot [10] registered Sep 12 18:07:28.883772 kernel: acpiphp: Slot [11] registered Sep 12 18:07:28.883781 kernel: acpiphp: Slot [12] registered Sep 12 18:07:28.883790 kernel: acpiphp: Slot [13] registered Sep 12 18:07:28.883798 kernel: acpiphp: Slot [14] registered Sep 12 18:07:28.883807 kernel: acpiphp: Slot [15] registered Sep 12 18:07:28.883818 kernel: acpiphp: Slot [16] registered Sep 12 18:07:28.883827 kernel: acpiphp: Slot [17] registered Sep 12 18:07:28.883836 kernel: acpiphp: Slot [18] registered Sep 12 18:07:28.883844 kernel: acpiphp: Slot [19] registered Sep 12 18:07:28.883854 kernel: acpiphp: Slot [20] registered Sep 12 18:07:28.883862 kernel: acpiphp: Slot [21] registered Sep 12 18:07:28.883871 kernel: acpiphp: Slot [22] registered Sep 12 18:07:28.883879 kernel: acpiphp: Slot [23] registered Sep 12 18:07:28.883888 kernel: acpiphp: Slot [24] registered Sep 12 18:07:28.883899 kernel: acpiphp: Slot [25] registered Sep 12 18:07:28.883908 kernel: acpiphp: Slot [26] registered Sep 12 18:07:28.883917 kernel: acpiphp: Slot [27] registered Sep 12 18:07:28.883926 kernel: acpiphp: Slot [28] registered Sep 12 18:07:28.883934 kernel: acpiphp: Slot [29] registered Sep 12 18:07:28.883943 kernel: acpiphp: Slot [30] registered Sep 12 18:07:28.883951 kernel: acpiphp: Slot [31] registered Sep 12 18:07:28.883960 kernel: PCI host bridge to bus 0000:00 Sep 12 18:07:28.884085 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 12 18:07:28.884177 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 12 18:07:28.884261 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 12 18:07:28.884343 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Sep 12 18:07:28.884425 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Sep 12 18:07:28.885596 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 12 18:07:28.885806 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint Sep 12 18:07:28.885972 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint Sep 12 18:07:28.886130 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 conventional PCI endpoint Sep 12 18:07:28.886266 kernel: pci 0000:00:01.1: BAR 4 [io 0xc1e0-0xc1ef] Sep 12 18:07:28.886408 kernel: pci 0000:00:01.1: BAR 0 [io 0x01f0-0x01f7]: legacy IDE quirk Sep 12 18:07:28.888650 kernel: pci 0000:00:01.1: BAR 1 [io 0x03f6]: legacy IDE quirk Sep 12 18:07:28.888776 kernel: pci 0000:00:01.1: BAR 2 [io 0x0170-0x0177]: legacy IDE quirk Sep 12 18:07:28.888872 kernel: pci 0000:00:01.1: BAR 3 [io 0x0376]: legacy IDE quirk Sep 12 18:07:28.888991 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 conventional PCI endpoint Sep 12 18:07:28.889085 kernel: pci 0000:00:01.2: BAR 4 [io 0xc180-0xc19f] Sep 12 18:07:28.889211 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 conventional PCI endpoint Sep 12 18:07:28.889304 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Sep 12 18:07:28.889397 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Sep 12 18:07:28.889507 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint Sep 12 18:07:28.889609 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfe000000-0xfe7fffff pref] Sep 12 18:07:28.889702 kernel: pci 0000:00:02.0: BAR 2 [mem 0xfe800000-0xfe803fff 64bit pref] Sep 12 18:07:28.889794 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfebf0000-0xfebf0fff] Sep 12 18:07:28.889884 kernel: pci 0000:00:02.0: ROM [mem 0xfebe0000-0xfebeffff pref] Sep 12 18:07:28.889975 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 12 18:07:28.890078 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Sep 12 18:07:28.890172 kernel: pci 0000:00:03.0: BAR 0 [io 0xc1a0-0xc1bf] Sep 12 18:07:28.890268 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebf1000-0xfebf1fff] Sep 12 18:07:28.890380 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe804000-0xfe807fff 64bit pref] Sep 12 18:07:28.892598 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Sep 12 18:07:28.892715 kernel: pci 0000:00:04.0: BAR 0 [io 0xc1c0-0xc1df] Sep 12 18:07:28.892812 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebf2000-0xfebf2fff] Sep 12 18:07:28.892904 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe808000-0xfe80bfff 64bit pref] Sep 12 18:07:28.893018 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint Sep 12 18:07:28.893118 kernel: pci 0000:00:05.0: BAR 0 [io 0xc100-0xc13f] Sep 12 18:07:28.893224 kernel: pci 0000:00:05.0: BAR 1 [mem 0xfebf3000-0xfebf3fff] Sep 12 18:07:28.893317 kernel: pci 0000:00:05.0: BAR 4 [mem 0xfe80c000-0xfe80ffff 64bit pref] Sep 12 18:07:28.893416 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Sep 12 18:07:28.893522 kernel: pci 0000:00:06.0: BAR 0 [io 0xc000-0xc07f] Sep 12 18:07:28.893613 kernel: pci 0000:00:06.0: BAR 1 [mem 0xfebf4000-0xfebf4fff] Sep 12 18:07:28.893703 kernel: pci 0000:00:06.0: BAR 4 [mem 0xfe810000-0xfe813fff 64bit pref] Sep 12 18:07:28.893809 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Sep 12 18:07:28.893901 kernel: pci 0000:00:07.0: BAR 0 [io 0xc080-0xc0ff] Sep 12 18:07:28.893991 kernel: pci 0000:00:07.0: BAR 1 [mem 0xfebf5000-0xfebf5fff] Sep 12 18:07:28.894082 kernel: pci 0000:00:07.0: BAR 4 [mem 0xfe814000-0xfe817fff 64bit pref] Sep 12 18:07:28.894188 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 conventional PCI endpoint Sep 12 18:07:28.894280 kernel: pci 0000:00:08.0: BAR 0 [io 0xc140-0xc17f] Sep 12 18:07:28.894374 kernel: pci 0000:00:08.0: BAR 4 [mem 0xfe818000-0xfe81bfff 64bit pref] Sep 12 18:07:28.894386 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 12 18:07:28.894395 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 12 18:07:28.894404 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 12 18:07:28.894413 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 12 18:07:28.894422 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 12 18:07:28.894431 kernel: iommu: Default domain type: Translated Sep 12 18:07:28.894439 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 12 18:07:28.894451 kernel: PCI: Using ACPI for IRQ routing Sep 12 18:07:28.894460 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 12 18:07:28.894469 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 12 18:07:28.896546 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Sep 12 18:07:28.896695 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Sep 12 18:07:28.896796 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Sep 12 18:07:28.896892 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 12 18:07:28.896904 kernel: vgaarb: loaded Sep 12 18:07:28.896913 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 12 18:07:28.896928 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 12 18:07:28.896937 kernel: clocksource: Switched to clocksource kvm-clock Sep 12 18:07:28.896946 kernel: VFS: Disk quotas dquot_6.6.0 Sep 12 18:07:28.896955 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 12 18:07:28.896964 kernel: pnp: PnP ACPI init Sep 12 18:07:28.896973 kernel: pnp: PnP ACPI: found 4 devices Sep 12 18:07:28.896982 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 12 18:07:28.896990 kernel: NET: Registered PF_INET protocol family Sep 12 18:07:28.896999 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 12 18:07:28.897011 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 12 18:07:28.897020 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 12 18:07:28.897029 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 12 18:07:28.897038 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 12 18:07:28.897046 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 12 18:07:28.897055 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 12 18:07:28.897064 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 12 18:07:28.897073 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 12 18:07:28.897082 kernel: NET: Registered PF_XDP protocol family Sep 12 18:07:28.897190 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 12 18:07:28.897276 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 12 18:07:28.897364 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 12 18:07:28.897446 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Sep 12 18:07:28.897541 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Sep 12 18:07:28.897639 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Sep 12 18:07:28.897736 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 12 18:07:28.897753 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Sep 12 18:07:28.897847 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x720 took 27667 usecs Sep 12 18:07:28.897859 kernel: PCI: CLS 0 bytes, default 64 Sep 12 18:07:28.897868 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 12 18:07:28.897877 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Sep 12 18:07:28.897886 kernel: Initialise system trusted keyrings Sep 12 18:07:28.897895 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 12 18:07:28.897904 kernel: Key type asymmetric registered Sep 12 18:07:28.897912 kernel: Asymmetric key parser 'x509' registered Sep 12 18:07:28.897924 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 12 18:07:28.897933 kernel: io scheduler mq-deadline registered Sep 12 18:07:28.897942 kernel: io scheduler kyber registered Sep 12 18:07:28.897950 kernel: io scheduler bfq registered Sep 12 18:07:28.897959 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 12 18:07:28.897968 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Sep 12 18:07:28.897976 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Sep 12 18:07:28.897985 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Sep 12 18:07:28.897993 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 12 18:07:28.898005 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 12 18:07:28.898013 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 12 18:07:28.898022 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 12 18:07:28.898031 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 12 18:07:28.898039 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 12 18:07:28.898156 kernel: rtc_cmos 00:03: RTC can wake from S4 Sep 12 18:07:28.898246 kernel: rtc_cmos 00:03: registered as rtc0 Sep 12 18:07:28.898332 kernel: rtc_cmos 00:03: setting system clock to 2025-09-12T18:07:28 UTC (1757700448) Sep 12 18:07:28.898421 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Sep 12 18:07:28.898432 kernel: intel_pstate: CPU model not supported Sep 12 18:07:28.898441 kernel: NET: Registered PF_INET6 protocol family Sep 12 18:07:28.898449 kernel: Segment Routing with IPv6 Sep 12 18:07:28.898458 kernel: In-situ OAM (IOAM) with IPv6 Sep 12 18:07:28.898467 kernel: NET: Registered PF_PACKET protocol family Sep 12 18:07:28.900510 kernel: Key type dns_resolver registered Sep 12 18:07:28.900527 kernel: IPI shorthand broadcast: enabled Sep 12 18:07:28.900537 kernel: sched_clock: Marking stable (3233003113, 140899919)->(3497318484, -123415452) Sep 12 18:07:28.900551 kernel: registered taskstats version 1 Sep 12 18:07:28.900560 kernel: Loading compiled-in X.509 certificates Sep 12 18:07:28.900569 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.47-flatcar: f1ae8d6e9bfae84d90f4136cf098b0465b2a5bd7' Sep 12 18:07:28.900577 kernel: Demotion targets for Node 0: null Sep 12 18:07:28.900586 kernel: Key type .fscrypt registered Sep 12 18:07:28.900595 kernel: Key type fscrypt-provisioning registered Sep 12 18:07:28.900624 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 12 18:07:28.900636 kernel: ima: Allocated hash algorithm: sha1 Sep 12 18:07:28.900645 kernel: ima: No architecture policies found Sep 12 18:07:28.900657 kernel: clk: Disabling unused clocks Sep 12 18:07:28.900666 kernel: Warning: unable to open an initial console. Sep 12 18:07:28.900675 kernel: Freeing unused kernel image (initmem) memory: 54040K Sep 12 18:07:28.900685 kernel: Write protecting the kernel read-only data: 24576k Sep 12 18:07:28.900694 kernel: Freeing unused kernel image (rodata/data gap) memory: 280K Sep 12 18:07:28.900703 kernel: Run /init as init process Sep 12 18:07:28.900712 kernel: with arguments: Sep 12 18:07:28.900722 kernel: /init Sep 12 18:07:28.900731 kernel: with environment: Sep 12 18:07:28.900742 kernel: HOME=/ Sep 12 18:07:28.900752 kernel: TERM=linux Sep 12 18:07:28.900761 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 12 18:07:28.900773 systemd[1]: Successfully made /usr/ read-only. Sep 12 18:07:28.900786 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 18:07:28.900797 systemd[1]: Detected virtualization kvm. Sep 12 18:07:28.900806 systemd[1]: Detected architecture x86-64. Sep 12 18:07:28.900818 systemd[1]: Running in initrd. Sep 12 18:07:28.900828 systemd[1]: No hostname configured, using default hostname. Sep 12 18:07:28.900838 systemd[1]: Hostname set to . Sep 12 18:07:28.900847 systemd[1]: Initializing machine ID from VM UUID. Sep 12 18:07:28.900857 systemd[1]: Queued start job for default target initrd.target. Sep 12 18:07:28.900867 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 18:07:28.900877 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 18:07:28.900888 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 12 18:07:28.900901 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 18:07:28.900911 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 12 18:07:28.900924 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 12 18:07:28.900935 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 12 18:07:28.900948 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 12 18:07:28.900959 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 18:07:28.900974 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 18:07:28.900988 systemd[1]: Reached target paths.target - Path Units. Sep 12 18:07:28.901003 systemd[1]: Reached target slices.target - Slice Units. Sep 12 18:07:28.901017 systemd[1]: Reached target swap.target - Swaps. Sep 12 18:07:28.901032 systemd[1]: Reached target timers.target - Timer Units. Sep 12 18:07:28.901046 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 18:07:28.901063 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 18:07:28.901073 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 12 18:07:28.901083 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 12 18:07:28.901093 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 18:07:28.901103 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 18:07:28.901113 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 18:07:28.901123 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 18:07:28.901132 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 12 18:07:28.901142 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 18:07:28.901170 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 12 18:07:28.901180 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 12 18:07:28.901190 systemd[1]: Starting systemd-fsck-usr.service... Sep 12 18:07:28.901199 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 18:07:28.901209 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 18:07:28.901218 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 18:07:28.901228 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 12 18:07:28.901241 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 18:07:28.901251 systemd[1]: Finished systemd-fsck-usr.service. Sep 12 18:07:28.901313 systemd-journald[212]: Collecting audit messages is disabled. Sep 12 18:07:28.901340 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 12 18:07:28.901352 systemd-journald[212]: Journal started Sep 12 18:07:28.901374 systemd-journald[212]: Runtime Journal (/run/log/journal/e96b813a14a54fcb878444dae4b94936) is 4.9M, max 39.5M, 34.6M free. Sep 12 18:07:28.882519 systemd-modules-load[213]: Inserted module 'overlay' Sep 12 18:07:28.926113 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 18:07:28.926148 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 12 18:07:28.926164 kernel: Bridge firewalling registered Sep 12 18:07:28.920582 systemd-modules-load[213]: Inserted module 'br_netfilter' Sep 12 18:07:28.927200 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 18:07:28.927927 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:28.928467 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 18:07:28.932448 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 12 18:07:28.933679 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 18:07:28.936604 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 18:07:28.944297 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 18:07:28.958972 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 18:07:28.964083 systemd-tmpfiles[233]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 12 18:07:28.965585 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 18:07:28.970235 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 18:07:28.972014 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 18:07:28.974436 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 18:07:28.976669 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 12 18:07:29.003381 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=271a44cc8ea1639cfb6fdf777202a5f025fda0b3ce9b293cc4e0e7047aecb858 Sep 12 18:07:29.025103 systemd-resolved[251]: Positive Trust Anchors: Sep 12 18:07:29.025118 systemd-resolved[251]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 18:07:29.025205 systemd-resolved[251]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 18:07:29.029322 systemd-resolved[251]: Defaulting to hostname 'linux'. Sep 12 18:07:29.031609 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 18:07:29.032276 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 18:07:29.113505 kernel: SCSI subsystem initialized Sep 12 18:07:29.124505 kernel: Loading iSCSI transport class v2.0-870. Sep 12 18:07:29.136511 kernel: iscsi: registered transport (tcp) Sep 12 18:07:29.160563 kernel: iscsi: registered transport (qla4xxx) Sep 12 18:07:29.160664 kernel: QLogic iSCSI HBA Driver Sep 12 18:07:29.189678 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 18:07:29.208655 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 18:07:29.211732 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 18:07:29.281122 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 12 18:07:29.283729 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 12 18:07:29.352567 kernel: raid6: avx2x4 gen() 14137 MB/s Sep 12 18:07:29.369551 kernel: raid6: avx2x2 gen() 15684 MB/s Sep 12 18:07:29.387037 kernel: raid6: avx2x1 gen() 13722 MB/s Sep 12 18:07:29.387141 kernel: raid6: using algorithm avx2x2 gen() 15684 MB/s Sep 12 18:07:29.404644 kernel: raid6: .... xor() 12681 MB/s, rmw enabled Sep 12 18:07:29.404758 kernel: raid6: using avx2x2 recovery algorithm Sep 12 18:07:29.436541 kernel: xor: automatically using best checksumming function avx Sep 12 18:07:29.697534 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 12 18:07:29.709054 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 12 18:07:29.711886 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 18:07:29.753330 systemd-udevd[461]: Using default interface naming scheme 'v255'. Sep 12 18:07:29.764370 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 18:07:29.769444 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 12 18:07:29.806890 dracut-pre-trigger[470]: rd.md=0: removing MD RAID activation Sep 12 18:07:29.843180 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 18:07:29.847229 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 18:07:29.920511 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 18:07:29.923132 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 12 18:07:30.015546 kernel: libata version 3.00 loaded. Sep 12 18:07:30.030507 kernel: cryptd: max_cpu_qlen set to 1000 Sep 12 18:07:30.030572 kernel: ata_piix 0000:00:01.1: version 2.13 Sep 12 18:07:30.035513 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Sep 12 18:07:30.043513 kernel: virtio_scsi virtio3: 2/0/0 default/read/poll queues Sep 12 18:07:30.048931 kernel: scsi host0: Virtio SCSI HBA Sep 12 18:07:30.048983 kernel: scsi host1: ata_piix Sep 12 18:07:30.053591 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Sep 12 18:07:30.056424 kernel: scsi host2: ata_piix Sep 12 18:07:30.056682 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 lpm-pol 0 Sep 12 18:07:30.056699 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 lpm-pol 0 Sep 12 18:07:30.073945 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 12 18:07:30.074035 kernel: GPT:9289727 != 125829119 Sep 12 18:07:30.074050 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 12 18:07:30.074080 kernel: GPT:9289727 != 125829119 Sep 12 18:07:30.074092 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 12 18:07:30.074106 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 18:07:30.082527 kernel: AES CTR mode by8 optimization enabled Sep 12 18:07:30.086076 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Sep 12 18:07:30.086457 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Sep 12 18:07:30.086498 kernel: virtio_blk virtio5: [vdb] 980 512-byte logical blocks (502 kB/490 KiB) Sep 12 18:07:30.112770 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 18:07:30.112914 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:30.113630 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 18:07:30.115784 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 18:07:30.122859 kernel: ACPI: bus type USB registered Sep 12 18:07:30.122922 kernel: usbcore: registered new interface driver usbfs Sep 12 18:07:30.122937 kernel: usbcore: registered new interface driver hub Sep 12 18:07:30.122344 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 18:07:30.124504 kernel: usbcore: registered new device driver usb Sep 12 18:07:30.180148 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:30.306230 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Sep 12 18:07:30.306568 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Sep 12 18:07:30.309378 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Sep 12 18:07:30.310938 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Sep 12 18:07:30.311126 kernel: hub 1-0:1.0: USB hub found Sep 12 18:07:30.312104 kernel: hub 1-0:1.0: 2 ports detected Sep 12 18:07:30.313356 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 12 18:07:30.326941 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 12 18:07:30.337231 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 12 18:07:30.345933 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 12 18:07:30.346398 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 12 18:07:30.355616 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 12 18:07:30.356158 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 18:07:30.356821 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 18:07:30.357760 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 18:07:30.359404 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 12 18:07:30.360491 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 12 18:07:30.388763 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 12 18:07:30.396544 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 18:07:30.397530 disk-uuid[618]: Primary Header is updated. Sep 12 18:07:30.397530 disk-uuid[618]: Secondary Entries is updated. Sep 12 18:07:30.397530 disk-uuid[618]: Secondary Header is updated. Sep 12 18:07:31.420185 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 18:07:31.420979 disk-uuid[626]: The operation has completed successfully. Sep 12 18:07:31.478799 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 12 18:07:31.478959 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 12 18:07:31.509375 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 12 18:07:31.535592 sh[637]: Success Sep 12 18:07:31.556669 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 12 18:07:31.556746 kernel: device-mapper: uevent: version 1.0.3 Sep 12 18:07:31.558039 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 12 18:07:31.569507 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" Sep 12 18:07:31.623199 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 12 18:07:31.626134 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 12 18:07:31.642344 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 12 18:07:31.656526 kernel: BTRFS: device fsid 74707491-1b86-4926-8bdb-c533ce2a0c32 devid 1 transid 38 /dev/mapper/usr (253:0) scanned by mount (649) Sep 12 18:07:31.660905 kernel: BTRFS info (device dm-0): first mount of filesystem 74707491-1b86-4926-8bdb-c533ce2a0c32 Sep 12 18:07:31.660978 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 12 18:07:31.667615 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 12 18:07:31.667696 kernel: BTRFS info (device dm-0): enabling free space tree Sep 12 18:07:31.670555 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 12 18:07:31.671454 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 12 18:07:31.672318 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 12 18:07:31.673344 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 12 18:07:31.675547 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 12 18:07:31.715524 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 (254:6) scanned by mount (679) Sep 12 18:07:31.718860 kernel: BTRFS info (device vda6): first mount of filesystem 5410dae6-8d31-4ea4-a4b4-868064445761 Sep 12 18:07:31.718926 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 12 18:07:31.723849 kernel: BTRFS info (device vda6): turning on async discard Sep 12 18:07:31.723936 kernel: BTRFS info (device vda6): enabling free space tree Sep 12 18:07:31.731500 kernel: BTRFS info (device vda6): last unmount of filesystem 5410dae6-8d31-4ea4-a4b4-868064445761 Sep 12 18:07:31.732350 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 12 18:07:31.736695 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 12 18:07:31.856396 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 18:07:31.867227 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 18:07:31.915056 systemd-networkd[819]: lo: Link UP Sep 12 18:07:31.915787 systemd-networkd[819]: lo: Gained carrier Sep 12 18:07:31.919558 systemd-networkd[819]: Enumeration completed Sep 12 18:07:31.919732 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 18:07:31.919983 systemd-networkd[819]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Sep 12 18:07:31.919989 systemd-networkd[819]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Sep 12 18:07:31.920293 systemd[1]: Reached target network.target - Network. Sep 12 18:07:31.923939 systemd-networkd[819]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 18:07:31.923945 systemd-networkd[819]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 12 18:07:31.926879 systemd-networkd[819]: eth0: Link UP Sep 12 18:07:31.927182 systemd-networkd[819]: eth1: Link UP Sep 12 18:07:31.927464 systemd-networkd[819]: eth0: Gained carrier Sep 12 18:07:31.927777 systemd-networkd[819]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Sep 12 18:07:31.934747 systemd-networkd[819]: eth1: Gained carrier Sep 12 18:07:31.934777 systemd-networkd[819]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 18:07:31.938041 ignition[726]: Ignition 2.21.0 Sep 12 18:07:31.938060 ignition[726]: Stage: fetch-offline Sep 12 18:07:31.938118 ignition[726]: no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:31.938132 ignition[726]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:31.941693 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 18:07:31.938255 ignition[726]: parsed url from cmdline: "" Sep 12 18:07:31.938260 ignition[726]: no config URL provided Sep 12 18:07:31.938267 ignition[726]: reading system config file "/usr/lib/ignition/user.ign" Sep 12 18:07:31.938275 ignition[726]: no config at "/usr/lib/ignition/user.ign" Sep 12 18:07:31.944718 systemd-networkd[819]: eth0: DHCPv4 address 161.35.232.141/20, gateway 161.35.224.1 acquired from 169.254.169.253 Sep 12 18:07:31.938282 ignition[726]: failed to fetch config: resource requires networking Sep 12 18:07:31.945897 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 12 18:07:31.939176 ignition[726]: Ignition finished successfully Sep 12 18:07:31.957636 systemd-networkd[819]: eth1: DHCPv4 address 10.124.0.32/20 acquired from 169.254.169.253 Sep 12 18:07:31.983703 ignition[827]: Ignition 2.21.0 Sep 12 18:07:31.983747 ignition[827]: Stage: fetch Sep 12 18:07:31.983992 ignition[827]: no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:31.984004 ignition[827]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:31.984181 ignition[827]: parsed url from cmdline: "" Sep 12 18:07:31.984186 ignition[827]: no config URL provided Sep 12 18:07:31.984197 ignition[827]: reading system config file "/usr/lib/ignition/user.ign" Sep 12 18:07:31.984206 ignition[827]: no config at "/usr/lib/ignition/user.ign" Sep 12 18:07:31.984255 ignition[827]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Sep 12 18:07:32.028536 ignition[827]: GET result: OK Sep 12 18:07:32.028730 ignition[827]: parsing config with SHA512: 4cb13429e7851d209be8049e228e15fa773b4870bb5a1376a76c378237be194d944512ad2c6d45ade3c79d21c345b65c73b105bfa1d755d21ff6bf5afb947887 Sep 12 18:07:32.033836 unknown[827]: fetched base config from "system" Sep 12 18:07:32.033848 unknown[827]: fetched base config from "system" Sep 12 18:07:32.034278 ignition[827]: fetch: fetch complete Sep 12 18:07:32.033855 unknown[827]: fetched user config from "digitalocean" Sep 12 18:07:32.034286 ignition[827]: fetch: fetch passed Sep 12 18:07:32.034377 ignition[827]: Ignition finished successfully Sep 12 18:07:32.037520 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 12 18:07:32.039157 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 12 18:07:32.079328 ignition[835]: Ignition 2.21.0 Sep 12 18:07:32.080182 ignition[835]: Stage: kargs Sep 12 18:07:32.080745 ignition[835]: no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:32.081124 ignition[835]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:32.083848 ignition[835]: kargs: kargs passed Sep 12 18:07:32.083979 ignition[835]: Ignition finished successfully Sep 12 18:07:32.086745 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 12 18:07:32.088574 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 12 18:07:32.121330 ignition[842]: Ignition 2.21.0 Sep 12 18:07:32.122081 ignition[842]: Stage: disks Sep 12 18:07:32.122281 ignition[842]: no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:32.122296 ignition[842]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:32.127682 ignition[842]: disks: disks passed Sep 12 18:07:32.127779 ignition[842]: Ignition finished successfully Sep 12 18:07:32.130277 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 12 18:07:32.131328 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 12 18:07:32.131791 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 12 18:07:32.132551 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 18:07:32.133390 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 18:07:32.134007 systemd[1]: Reached target basic.target - Basic System. Sep 12 18:07:32.135946 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 12 18:07:32.167539 systemd-fsck[851]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 12 18:07:32.171374 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 12 18:07:32.174359 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 12 18:07:32.331528 kernel: EXT4-fs (vda9): mounted filesystem 26739aba-b0be-4ce3-bfbd-ca4dbcbe2426 r/w with ordered data mode. Quota mode: none. Sep 12 18:07:32.332652 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 12 18:07:32.333904 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 12 18:07:32.336402 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 18:07:32.339214 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 12 18:07:32.344923 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... Sep 12 18:07:32.353722 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Sep 12 18:07:32.355300 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 12 18:07:32.355447 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 18:07:32.367531 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 (254:6) scanned by mount (859) Sep 12 18:07:32.368646 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 12 18:07:32.372507 kernel: BTRFS info (device vda6): first mount of filesystem 5410dae6-8d31-4ea4-a4b4-868064445761 Sep 12 18:07:32.374805 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 12 18:07:32.380707 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 12 18:07:32.387511 kernel: BTRFS info (device vda6): turning on async discard Sep 12 18:07:32.387608 kernel: BTRFS info (device vda6): enabling free space tree Sep 12 18:07:32.391405 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 18:07:32.466550 coreos-metadata[862]: Sep 12 18:07:32.464 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 18:07:32.473823 coreos-metadata[861]: Sep 12 18:07:32.473 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 18:07:32.479889 coreos-metadata[862]: Sep 12 18:07:32.479 INFO Fetch successful Sep 12 18:07:32.483688 initrd-setup-root[889]: cut: /sysroot/etc/passwd: No such file or directory Sep 12 18:07:32.488101 coreos-metadata[861]: Sep 12 18:07:32.486 INFO Fetch successful Sep 12 18:07:32.488805 coreos-metadata[862]: Sep 12 18:07:32.488 INFO wrote hostname ci-4426.1.0-c-cd41a2aea4 to /sysroot/etc/hostname Sep 12 18:07:32.490130 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 12 18:07:32.497050 initrd-setup-root[897]: cut: /sysroot/etc/group: No such file or directory Sep 12 18:07:32.499211 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. Sep 12 18:07:32.499545 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. Sep 12 18:07:32.506321 initrd-setup-root[905]: cut: /sysroot/etc/shadow: No such file or directory Sep 12 18:07:32.512511 initrd-setup-root[912]: cut: /sysroot/etc/gshadow: No such file or directory Sep 12 18:07:32.636571 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 12 18:07:32.638366 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 12 18:07:32.639746 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 12 18:07:32.662562 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 12 18:07:32.664579 kernel: BTRFS info (device vda6): last unmount of filesystem 5410dae6-8d31-4ea4-a4b4-868064445761 Sep 12 18:07:32.684855 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 12 18:07:32.711532 ignition[981]: INFO : Ignition 2.21.0 Sep 12 18:07:32.711532 ignition[981]: INFO : Stage: mount Sep 12 18:07:32.711532 ignition[981]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:32.711532 ignition[981]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:32.715386 ignition[981]: INFO : mount: mount passed Sep 12 18:07:32.716804 ignition[981]: INFO : Ignition finished successfully Sep 12 18:07:32.717634 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 12 18:07:32.720184 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 12 18:07:32.744009 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 18:07:32.779532 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 (254:6) scanned by mount (992) Sep 12 18:07:32.782891 kernel: BTRFS info (device vda6): first mount of filesystem 5410dae6-8d31-4ea4-a4b4-868064445761 Sep 12 18:07:32.782968 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 12 18:07:32.787989 kernel: BTRFS info (device vda6): turning on async discard Sep 12 18:07:32.788088 kernel: BTRFS info (device vda6): enabling free space tree Sep 12 18:07:32.791155 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 18:07:32.825587 ignition[1008]: INFO : Ignition 2.21.0 Sep 12 18:07:32.825587 ignition[1008]: INFO : Stage: files Sep 12 18:07:32.826776 ignition[1008]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:32.826776 ignition[1008]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:32.829253 ignition[1008]: DEBUG : files: compiled without relabeling support, skipping Sep 12 18:07:32.830653 ignition[1008]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 12 18:07:32.830653 ignition[1008]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 12 18:07:32.833095 ignition[1008]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 12 18:07:32.833923 ignition[1008]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 12 18:07:32.835086 unknown[1008]: wrote ssh authorized keys file for user: core Sep 12 18:07:32.835854 ignition[1008]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 12 18:07:32.838562 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 12 18:07:32.839358 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 12 18:07:32.962922 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 12 18:07:33.030680 systemd-networkd[819]: eth0: Gained IPv6LL Sep 12 18:07:33.088918 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 12 18:07:33.088918 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 18:07:33.091008 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 18:07:33.099297 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 18:07:33.099297 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 12 18:07:33.099297 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 12 18:07:33.099297 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 12 18:07:33.099297 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Sep 12 18:07:33.351939 systemd-networkd[819]: eth1: Gained IPv6LL Sep 12 18:07:33.706221 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 12 18:07:34.763846 ignition[1008]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 12 18:07:34.763846 ignition[1008]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 12 18:07:34.765807 ignition[1008]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 18:07:34.766692 ignition[1008]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 18:07:34.766692 ignition[1008]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 12 18:07:34.766692 ignition[1008]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Sep 12 18:07:34.769816 ignition[1008]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Sep 12 18:07:34.769816 ignition[1008]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 12 18:07:34.769816 ignition[1008]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 12 18:07:34.769816 ignition[1008]: INFO : files: files passed Sep 12 18:07:34.769816 ignition[1008]: INFO : Ignition finished successfully Sep 12 18:07:34.768642 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 12 18:07:34.771693 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 12 18:07:34.774965 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 12 18:07:34.786435 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 12 18:07:34.786989 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 12 18:07:34.797391 initrd-setup-root-after-ignition[1040]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 18:07:34.797391 initrd-setup-root-after-ignition[1040]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 12 18:07:34.799639 initrd-setup-root-after-ignition[1044]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 18:07:34.801683 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 18:07:34.802552 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 12 18:07:34.804509 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 12 18:07:34.858521 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 12 18:07:34.858646 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 12 18:07:34.859649 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 12 18:07:34.860254 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 12 18:07:34.861152 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 12 18:07:34.862426 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 12 18:07:34.890413 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 18:07:34.893689 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 12 18:07:34.919719 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 12 18:07:34.920726 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 18:07:34.921159 systemd[1]: Stopped target timers.target - Timer Units. Sep 12 18:07:34.922519 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 12 18:07:34.922662 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 18:07:34.924046 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 12 18:07:34.924613 systemd[1]: Stopped target basic.target - Basic System. Sep 12 18:07:34.925126 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 12 18:07:34.925883 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 18:07:34.926630 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 12 18:07:34.927560 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 12 18:07:34.928161 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 12 18:07:34.928805 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 18:07:34.929724 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 12 18:07:34.930485 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 12 18:07:34.931193 systemd[1]: Stopped target swap.target - Swaps. Sep 12 18:07:34.931885 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 12 18:07:34.932094 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 12 18:07:34.933191 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 12 18:07:34.933744 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 18:07:34.934360 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 12 18:07:34.934509 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 18:07:34.935177 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 12 18:07:34.935365 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 12 18:07:34.936165 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 12 18:07:34.936322 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 18:07:34.936980 systemd[1]: ignition-files.service: Deactivated successfully. Sep 12 18:07:34.937128 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 12 18:07:34.937765 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 12 18:07:34.937952 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 12 18:07:34.939644 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 12 18:07:34.941808 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 12 18:07:34.942018 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 18:07:34.953774 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 12 18:07:34.954866 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 12 18:07:34.955721 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 18:07:34.962290 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 12 18:07:34.963809 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 18:07:34.968653 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 12 18:07:34.973652 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 12 18:07:34.976860 ignition[1064]: INFO : Ignition 2.21.0 Sep 12 18:07:34.976860 ignition[1064]: INFO : Stage: umount Sep 12 18:07:34.977840 ignition[1064]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 18:07:34.977840 ignition[1064]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 18:07:34.978698 ignition[1064]: INFO : umount: umount passed Sep 12 18:07:34.978698 ignition[1064]: INFO : Ignition finished successfully Sep 12 18:07:34.980968 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 12 18:07:34.981547 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 12 18:07:34.982056 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 12 18:07:34.982103 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 12 18:07:34.982418 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 12 18:07:34.982451 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 12 18:07:34.983622 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 12 18:07:34.983666 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 12 18:07:34.983998 systemd[1]: Stopped target network.target - Network. Sep 12 18:07:34.984667 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 12 18:07:34.984720 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 18:07:34.985176 systemd[1]: Stopped target paths.target - Path Units. Sep 12 18:07:34.985624 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 12 18:07:34.989817 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 18:07:34.990553 systemd[1]: Stopped target slices.target - Slice Units. Sep 12 18:07:34.991013 systemd[1]: Stopped target sockets.target - Socket Units. Sep 12 18:07:34.991356 systemd[1]: iscsid.socket: Deactivated successfully. Sep 12 18:07:34.991419 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 18:07:34.991780 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 12 18:07:34.991816 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 18:07:34.992280 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 12 18:07:34.992347 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 12 18:07:34.994738 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 12 18:07:34.994800 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 12 18:07:34.995464 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 12 18:07:35.010650 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 12 18:07:35.019638 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 12 18:07:35.026229 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 12 18:07:35.026402 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 12 18:07:35.028879 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 12 18:07:35.030722 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 12 18:07:35.033074 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 12 18:07:35.033128 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 12 18:07:35.038579 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 12 18:07:35.039338 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 12 18:07:35.039840 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 18:07:35.040275 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 18:07:35.043944 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 12 18:07:35.047849 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 12 18:07:35.052627 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 12 18:07:35.053229 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 12 18:07:35.054572 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 12 18:07:35.056244 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 12 18:07:35.056427 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 18:07:35.059224 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 12 18:07:35.059329 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 12 18:07:35.060227 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 12 18:07:35.060264 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 18:07:35.061121 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 12 18:07:35.061273 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 12 18:07:35.064621 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 12 18:07:35.064691 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 12 18:07:35.065102 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 12 18:07:35.065154 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 18:07:35.065731 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 12 18:07:35.065778 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 12 18:07:35.071510 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 12 18:07:35.072014 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 12 18:07:35.072109 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 18:07:35.075421 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 12 18:07:35.075520 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 12 18:07:35.076626 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 12 18:07:35.076679 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 12 18:07:35.077304 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 12 18:07:35.077361 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 18:07:35.080157 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 12 18:07:35.080244 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 18:07:35.081621 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 18:07:35.081710 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:35.086429 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Sep 12 18:07:35.088179 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 12 18:07:35.088263 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 12 18:07:35.088337 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 18:07:35.088393 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 12 18:07:35.089151 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 12 18:07:35.090158 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 12 18:07:35.091428 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 12 18:07:35.091546 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 12 18:07:35.093508 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 12 18:07:35.094780 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 12 18:07:35.120892 systemd[1]: Switching root. Sep 12 18:07:35.180170 systemd-journald[212]: Journal stopped Sep 12 18:07:36.348285 systemd-journald[212]: Received SIGTERM from PID 1 (systemd). Sep 12 18:07:36.348353 kernel: SELinux: policy capability network_peer_controls=1 Sep 12 18:07:36.348371 kernel: SELinux: policy capability open_perms=1 Sep 12 18:07:36.348384 kernel: SELinux: policy capability extended_socket_class=1 Sep 12 18:07:36.348413 kernel: SELinux: policy capability always_check_network=0 Sep 12 18:07:36.348426 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 12 18:07:36.348445 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 12 18:07:36.348458 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 12 18:07:36.348484 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 12 18:07:36.348501 kernel: SELinux: policy capability userspace_initial_context=0 Sep 12 18:07:36.348514 kernel: audit: type=1403 audit(1757700455.312:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 12 18:07:36.348530 systemd[1]: Successfully loaded SELinux policy in 66.926ms. Sep 12 18:07:36.348560 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.490ms. Sep 12 18:07:36.348579 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 18:07:36.348598 systemd[1]: Detected virtualization kvm. Sep 12 18:07:36.348612 systemd[1]: Detected architecture x86-64. Sep 12 18:07:36.348626 systemd[1]: Detected first boot. Sep 12 18:07:36.348641 systemd[1]: Hostname set to . Sep 12 18:07:36.348655 systemd[1]: Initializing machine ID from VM UUID. Sep 12 18:07:36.348669 zram_generator::config[1110]: No configuration found. Sep 12 18:07:36.348690 kernel: Guest personality initialized and is inactive Sep 12 18:07:36.348703 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 12 18:07:36.348716 kernel: Initialized host personality Sep 12 18:07:36.348729 kernel: NET: Registered PF_VSOCK protocol family Sep 12 18:07:36.348743 systemd[1]: Populated /etc with preset unit settings. Sep 12 18:07:36.348758 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 12 18:07:36.348772 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 12 18:07:36.348786 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 12 18:07:36.348800 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 12 18:07:36.348820 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 12 18:07:36.348835 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 12 18:07:36.354611 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 12 18:07:36.354638 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 12 18:07:36.354654 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 12 18:07:36.354670 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 12 18:07:36.354684 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 12 18:07:36.354705 systemd[1]: Created slice user.slice - User and Session Slice. Sep 12 18:07:36.354719 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 18:07:36.354744 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 18:07:36.354759 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 12 18:07:36.354773 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 12 18:07:36.354788 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 12 18:07:36.354803 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 18:07:36.354817 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 12 18:07:36.354837 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 18:07:36.354856 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 18:07:36.354871 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 12 18:07:36.354892 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 12 18:07:36.354906 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 12 18:07:36.354920 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 12 18:07:36.354934 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 18:07:36.354948 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 18:07:36.354963 systemd[1]: Reached target slices.target - Slice Units. Sep 12 18:07:36.354983 systemd[1]: Reached target swap.target - Swaps. Sep 12 18:07:36.354998 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 12 18:07:36.355012 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 12 18:07:36.355026 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 12 18:07:36.355040 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 18:07:36.355054 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 18:07:36.355069 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 18:07:36.355082 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 12 18:07:36.355097 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 12 18:07:36.355124 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 12 18:07:36.355138 systemd[1]: Mounting media.mount - External Media Directory... Sep 12 18:07:36.355153 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:36.355167 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 12 18:07:36.355182 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 12 18:07:36.355196 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 12 18:07:36.355216 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 12 18:07:36.355229 systemd[1]: Reached target machines.target - Containers. Sep 12 18:07:36.355243 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 12 18:07:36.355264 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 18:07:36.355278 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 18:07:36.355292 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 12 18:07:36.355307 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 18:07:36.355320 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 18:07:36.355334 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 18:07:36.355349 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 12 18:07:36.355362 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 18:07:36.355383 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 12 18:07:36.355398 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 12 18:07:36.355412 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 12 18:07:36.355426 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 12 18:07:36.355439 systemd[1]: Stopped systemd-fsck-usr.service. Sep 12 18:07:36.355454 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 18:07:36.355468 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 18:07:36.358596 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 18:07:36.358622 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 18:07:36.358644 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 12 18:07:36.358659 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 12 18:07:36.358674 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 18:07:36.358696 systemd[1]: verity-setup.service: Deactivated successfully. Sep 12 18:07:36.358711 systemd[1]: Stopped verity-setup.service. Sep 12 18:07:36.358726 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:36.358740 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 12 18:07:36.358754 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 12 18:07:36.358769 systemd[1]: Mounted media.mount - External Media Directory. Sep 12 18:07:36.358784 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 12 18:07:36.358807 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 12 18:07:36.358822 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 12 18:07:36.358836 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 18:07:36.358850 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 12 18:07:36.358865 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 12 18:07:36.358879 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 18:07:36.358893 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 18:07:36.358907 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 18:07:36.358929 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 18:07:36.358943 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 18:07:36.358958 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 12 18:07:36.358972 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 12 18:07:36.358987 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 12 18:07:36.359001 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 18:07:36.359015 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 12 18:07:36.359029 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 12 18:07:36.359043 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 18:07:36.359064 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 12 18:07:36.359130 systemd-journald[1177]: Collecting audit messages is disabled. Sep 12 18:07:36.359158 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 18:07:36.359182 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 12 18:07:36.359199 systemd-journald[1177]: Journal started Sep 12 18:07:36.359230 systemd-journald[1177]: Runtime Journal (/run/log/journal/e96b813a14a54fcb878444dae4b94936) is 4.9M, max 39.5M, 34.6M free. Sep 12 18:07:36.038025 systemd[1]: Queued start job for default target multi-user.target. Sep 12 18:07:36.062286 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 12 18:07:36.062745 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 12 18:07:36.366950 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 18:07:36.372495 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 12 18:07:36.381554 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 18:07:36.383673 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 18:07:36.384430 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 12 18:07:36.408506 kernel: loop: module loaded Sep 12 18:07:36.404432 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 18:07:36.418603 kernel: fuse: init (API version 7.41) Sep 12 18:07:36.420078 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 12 18:07:36.422133 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 12 18:07:36.422302 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 12 18:07:36.427907 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 18:07:36.430024 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 18:07:36.430775 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 12 18:07:36.439720 systemd-journald[1177]: Time spent on flushing to /var/log/journal/e96b813a14a54fcb878444dae4b94936 is 96.826ms for 1003 entries. Sep 12 18:07:36.439720 systemd-journald[1177]: System Journal (/var/log/journal/e96b813a14a54fcb878444dae4b94936) is 8M, max 195.6M, 187.6M free. Sep 12 18:07:36.559254 systemd-journald[1177]: Received client request to flush runtime journal. Sep 12 18:07:36.559330 kernel: loop0: detected capacity change from 0 to 221472 Sep 12 18:07:36.559381 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 12 18:07:36.559397 kernel: loop1: detected capacity change from 0 to 111000 Sep 12 18:07:36.438670 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 12 18:07:36.443385 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 12 18:07:36.450610 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 12 18:07:36.457443 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 12 18:07:36.457869 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 18:07:36.459721 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 12 18:07:36.472625 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 12 18:07:36.474119 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 12 18:07:36.524270 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 18:07:36.540555 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 12 18:07:36.566053 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 12 18:07:36.612798 kernel: loop2: detected capacity change from 0 to 8 Sep 12 18:07:36.614941 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 12 18:07:36.616833 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 18:07:36.620581 kernel: ACPI: bus type drm_connector registered Sep 12 18:07:36.624358 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 18:07:36.624588 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 18:07:36.656501 kernel: loop3: detected capacity change from 0 to 128016 Sep 12 18:07:36.671006 systemd-tmpfiles[1252]: ACLs are not supported, ignoring. Sep 12 18:07:36.672517 systemd-tmpfiles[1252]: ACLs are not supported, ignoring. Sep 12 18:07:36.686268 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 18:07:36.693310 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 18:07:36.699540 kernel: loop4: detected capacity change from 0 to 221472 Sep 12 18:07:36.744506 kernel: loop5: detected capacity change from 0 to 111000 Sep 12 18:07:36.776227 kernel: loop6: detected capacity change from 0 to 8 Sep 12 18:07:36.785424 kernel: loop7: detected capacity change from 0 to 128016 Sep 12 18:07:36.814536 (sd-merge)[1259]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Sep 12 18:07:36.819832 (sd-merge)[1259]: Merged extensions into '/usr'. Sep 12 18:07:36.833687 systemd[1]: Reload requested from client PID 1212 ('systemd-sysext') (unit systemd-sysext.service)... Sep 12 18:07:36.834119 systemd[1]: Reloading... Sep 12 18:07:36.946573 zram_generator::config[1284]: No configuration found. Sep 12 18:07:37.091518 ldconfig[1204]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 12 18:07:37.225928 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 12 18:07:37.226255 systemd[1]: Reloading finished in 390 ms. Sep 12 18:07:37.254345 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 12 18:07:37.256293 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 12 18:07:37.270689 systemd[1]: Starting ensure-sysext.service... Sep 12 18:07:37.272299 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 18:07:37.306921 systemd[1]: Reload requested from client PID 1328 ('systemctl') (unit ensure-sysext.service)... Sep 12 18:07:37.307103 systemd[1]: Reloading... Sep 12 18:07:37.324335 systemd-tmpfiles[1329]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 12 18:07:37.324372 systemd-tmpfiles[1329]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 12 18:07:37.324697 systemd-tmpfiles[1329]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 12 18:07:37.324985 systemd-tmpfiles[1329]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 12 18:07:37.327842 systemd-tmpfiles[1329]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 12 18:07:37.328179 systemd-tmpfiles[1329]: ACLs are not supported, ignoring. Sep 12 18:07:37.328240 systemd-tmpfiles[1329]: ACLs are not supported, ignoring. Sep 12 18:07:37.337087 systemd-tmpfiles[1329]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 18:07:37.337100 systemd-tmpfiles[1329]: Skipping /boot Sep 12 18:07:37.352175 systemd-tmpfiles[1329]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 18:07:37.352189 systemd-tmpfiles[1329]: Skipping /boot Sep 12 18:07:37.405572 zram_generator::config[1356]: No configuration found. Sep 12 18:07:37.639889 systemd[1]: Reloading finished in 332 ms. Sep 12 18:07:37.666709 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 12 18:07:37.676963 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 18:07:37.687470 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 18:07:37.691887 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 12 18:07:37.694437 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 12 18:07:37.703600 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 18:07:37.708000 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 18:07:37.712831 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 12 18:07:37.716568 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:37.716873 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 18:07:37.721843 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 18:07:37.732660 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 18:07:37.745970 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 18:07:37.746912 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 18:07:37.747057 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 18:07:37.747159 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:37.751179 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:37.751554 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 18:07:37.751793 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 18:07:37.751876 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 18:07:37.759770 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 12 18:07:37.760208 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:37.769096 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:37.769425 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 18:07:37.771430 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 18:07:37.772051 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 18:07:37.772166 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 18:07:37.772304 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:37.777705 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 12 18:07:37.785413 systemd[1]: Finished ensure-sysext.service. Sep 12 18:07:37.791187 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 12 18:07:37.798844 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 18:07:37.802872 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 18:07:37.804559 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 12 18:07:37.811108 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 12 18:07:37.826978 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 18:07:37.831048 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 18:07:37.836935 systemd-udevd[1405]: Using default interface naming scheme 'v255'. Sep 12 18:07:37.837369 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 12 18:07:37.838968 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 18:07:37.840888 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 18:07:37.844751 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 18:07:37.844818 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 18:07:37.844843 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 18:07:37.856921 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 18:07:37.857130 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 18:07:37.868565 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 12 18:07:37.874110 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 18:07:37.880411 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 18:07:37.893306 augenrules[1460]: No rules Sep 12 18:07:37.897242 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 18:07:37.898013 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 18:07:37.927951 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 12 18:07:38.072654 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. Sep 12 18:07:38.075974 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Sep 12 18:07:38.077563 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:38.077786 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 18:07:38.079657 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 18:07:38.082818 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 18:07:38.085692 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 18:07:38.086973 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 18:07:38.087010 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 18:07:38.087042 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 18:07:38.087060 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 18:07:38.094678 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 12 18:07:38.099940 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 18:07:38.100693 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 18:07:38.101943 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 18:07:38.140497 kernel: ISO 9660 Extensions: RRIP_1991A Sep 12 18:07:38.144882 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 18:07:38.145370 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 18:07:38.151222 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Sep 12 18:07:38.153826 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 18:07:38.155569 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 18:07:38.157346 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 18:07:38.164504 kernel: mousedev: PS/2 mouse device common for all mice Sep 12 18:07:38.248392 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 12 18:07:38.250908 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 12 18:07:38.251728 systemd[1]: Reached target time-set.target - System Time Set. Sep 12 18:07:38.257925 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 12 18:07:38.264575 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 12 18:07:38.305610 systemd-networkd[1450]: lo: Link UP Sep 12 18:07:38.305621 systemd-networkd[1450]: lo: Gained carrier Sep 12 18:07:38.312552 kernel: ACPI: button: Power Button [PWRF] Sep 12 18:07:38.313735 systemd-networkd[1450]: Enumeration completed Sep 12 18:07:38.315612 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 18:07:38.319285 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 12 18:07:38.322932 systemd-networkd[1450]: eth0: Configuring with /run/systemd/network/10-0e:b6:0d:8e:18:a1.network. Sep 12 18:07:38.323702 systemd-networkd[1450]: eth1: Configuring with /run/systemd/network/10-be:31:5b:b7:58:a0.network. Sep 12 18:07:38.325335 systemd-networkd[1450]: eth0: Link UP Sep 12 18:07:38.327441 systemd-networkd[1450]: eth0: Gained carrier Sep 12 18:07:38.327718 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 12 18:07:38.330020 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 12 18:07:38.335824 systemd-networkd[1450]: eth1: Link UP Sep 12 18:07:38.337096 systemd-networkd[1450]: eth1: Gained carrier Sep 12 18:07:38.343221 systemd-resolved[1404]: Positive Trust Anchors: Sep 12 18:07:38.343597 systemd-resolved[1404]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 18:07:38.343698 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:38.344029 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:38.344193 systemd-resolved[1404]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 18:07:38.350094 systemd-resolved[1404]: Using system hostname 'ci-4426.1.0-c-cd41a2aea4'. Sep 12 18:07:38.352191 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 18:07:38.352856 systemd[1]: Reached target network.target - Network. Sep 12 18:07:38.353367 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 18:07:38.353920 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 18:07:38.355227 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 12 18:07:38.355638 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 12 18:07:38.355979 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Sep 12 18:07:38.356463 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 12 18:07:38.356894 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 12 18:07:38.357518 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 12 18:07:38.357866 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 12 18:07:38.357892 systemd[1]: Reached target paths.target - Path Units. Sep 12 18:07:38.358332 systemd[1]: Reached target timers.target - Timer Units. Sep 12 18:07:38.359521 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 12 18:07:38.362821 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 12 18:07:38.366296 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 12 18:07:38.367537 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 12 18:07:38.367925 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 12 18:07:38.372260 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 12 18:07:38.373142 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 12 18:07:38.375354 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 12 18:07:38.376036 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 12 18:07:38.378470 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 18:07:38.380813 systemd[1]: Reached target basic.target - Basic System. Sep 12 18:07:38.381346 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 12 18:07:38.381377 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 12 18:07:38.383465 systemd[1]: Starting containerd.service - containerd container runtime... Sep 12 18:07:38.388716 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 12 18:07:38.391643 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 12 18:07:38.397109 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 12 18:07:38.404593 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 12 18:07:38.418584 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Sep 12 18:07:38.418881 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 12 18:07:38.416426 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 12 18:07:38.417567 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 12 18:07:38.420927 jq[1517]: false Sep 12 18:07:38.421359 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Sep 12 18:07:38.423259 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 12 18:07:38.428680 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 12 18:07:38.431630 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 12 18:07:38.437718 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 12 18:07:38.444706 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 12 18:07:38.445987 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 12 18:07:38.447600 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 12 18:07:38.454742 systemd[1]: Starting update-engine.service - Update Engine... Sep 12 18:07:38.471660 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 12 18:07:38.479677 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 12 18:07:38.481716 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 12 18:07:38.482022 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 12 18:07:38.499274 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 12 18:07:38.499850 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 12 18:07:38.503257 google_oslogin_nss_cache[1520]: oslogin_cache_refresh[1520]: Refreshing passwd entry cache Sep 12 18:07:38.501081 oslogin_cache_refresh[1520]: Refreshing passwd entry cache Sep 12 18:07:38.513156 extend-filesystems[1519]: Found /dev/vda6 Sep 12 18:07:38.519533 google_oslogin_nss_cache[1520]: oslogin_cache_refresh[1520]: Failure getting users, quitting Sep 12 18:07:38.519533 google_oslogin_nss_cache[1520]: oslogin_cache_refresh[1520]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 12 18:07:38.519533 google_oslogin_nss_cache[1520]: oslogin_cache_refresh[1520]: Refreshing group entry cache Sep 12 18:07:38.517224 oslogin_cache_refresh[1520]: Failure getting users, quitting Sep 12 18:07:38.517247 oslogin_cache_refresh[1520]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 12 18:07:38.517301 oslogin_cache_refresh[1520]: Refreshing group entry cache Sep 12 18:07:38.525548 google_oslogin_nss_cache[1520]: oslogin_cache_refresh[1520]: Failure getting groups, quitting Sep 12 18:07:38.525548 google_oslogin_nss_cache[1520]: oslogin_cache_refresh[1520]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 12 18:07:38.522363 oslogin_cache_refresh[1520]: Failure getting groups, quitting Sep 12 18:07:38.522378 oslogin_cache_refresh[1520]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 12 18:07:38.529545 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Sep 12 18:07:38.530826 jq[1533]: true Sep 12 18:07:38.536585 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Sep 12 18:07:38.545870 coreos-metadata[1514]: Sep 12 18:07:38.545 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 18:07:38.543246 systemd[1]: motdgen.service: Deactivated successfully. Sep 12 18:07:38.543975 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 12 18:07:38.559510 update_engine[1530]: I20250912 18:07:38.552136 1530 main.cc:92] Flatcar Update Engine starting Sep 12 18:07:38.559913 extend-filesystems[1519]: Found /dev/vda9 Sep 12 18:07:38.563495 (ntainerd)[1562]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 12 18:07:38.567880 coreos-metadata[1514]: Sep 12 18:07:38.564 INFO Fetch successful Sep 12 18:07:38.574535 extend-filesystems[1519]: Checking size of /dev/vda9 Sep 12 18:07:38.597982 tar[1542]: linux-amd64/helm Sep 12 18:07:38.615153 jq[1558]: true Sep 12 18:07:38.643847 extend-filesystems[1519]: Resized partition /dev/vda9 Sep 12 18:07:38.647404 dbus-daemon[1515]: [system] SELinux support is enabled Sep 12 18:07:38.647829 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 12 18:07:38.654084 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 12 18:07:38.654126 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 12 18:07:38.657861 extend-filesystems[1576]: resize2fs 1.47.2 (1-Jan-2025) Sep 12 18:07:38.668925 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Sep 12 18:07:38.656867 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 12 18:07:38.656987 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Sep 12 18:07:38.657012 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 12 18:07:38.671682 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 12 18:07:38.672393 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 12 18:07:38.688503 systemd[1]: Started update-engine.service - Update Engine. Sep 12 18:07:38.696136 update_engine[1530]: I20250912 18:07:38.695181 1530 update_check_scheduler.cc:74] Next update check in 3m11s Sep 12 18:07:38.694343 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 12 18:07:38.729444 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Sep 12 18:07:38.768503 extend-filesystems[1576]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 12 18:07:38.768503 extend-filesystems[1576]: old_desc_blocks = 1, new_desc_blocks = 8 Sep 12 18:07:38.768503 extend-filesystems[1576]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Sep 12 18:07:38.766109 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 12 18:07:38.780675 extend-filesystems[1519]: Resized filesystem in /dev/vda9 Sep 12 18:07:38.766320 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 12 18:07:38.787958 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 12 18:07:38.801318 systemd-logind[1528]: New seat seat0. Sep 12 18:07:38.808142 systemd[1]: Started systemd-logind.service - User Login Management. Sep 12 18:07:38.831164 bash[1595]: Updated "/home/core/.ssh/authorized_keys" Sep 12 18:07:38.816567 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 12 18:07:38.823688 systemd[1]: Starting sshkeys.service... Sep 12 18:07:38.862825 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Sep 12 18:07:38.867008 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Sep 12 18:07:38.981990 coreos-metadata[1599]: Sep 12 18:07:38.981 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 18:07:39.001401 coreos-metadata[1599]: Sep 12 18:07:38.999 INFO Fetch successful Sep 12 18:07:39.013374 unknown[1599]: wrote ssh authorized keys file for user: core Sep 12 18:07:39.051289 update-ssh-keys[1609]: Updated "/home/core/.ssh/authorized_keys" Sep 12 18:07:39.055046 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Sep 12 18:07:39.058633 systemd[1]: Finished sshkeys.service. Sep 12 18:07:39.119898 locksmithd[1578]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 12 18:07:39.135200 containerd[1562]: time="2025-09-12T18:07:39Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 12 18:07:39.137746 containerd[1562]: time="2025-09-12T18:07:39.135769795Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 12 18:07:39.148273 containerd[1562]: time="2025-09-12T18:07:39.148221428Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="10.604µs" Sep 12 18:07:39.148273 containerd[1562]: time="2025-09-12T18:07:39.148257964Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 12 18:07:39.148273 containerd[1562]: time="2025-09-12T18:07:39.148276559Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148446443Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148464667Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148506469Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148562762Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148573656Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148804450Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148818737Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148829848Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148837440Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.148907891Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150023 containerd[1562]: time="2025-09-12T18:07:39.149142034Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150330 containerd[1562]: time="2025-09-12T18:07:39.149190949Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 12 18:07:39.150330 containerd[1562]: time="2025-09-12T18:07:39.149204912Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 12 18:07:39.150330 containerd[1562]: time="2025-09-12T18:07:39.149247572Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 12 18:07:39.150330 containerd[1562]: time="2025-09-12T18:07:39.149462493Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 12 18:07:39.150330 containerd[1562]: time="2025-09-12T18:07:39.149541305Z" level=info msg="metadata content store policy set" policy=shared Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156296567Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156363501Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156378795Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156391423Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156405327Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156416245Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156444695Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156486542Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156500652Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156510483Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156520674Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 12 18:07:39.156627 containerd[1562]: time="2025-09-12T18:07:39.156533274Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156672252Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156691036Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156727639Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156742076Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156762101Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156774385Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156784803Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156794269Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156804645Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156815525Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156825585Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156903469Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 12 18:07:39.156941 containerd[1562]: time="2025-09-12T18:07:39.156924964Z" level=info msg="Start snapshots syncer" Sep 12 18:07:39.157213 containerd[1562]: time="2025-09-12T18:07:39.156954795Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 12 18:07:39.157241 containerd[1562]: time="2025-09-12T18:07:39.157202170Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 12 18:07:39.157372 containerd[1562]: time="2025-09-12T18:07:39.157274063Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157407556Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157561363Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157593782Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157613709Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157631308Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157650630Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157666384Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157676323Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157709569Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157724864Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157739428Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157780588Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157798653Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 12 18:07:39.160821 containerd[1562]: time="2025-09-12T18:07:39.157807500Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157816508Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157824168Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157835501Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157844863Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157860691Z" level=info msg="runtime interface created" Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157866456Z" level=info msg="created NRI interface" Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157874756Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157886474Z" level=info msg="Connect containerd service" Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.157910358Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 12 18:07:39.161209 containerd[1562]: time="2025-09-12T18:07:39.158670482Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 12 18:07:39.204215 systemd-logind[1528]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 12 18:07:39.206750 systemd-logind[1528]: Watching system buttons on /dev/input/event2 (Power Button) Sep 12 18:07:39.207996 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 18:07:39.218507 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Sep 12 18:07:39.218584 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Sep 12 18:07:39.258509 kernel: Console: switching to colour dummy device 80x25 Sep 12 18:07:39.261508 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Sep 12 18:07:39.261578 kernel: [drm] features: -context_init Sep 12 18:07:39.267898 kernel: [drm] number of scanouts: 1 Sep 12 18:07:39.267978 kernel: [drm] number of cap sets: 0 Sep 12 18:07:39.267993 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:02.0 on minor 0 Sep 12 18:07:39.277663 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Sep 12 18:07:39.277784 kernel: Console: switching to colour frame buffer device 128x48 Sep 12 18:07:39.280573 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Sep 12 18:07:39.364883 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 18:07:39.367655 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:39.375138 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 18:07:39.380607 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426137789Z" level=info msg="Start subscribing containerd event" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426196563Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426212666Z" level=info msg="Start recovering state" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426282494Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426344494Z" level=info msg="Start event monitor" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426358183Z" level=info msg="Start cni network conf syncer for default" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426376672Z" level=info msg="Start streaming server" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426387818Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426395138Z" level=info msg="runtime interface starting up..." Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426400593Z" level=info msg="starting plugins..." Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426414026Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 12 18:07:39.426929 containerd[1562]: time="2025-09-12T18:07:39.426623626Z" level=info msg="containerd successfully booted in 0.292214s" Sep 12 18:07:39.426900 systemd[1]: Started containerd.service - containerd container runtime. Sep 12 18:07:39.446417 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 18:07:39.446874 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:39.450391 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 18:07:39.456624 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 18:07:39.557616 systemd-networkd[1450]: eth1: Gained IPv6LL Sep 12 18:07:39.558718 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:39.564410 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 12 18:07:39.576381 systemd[1]: Reached target network-online.target - Network is Online. Sep 12 18:07:39.587284 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 18:07:39.591037 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 12 18:07:39.612479 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 18:07:39.678166 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 12 18:07:39.688527 kernel: EDAC MC: Ver: 3.0.0 Sep 12 18:07:39.741272 sshd_keygen[1559]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 12 18:07:39.769291 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 12 18:07:39.774730 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 12 18:07:39.778886 systemd[1]: Started sshd@0-161.35.232.141:22-185.156.73.233:32352.service - OpenSSH per-connection server daemon (185.156.73.233:32352). Sep 12 18:07:39.832671 systemd[1]: issuegen.service: Deactivated successfully. Sep 12 18:07:39.833633 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 12 18:07:39.843903 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 12 18:07:39.846849 tar[1542]: linux-amd64/LICENSE Sep 12 18:07:39.846849 tar[1542]: linux-amd64/README.md Sep 12 18:07:39.870412 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 12 18:07:39.879047 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 12 18:07:39.883304 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 12 18:07:39.886756 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 12 18:07:39.888366 systemd[1]: Reached target getty.target - Login Prompts. Sep 12 18:07:40.133705 systemd-networkd[1450]: eth0: Gained IPv6LL Sep 12 18:07:40.134314 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:40.886637 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:07:40.889859 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 12 18:07:40.891918 systemd[1]: Startup finished in 3.307s (kernel) + 6.665s (initrd) + 5.644s (userspace) = 15.616s. Sep 12 18:07:40.900670 (kubelet)[1688]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 18:07:41.514308 sshd[1669]: Invalid user admin from 185.156.73.233 port 32352 Sep 12 18:07:41.609900 kubelet[1688]: E0912 18:07:41.609791 1688 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 18:07:41.613980 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 18:07:41.614236 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 18:07:41.615052 systemd[1]: kubelet.service: Consumed 1.345s CPU time, 263.2M memory peak. Sep 12 18:07:41.790690 sshd[1669]: Connection closed by invalid user admin 185.156.73.233 port 32352 [preauth] Sep 12 18:07:41.792686 systemd[1]: sshd@0-161.35.232.141:22-185.156.73.233:32352.service: Deactivated successfully. Sep 12 18:07:42.404889 systemd[1]: Started sshd@1-161.35.232.141:22-139.178.89.65:55058.service - OpenSSH per-connection server daemon (139.178.89.65:55058). Sep 12 18:07:42.488097 sshd[1703]: Accepted publickey for core from 139.178.89.65 port 55058 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:42.490072 sshd-session[1703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:42.498216 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 12 18:07:42.499389 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 12 18:07:42.508605 systemd-logind[1528]: New session 1 of user core. Sep 12 18:07:42.533068 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 12 18:07:42.537417 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 12 18:07:42.553404 (systemd)[1708]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 12 18:07:42.557891 systemd-logind[1528]: New session c1 of user core. Sep 12 18:07:42.781993 systemd[1708]: Queued start job for default target default.target. Sep 12 18:07:42.790912 systemd[1708]: Created slice app.slice - User Application Slice. Sep 12 18:07:42.791202 systemd[1708]: Reached target paths.target - Paths. Sep 12 18:07:42.791336 systemd[1708]: Reached target timers.target - Timers. Sep 12 18:07:42.793129 systemd[1708]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 12 18:07:42.809501 systemd[1708]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 12 18:07:42.809689 systemd[1708]: Reached target sockets.target - Sockets. Sep 12 18:07:42.809827 systemd[1708]: Reached target basic.target - Basic System. Sep 12 18:07:42.809909 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 12 18:07:42.811186 systemd[1708]: Reached target default.target - Main User Target. Sep 12 18:07:42.811243 systemd[1708]: Startup finished in 243ms. Sep 12 18:07:42.816820 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 12 18:07:42.885813 systemd[1]: Started sshd@2-161.35.232.141:22-139.178.89.65:55064.service - OpenSSH per-connection server daemon (139.178.89.65:55064). Sep 12 18:07:42.949616 sshd[1719]: Accepted publickey for core from 139.178.89.65 port 55064 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:42.951324 sshd-session[1719]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:42.958366 systemd-logind[1528]: New session 2 of user core. Sep 12 18:07:42.973838 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 12 18:07:43.035550 sshd[1722]: Connection closed by 139.178.89.65 port 55064 Sep 12 18:07:43.036339 sshd-session[1719]: pam_unix(sshd:session): session closed for user core Sep 12 18:07:43.049787 systemd[1]: sshd@2-161.35.232.141:22-139.178.89.65:55064.service: Deactivated successfully. Sep 12 18:07:43.051866 systemd[1]: session-2.scope: Deactivated successfully. Sep 12 18:07:43.052780 systemd-logind[1528]: Session 2 logged out. Waiting for processes to exit. Sep 12 18:07:43.056390 systemd[1]: Started sshd@3-161.35.232.141:22-139.178.89.65:55072.service - OpenSSH per-connection server daemon (139.178.89.65:55072). Sep 12 18:07:43.057589 systemd-logind[1528]: Removed session 2. Sep 12 18:07:43.118423 sshd[1728]: Accepted publickey for core from 139.178.89.65 port 55072 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:43.120364 sshd-session[1728]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:43.125871 systemd-logind[1528]: New session 3 of user core. Sep 12 18:07:43.143820 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 12 18:07:43.203260 sshd[1731]: Connection closed by 139.178.89.65 port 55072 Sep 12 18:07:43.203119 sshd-session[1728]: pam_unix(sshd:session): session closed for user core Sep 12 18:07:43.217083 systemd[1]: sshd@3-161.35.232.141:22-139.178.89.65:55072.service: Deactivated successfully. Sep 12 18:07:43.219528 systemd[1]: session-3.scope: Deactivated successfully. Sep 12 18:07:43.220393 systemd-logind[1528]: Session 3 logged out. Waiting for processes to exit. Sep 12 18:07:43.224259 systemd[1]: Started sshd@4-161.35.232.141:22-139.178.89.65:55076.service - OpenSSH per-connection server daemon (139.178.89.65:55076). Sep 12 18:07:43.226264 systemd-logind[1528]: Removed session 3. Sep 12 18:07:43.286834 sshd[1737]: Accepted publickey for core from 139.178.89.65 port 55076 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:43.288579 sshd-session[1737]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:43.294981 systemd-logind[1528]: New session 4 of user core. Sep 12 18:07:43.300723 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 12 18:07:43.364358 sshd[1740]: Connection closed by 139.178.89.65 port 55076 Sep 12 18:07:43.364957 sshd-session[1737]: pam_unix(sshd:session): session closed for user core Sep 12 18:07:43.381023 systemd[1]: sshd@4-161.35.232.141:22-139.178.89.65:55076.service: Deactivated successfully. Sep 12 18:07:43.382875 systemd[1]: session-4.scope: Deactivated successfully. Sep 12 18:07:43.383768 systemd-logind[1528]: Session 4 logged out. Waiting for processes to exit. Sep 12 18:07:43.387623 systemd[1]: Started sshd@5-161.35.232.141:22-139.178.89.65:55080.service - OpenSSH per-connection server daemon (139.178.89.65:55080). Sep 12 18:07:43.388805 systemd-logind[1528]: Removed session 4. Sep 12 18:07:43.445106 sshd[1746]: Accepted publickey for core from 139.178.89.65 port 55080 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:43.446884 sshd-session[1746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:43.452893 systemd-logind[1528]: New session 5 of user core. Sep 12 18:07:43.463770 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 12 18:07:43.533695 sudo[1750]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 12 18:07:43.534342 sudo[1750]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 18:07:43.549846 sudo[1750]: pam_unix(sudo:session): session closed for user root Sep 12 18:07:43.553461 sshd[1749]: Connection closed by 139.178.89.65 port 55080 Sep 12 18:07:43.554126 sshd-session[1746]: pam_unix(sshd:session): session closed for user core Sep 12 18:07:43.565167 systemd[1]: sshd@5-161.35.232.141:22-139.178.89.65:55080.service: Deactivated successfully. Sep 12 18:07:43.567385 systemd[1]: session-5.scope: Deactivated successfully. Sep 12 18:07:43.569130 systemd-logind[1528]: Session 5 logged out. Waiting for processes to exit. Sep 12 18:07:43.571988 systemd[1]: Started sshd@6-161.35.232.141:22-139.178.89.65:55088.service - OpenSSH per-connection server daemon (139.178.89.65:55088). Sep 12 18:07:43.573233 systemd-logind[1528]: Removed session 5. Sep 12 18:07:43.640927 sshd[1756]: Accepted publickey for core from 139.178.89.65 port 55088 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:43.642599 sshd-session[1756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:43.650065 systemd-logind[1528]: New session 6 of user core. Sep 12 18:07:43.655855 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 12 18:07:43.717229 sudo[1761]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 12 18:07:43.717784 sudo[1761]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 18:07:43.747168 sudo[1761]: pam_unix(sudo:session): session closed for user root Sep 12 18:07:43.756435 sudo[1760]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 12 18:07:43.757324 sudo[1760]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 18:07:43.770912 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 18:07:43.817905 augenrules[1783]: No rules Sep 12 18:07:43.819617 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 18:07:43.819833 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 18:07:43.821133 sudo[1760]: pam_unix(sudo:session): session closed for user root Sep 12 18:07:43.824988 sshd[1759]: Connection closed by 139.178.89.65 port 55088 Sep 12 18:07:43.824293 sshd-session[1756]: pam_unix(sshd:session): session closed for user core Sep 12 18:07:43.839696 systemd[1]: sshd@6-161.35.232.141:22-139.178.89.65:55088.service: Deactivated successfully. Sep 12 18:07:43.842059 systemd[1]: session-6.scope: Deactivated successfully. Sep 12 18:07:43.843769 systemd-logind[1528]: Session 6 logged out. Waiting for processes to exit. Sep 12 18:07:43.846291 systemd[1]: Started sshd@7-161.35.232.141:22-139.178.89.65:55096.service - OpenSSH per-connection server daemon (139.178.89.65:55096). Sep 12 18:07:43.847719 systemd-logind[1528]: Removed session 6. Sep 12 18:07:43.917011 sshd[1792]: Accepted publickey for core from 139.178.89.65 port 55096 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:07:43.918952 sshd-session[1792]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:07:43.926041 systemd-logind[1528]: New session 7 of user core. Sep 12 18:07:43.932815 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 12 18:07:43.994787 sudo[1796]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 12 18:07:43.995126 sudo[1796]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 18:07:44.502126 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 12 18:07:44.513046 (dockerd)[1814]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 12 18:07:44.913364 dockerd[1814]: time="2025-09-12T18:07:44.913203359Z" level=info msg="Starting up" Sep 12 18:07:44.918975 dockerd[1814]: time="2025-09-12T18:07:44.918918600Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 12 18:07:44.940680 dockerd[1814]: time="2025-09-12T18:07:44.940623662Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 12 18:07:44.960562 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1131933341-merged.mount: Deactivated successfully. Sep 12 18:07:44.984313 dockerd[1814]: time="2025-09-12T18:07:44.984254298Z" level=info msg="Loading containers: start." Sep 12 18:07:44.994499 kernel: Initializing XFRM netlink socket Sep 12 18:07:45.238752 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:45.240393 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:45.254243 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:45.293171 systemd-networkd[1450]: docker0: Link UP Sep 12 18:07:45.293562 systemd-timesyncd[1421]: Network configuration changed, trying to establish connection. Sep 12 18:07:45.297894 dockerd[1814]: time="2025-09-12T18:07:45.297827768Z" level=info msg="Loading containers: done." Sep 12 18:07:45.316106 dockerd[1814]: time="2025-09-12T18:07:45.316035988Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 12 18:07:45.316342 dockerd[1814]: time="2025-09-12T18:07:45.316145341Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 12 18:07:45.316342 dockerd[1814]: time="2025-09-12T18:07:45.316268083Z" level=info msg="Initializing buildkit" Sep 12 18:07:45.341541 dockerd[1814]: time="2025-09-12T18:07:45.341469129Z" level=info msg="Completed buildkit initialization" Sep 12 18:07:45.351031 dockerd[1814]: time="2025-09-12T18:07:45.350862807Z" level=info msg="Daemon has completed initialization" Sep 12 18:07:45.351511 dockerd[1814]: time="2025-09-12T18:07:45.351275323Z" level=info msg="API listen on /run/docker.sock" Sep 12 18:07:45.351821 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 12 18:07:46.306395 containerd[1562]: time="2025-09-12T18:07:46.306320563Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\"" Sep 12 18:07:46.924184 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1158712675.mount: Deactivated successfully. Sep 12 18:07:48.195770 containerd[1562]: time="2025-09-12T18:07:48.195720136Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:48.197730 containerd[1562]: time="2025-09-12T18:07:48.197685514Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.13: active requests=0, bytes read=28117124" Sep 12 18:07:48.198754 containerd[1562]: time="2025-09-12T18:07:48.198698540Z" level=info msg="ImageCreate event name:\"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:48.201665 containerd[1562]: time="2025-09-12T18:07:48.201630853Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:48.203504 containerd[1562]: time="2025-09-12T18:07:48.203209632Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.13\" with image id \"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.13\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e\", size \"28113723\" in 1.896707459s" Sep 12 18:07:48.203504 containerd[1562]: time="2025-09-12T18:07:48.203262818Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\" returns image reference \"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\"" Sep 12 18:07:48.204075 containerd[1562]: time="2025-09-12T18:07:48.204045863Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\"" Sep 12 18:07:49.848073 containerd[1562]: time="2025-09-12T18:07:49.847948012Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:49.849649 containerd[1562]: time="2025-09-12T18:07:49.849586696Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.13: active requests=0, bytes read=24716632" Sep 12 18:07:49.851508 containerd[1562]: time="2025-09-12T18:07:49.850158476Z" level=info msg="ImageCreate event name:\"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:49.853661 containerd[1562]: time="2025-09-12T18:07:49.853595262Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:49.855511 containerd[1562]: time="2025-09-12T18:07:49.855066277Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.13\" with image id \"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.13\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09\", size \"26351311\" in 1.65062924s" Sep 12 18:07:49.855511 containerd[1562]: time="2025-09-12T18:07:49.855114074Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\" returns image reference \"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\"" Sep 12 18:07:49.856061 containerd[1562]: time="2025-09-12T18:07:49.856025382Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\"" Sep 12 18:07:51.169635 containerd[1562]: time="2025-09-12T18:07:51.169558543Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:51.172101 containerd[1562]: time="2025-09-12T18:07:51.172024801Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.13: active requests=0, bytes read=18787698" Sep 12 18:07:51.172633 containerd[1562]: time="2025-09-12T18:07:51.172591035Z" level=info msg="ImageCreate event name:\"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:51.177537 containerd[1562]: time="2025-09-12T18:07:51.177423515Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:51.179182 containerd[1562]: time="2025-09-12T18:07:51.178984945Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.13\" with image id \"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.13\", repo digest \"registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241\", size \"20422395\" in 1.322913225s" Sep 12 18:07:51.179182 containerd[1562]: time="2025-09-12T18:07:51.179041220Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\" returns image reference \"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\"" Sep 12 18:07:51.179937 containerd[1562]: time="2025-09-12T18:07:51.179878775Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\"" Sep 12 18:07:51.747969 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 12 18:07:51.752638 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 18:07:52.004709 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:07:52.021094 (kubelet)[2106]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 18:07:52.106773 kubelet[2106]: E0912 18:07:52.106689 2106 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 18:07:52.115463 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 18:07:52.115716 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 18:07:52.116578 systemd[1]: kubelet.service: Consumed 249ms CPU time, 108.7M memory peak. Sep 12 18:07:52.513108 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2755248239.mount: Deactivated successfully. Sep 12 18:07:53.064450 containerd[1562]: time="2025-09-12T18:07:53.064392075Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:53.066912 containerd[1562]: time="2025-09-12T18:07:53.066838366Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.13: active requests=0, bytes read=30410252" Sep 12 18:07:53.067841 containerd[1562]: time="2025-09-12T18:07:53.067801112Z" level=info msg="ImageCreate event name:\"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:53.071112 containerd[1562]: time="2025-09-12T18:07:53.071034362Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:53.072934 containerd[1562]: time="2025-09-12T18:07:53.072750201Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.13\" with image id \"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\", repo tag \"registry.k8s.io/kube-proxy:v1.31.13\", repo digest \"registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae\", size \"30409271\" in 1.892679568s" Sep 12 18:07:53.072934 containerd[1562]: time="2025-09-12T18:07:53.072806982Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\" returns image reference \"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\"" Sep 12 18:07:53.073768 containerd[1562]: time="2025-09-12T18:07:53.073737070Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 12 18:07:53.075451 systemd-resolved[1404]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Sep 12 18:07:53.572322 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1671694624.mount: Deactivated successfully. Sep 12 18:07:54.533923 containerd[1562]: time="2025-09-12T18:07:54.533859485Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:54.534604 containerd[1562]: time="2025-09-12T18:07:54.534559991Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Sep 12 18:07:54.536038 containerd[1562]: time="2025-09-12T18:07:54.535992379Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:54.539379 containerd[1562]: time="2025-09-12T18:07:54.539331683Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:54.541081 containerd[1562]: time="2025-09-12T18:07:54.541034272Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.467154557s" Sep 12 18:07:54.541281 containerd[1562]: time="2025-09-12T18:07:54.541257573Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 12 18:07:54.542366 containerd[1562]: time="2025-09-12T18:07:54.542102786Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 12 18:07:55.083088 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2742278003.mount: Deactivated successfully. Sep 12 18:07:55.087590 containerd[1562]: time="2025-09-12T18:07:55.087249743Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 18:07:55.088286 containerd[1562]: time="2025-09-12T18:07:55.088253774Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 12 18:07:55.089155 containerd[1562]: time="2025-09-12T18:07:55.089123044Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 18:07:55.091337 containerd[1562]: time="2025-09-12T18:07:55.091265133Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 18:07:55.092619 containerd[1562]: time="2025-09-12T18:07:55.092148008Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 550.007569ms" Sep 12 18:07:55.092619 containerd[1562]: time="2025-09-12T18:07:55.092201625Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 12 18:07:55.092999 containerd[1562]: time="2025-09-12T18:07:55.092966884Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 12 18:07:55.578657 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3237386863.mount: Deactivated successfully. Sep 12 18:07:56.134189 systemd-resolved[1404]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Sep 12 18:07:57.422655 containerd[1562]: time="2025-09-12T18:07:57.421528898Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:57.422655 containerd[1562]: time="2025-09-12T18:07:57.422526378Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56910709" Sep 12 18:07:57.423312 containerd[1562]: time="2025-09-12T18:07:57.423131925Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:57.428054 containerd[1562]: time="2025-09-12T18:07:57.427990353Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:07:57.429845 containerd[1562]: time="2025-09-12T18:07:57.429779956Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.336577643s" Sep 12 18:07:57.429845 containerd[1562]: time="2025-09-12T18:07:57.429843348Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Sep 12 18:07:59.939331 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:07:59.939633 systemd[1]: kubelet.service: Consumed 249ms CPU time, 108.7M memory peak. Sep 12 18:07:59.943927 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 18:07:59.988095 systemd[1]: Reload requested from client PID 2256 ('systemctl') (unit session-7.scope)... Sep 12 18:07:59.988329 systemd[1]: Reloading... Sep 12 18:08:00.140581 zram_generator::config[2299]: No configuration found. Sep 12 18:08:00.407826 systemd[1]: Reloading finished in 418 ms. Sep 12 18:08:00.482882 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 12 18:08:00.483022 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 12 18:08:00.483430 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:08:00.483579 systemd[1]: kubelet.service: Consumed 133ms CPU time, 98.2M memory peak. Sep 12 18:08:00.486351 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 18:08:00.657981 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:08:00.671307 (kubelet)[2354]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 18:08:00.729513 kubelet[2354]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 18:08:00.729513 kubelet[2354]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 12 18:08:00.729513 kubelet[2354]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 18:08:00.729513 kubelet[2354]: I0912 18:08:00.729029 2354 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 18:08:00.971377 kubelet[2354]: I0912 18:08:00.970913 2354 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 12 18:08:00.971566 kubelet[2354]: I0912 18:08:00.971548 2354 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 18:08:00.972033 kubelet[2354]: I0912 18:08:00.972012 2354 server.go:934] "Client rotation is on, will bootstrap in background" Sep 12 18:08:01.000530 kubelet[2354]: E0912 18:08:01.000184 2354 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://161.35.232.141:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:01.003021 kubelet[2354]: I0912 18:08:01.002757 2354 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 18:08:01.020331 kubelet[2354]: I0912 18:08:01.020262 2354 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 12 18:08:01.025793 kubelet[2354]: I0912 18:08:01.025751 2354 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 18:08:01.026397 kubelet[2354]: I0912 18:08:01.026368 2354 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 12 18:08:01.026604 kubelet[2354]: I0912 18:08:01.026559 2354 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 18:08:01.026796 kubelet[2354]: I0912 18:08:01.026604 2354 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4426.1.0-c-cd41a2aea4","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 18:08:01.026910 kubelet[2354]: I0912 18:08:01.026806 2354 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 18:08:01.026910 kubelet[2354]: I0912 18:08:01.026817 2354 container_manager_linux.go:300] "Creating device plugin manager" Sep 12 18:08:01.027008 kubelet[2354]: I0912 18:08:01.026993 2354 state_mem.go:36] "Initialized new in-memory state store" Sep 12 18:08:01.029457 kubelet[2354]: I0912 18:08:01.029403 2354 kubelet.go:408] "Attempting to sync node with API server" Sep 12 18:08:01.029457 kubelet[2354]: I0912 18:08:01.029462 2354 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 18:08:01.029670 kubelet[2354]: I0912 18:08:01.029532 2354 kubelet.go:314] "Adding apiserver pod source" Sep 12 18:08:01.029670 kubelet[2354]: I0912 18:08:01.029572 2354 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 18:08:01.036168 kubelet[2354]: W0912 18:08:01.036084 2354 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://161.35.232.141:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4426.1.0-c-cd41a2aea4&limit=500&resourceVersion=0": dial tcp 161.35.232.141:6443: connect: connection refused Sep 12 18:08:01.036432 kubelet[2354]: E0912 18:08:01.036406 2354 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://161.35.232.141:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4426.1.0-c-cd41a2aea4&limit=500&resourceVersion=0\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:01.038148 kubelet[2354]: W0912 18:08:01.038092 2354 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://161.35.232.141:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 161.35.232.141:6443: connect: connection refused Sep 12 18:08:01.038372 kubelet[2354]: E0912 18:08:01.038323 2354 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://161.35.232.141:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:01.038649 kubelet[2354]: I0912 18:08:01.038523 2354 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 12 18:08:01.041932 kubelet[2354]: I0912 18:08:01.041892 2354 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 12 18:08:01.043134 kubelet[2354]: W0912 18:08:01.042874 2354 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 12 18:08:01.044224 kubelet[2354]: I0912 18:08:01.044202 2354 server.go:1274] "Started kubelet" Sep 12 18:08:01.047360 kubelet[2354]: I0912 18:08:01.047329 2354 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 18:08:01.052762 kubelet[2354]: E0912 18:08:01.049653 2354 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://161.35.232.141:6443/api/v1/namespaces/default/events\": dial tcp 161.35.232.141:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4426.1.0-c-cd41a2aea4.18649b465a2b7d84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4426.1.0-c-cd41a2aea4,UID:ci-4426.1.0-c-cd41a2aea4,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4426.1.0-c-cd41a2aea4,},FirstTimestamp:2025-09-12 18:08:01.044151684 +0000 UTC m=+0.366474732,LastTimestamp:2025-09-12 18:08:01.044151684 +0000 UTC m=+0.366474732,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4426.1.0-c-cd41a2aea4,}" Sep 12 18:08:01.055887 kubelet[2354]: I0912 18:08:01.055828 2354 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 18:08:01.056590 kubelet[2354]: I0912 18:08:01.056562 2354 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 12 18:08:01.056821 kubelet[2354]: E0912 18:08:01.056803 2354 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4426.1.0-c-cd41a2aea4\" not found" Sep 12 18:08:01.057137 kubelet[2354]: I0912 18:08:01.057119 2354 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 12 18:08:01.057211 kubelet[2354]: I0912 18:08:01.057180 2354 reconciler.go:26] "Reconciler: start to sync state" Sep 12 18:08:01.058353 kubelet[2354]: I0912 18:08:01.058331 2354 server.go:449] "Adding debug handlers to kubelet server" Sep 12 18:08:01.059737 kubelet[2354]: W0912 18:08:01.059672 2354 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://161.35.232.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 161.35.232.141:6443: connect: connection refused Sep 12 18:08:01.059807 kubelet[2354]: E0912 18:08:01.059761 2354 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://161.35.232.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:01.059879 kubelet[2354]: E0912 18:08:01.059849 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://161.35.232.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4426.1.0-c-cd41a2aea4?timeout=10s\": dial tcp 161.35.232.141:6443: connect: connection refused" interval="200ms" Sep 12 18:08:01.060886 kubelet[2354]: I0912 18:08:01.060834 2354 factory.go:221] Registration of the systemd container factory successfully Sep 12 18:08:01.062571 kubelet[2354]: I0912 18:08:01.062520 2354 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 18:08:01.063274 kubelet[2354]: I0912 18:08:01.063240 2354 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 18:08:01.063733 kubelet[2354]: I0912 18:08:01.063699 2354 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 18:08:01.066170 kubelet[2354]: I0912 18:08:01.064526 2354 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 18:08:01.066170 kubelet[2354]: I0912 18:08:01.066089 2354 factory.go:221] Registration of the containerd container factory successfully Sep 12 18:08:01.066827 kubelet[2354]: E0912 18:08:01.066795 2354 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 18:08:01.086100 kubelet[2354]: I0912 18:08:01.086056 2354 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 12 18:08:01.086100 kubelet[2354]: I0912 18:08:01.086087 2354 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 12 18:08:01.086100 kubelet[2354]: I0912 18:08:01.086113 2354 state_mem.go:36] "Initialized new in-memory state store" Sep 12 18:08:01.087993 kubelet[2354]: I0912 18:08:01.087953 2354 policy_none.go:49] "None policy: Start" Sep 12 18:08:01.091081 kubelet[2354]: I0912 18:08:01.090955 2354 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 12 18:08:01.091081 kubelet[2354]: I0912 18:08:01.091017 2354 state_mem.go:35] "Initializing new in-memory state store" Sep 12 18:08:01.099809 kubelet[2354]: I0912 18:08:01.099743 2354 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 12 18:08:01.102514 kubelet[2354]: I0912 18:08:01.101873 2354 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 12 18:08:01.102514 kubelet[2354]: I0912 18:08:01.101909 2354 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 12 18:08:01.102514 kubelet[2354]: I0912 18:08:01.101945 2354 kubelet.go:2321] "Starting kubelet main sync loop" Sep 12 18:08:01.102514 kubelet[2354]: E0912 18:08:01.102003 2354 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 18:08:01.108006 kubelet[2354]: W0912 18:08:01.107957 2354 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://161.35.232.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 161.35.232.141:6443: connect: connection refused Sep 12 18:08:01.108006 kubelet[2354]: E0912 18:08:01.108006 2354 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://161.35.232.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:01.113628 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 12 18:08:01.132707 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 12 18:08:01.138973 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 12 18:08:01.149095 kubelet[2354]: I0912 18:08:01.149030 2354 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 12 18:08:01.149468 kubelet[2354]: I0912 18:08:01.149433 2354 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 18:08:01.149564 kubelet[2354]: I0912 18:08:01.149462 2354 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 18:08:01.150154 kubelet[2354]: I0912 18:08:01.150102 2354 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 18:08:01.152551 kubelet[2354]: E0912 18:08:01.152497 2354 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4426.1.0-c-cd41a2aea4\" not found" Sep 12 18:08:01.220083 systemd[1]: Created slice kubepods-burstable-podf791c201e896ce34810ec8b08354135b.slice - libcontainer container kubepods-burstable-podf791c201e896ce34810ec8b08354135b.slice. Sep 12 18:08:01.240093 systemd[1]: Created slice kubepods-burstable-pod340546bd497b3f71404b0f6daba1be5c.slice - libcontainer container kubepods-burstable-pod340546bd497b3f71404b0f6daba1be5c.slice. Sep 12 18:08:01.250694 systemd[1]: Created slice kubepods-burstable-pod57f30eb7298f4b27fdb24c17107fca13.slice - libcontainer container kubepods-burstable-pod57f30eb7298f4b27fdb24c17107fca13.slice. Sep 12 18:08:01.252384 kubelet[2354]: I0912 18:08:01.252336 2354 kubelet_node_status.go:72] "Attempting to register node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.253310 kubelet[2354]: E0912 18:08:01.253122 2354 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://161.35.232.141:6443/api/v1/nodes\": dial tcp 161.35.232.141:6443: connect: connection refused" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.258771 kubelet[2354]: I0912 18:08:01.258598 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-kubeconfig\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.258771 kubelet[2354]: I0912 18:08:01.258745 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f791c201e896ce34810ec8b08354135b-k8s-certs\") pod \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" (UID: \"f791c201e896ce34810ec8b08354135b\") " pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259022 kubelet[2354]: I0912 18:08:01.258817 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-ca-certs\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259022 kubelet[2354]: I0912 18:08:01.258852 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-flexvolume-dir\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259022 kubelet[2354]: I0912 18:08:01.258882 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-k8s-certs\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259022 kubelet[2354]: I0912 18:08:01.258910 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259022 kubelet[2354]: I0912 18:08:01.258938 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/57f30eb7298f4b27fdb24c17107fca13-kubeconfig\") pod \"kube-scheduler-ci-4426.1.0-c-cd41a2aea4\" (UID: \"57f30eb7298f4b27fdb24c17107fca13\") " pod="kube-system/kube-scheduler-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259230 kubelet[2354]: I0912 18:08:01.258965 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f791c201e896ce34810ec8b08354135b-ca-certs\") pod \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" (UID: \"f791c201e896ce34810ec8b08354135b\") " pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.259230 kubelet[2354]: I0912 18:08:01.258991 2354 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f791c201e896ce34810ec8b08354135b-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" (UID: \"f791c201e896ce34810ec8b08354135b\") " pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.261217 kubelet[2354]: E0912 18:08:01.261147 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://161.35.232.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4426.1.0-c-cd41a2aea4?timeout=10s\": dial tcp 161.35.232.141:6443: connect: connection refused" interval="400ms" Sep 12 18:08:01.455515 kubelet[2354]: I0912 18:08:01.455460 2354 kubelet_node_status.go:72] "Attempting to register node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.456143 kubelet[2354]: E0912 18:08:01.456092 2354 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://161.35.232.141:6443/api/v1/nodes\": dial tcp 161.35.232.141:6443: connect: connection refused" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.536517 kubelet[2354]: E0912 18:08:01.536353 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:01.537496 containerd[1562]: time="2025-09-12T18:08:01.537413536Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4426.1.0-c-cd41a2aea4,Uid:f791c201e896ce34810ec8b08354135b,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:01.548468 kubelet[2354]: E0912 18:08:01.548122 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:01.554983 containerd[1562]: time="2025-09-12T18:08:01.554915616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4426.1.0-c-cd41a2aea4,Uid:340546bd497b3f71404b0f6daba1be5c,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:01.555989 kubelet[2354]: E0912 18:08:01.555950 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:01.557536 containerd[1562]: time="2025-09-12T18:08:01.557492664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4426.1.0-c-cd41a2aea4,Uid:57f30eb7298f4b27fdb24c17107fca13,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:01.666070 kubelet[2354]: E0912 18:08:01.665992 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://161.35.232.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4426.1.0-c-cd41a2aea4?timeout=10s\": dial tcp 161.35.232.141:6443: connect: connection refused" interval="800ms" Sep 12 18:08:01.682002 containerd[1562]: time="2025-09-12T18:08:01.681933232Z" level=info msg="connecting to shim fe99c39584c9d34160cd937606bb520c6d305df2f9143f59917343cfa0bbc0ba" address="unix:///run/containerd/s/569736749cda54410fb5ecdd03d50c385c156f13fc4f394b4e1b6de12e4e15bf" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:01.685783 containerd[1562]: time="2025-09-12T18:08:01.685711736Z" level=info msg="connecting to shim 94660ab30f85179ac7c659be434f4affb843a43caf0992f1852c2ad78b74f451" address="unix:///run/containerd/s/21a684ae0450a0703453fe82c56ff79e61e00bc62ef1e1c59936eea0858f037f" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:01.694671 containerd[1562]: time="2025-09-12T18:08:01.694608836Z" level=info msg="connecting to shim 117341ba7e2fdfa3c0136601be7267e7e9fc9f78b0a07adc3c049a88b39c206f" address="unix:///run/containerd/s/319cc8bfb30cac887dbdd0d268da34d19d2b0f18f39039b92deaab73a47e2a98" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:01.846761 systemd[1]: Started cri-containerd-117341ba7e2fdfa3c0136601be7267e7e9fc9f78b0a07adc3c049a88b39c206f.scope - libcontainer container 117341ba7e2fdfa3c0136601be7267e7e9fc9f78b0a07adc3c049a88b39c206f. Sep 12 18:08:01.849005 systemd[1]: Started cri-containerd-fe99c39584c9d34160cd937606bb520c6d305df2f9143f59917343cfa0bbc0ba.scope - libcontainer container fe99c39584c9d34160cd937606bb520c6d305df2f9143f59917343cfa0bbc0ba. Sep 12 18:08:01.858530 systemd[1]: Started cri-containerd-94660ab30f85179ac7c659be434f4affb843a43caf0992f1852c2ad78b74f451.scope - libcontainer container 94660ab30f85179ac7c659be434f4affb843a43caf0992f1852c2ad78b74f451. Sep 12 18:08:01.864754 kubelet[2354]: I0912 18:08:01.864629 2354 kubelet_node_status.go:72] "Attempting to register node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.865247 kubelet[2354]: E0912 18:08:01.864981 2354 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://161.35.232.141:6443/api/v1/nodes\": dial tcp 161.35.232.141:6443: connect: connection refused" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:01.947994 containerd[1562]: time="2025-09-12T18:08:01.947945098Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4426.1.0-c-cd41a2aea4,Uid:f791c201e896ce34810ec8b08354135b,Namespace:kube-system,Attempt:0,} returns sandbox id \"fe99c39584c9d34160cd937606bb520c6d305df2f9143f59917343cfa0bbc0ba\"" Sep 12 18:08:01.950598 kubelet[2354]: E0912 18:08:01.950467 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:01.954436 containerd[1562]: time="2025-09-12T18:08:01.954366020Z" level=info msg="CreateContainer within sandbox \"fe99c39584c9d34160cd937606bb520c6d305df2f9143f59917343cfa0bbc0ba\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 12 18:08:01.976245 containerd[1562]: time="2025-09-12T18:08:01.976165868Z" level=info msg="Container 5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:01.987555 kubelet[2354]: W0912 18:08:01.987290 2354 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://161.35.232.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 161.35.232.141:6443: connect: connection refused Sep 12 18:08:01.987555 kubelet[2354]: E0912 18:08:01.987383 2354 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://161.35.232.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:01.993454 containerd[1562]: time="2025-09-12T18:08:01.993384624Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4426.1.0-c-cd41a2aea4,Uid:340546bd497b3f71404b0f6daba1be5c,Namespace:kube-system,Attempt:0,} returns sandbox id \"94660ab30f85179ac7c659be434f4affb843a43caf0992f1852c2ad78b74f451\"" Sep 12 18:08:01.999350 kubelet[2354]: E0912 18:08:01.999293 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:02.003117 containerd[1562]: time="2025-09-12T18:08:02.002957760Z" level=info msg="CreateContainer within sandbox \"fe99c39584c9d34160cd937606bb520c6d305df2f9143f59917343cfa0bbc0ba\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895\"" Sep 12 18:08:02.006715 containerd[1562]: time="2025-09-12T18:08:02.006464292Z" level=info msg="CreateContainer within sandbox \"94660ab30f85179ac7c659be434f4affb843a43caf0992f1852c2ad78b74f451\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 12 18:08:02.006715 containerd[1562]: time="2025-09-12T18:08:02.006629915Z" level=info msg="StartContainer for \"5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895\"" Sep 12 18:08:02.009164 containerd[1562]: time="2025-09-12T18:08:02.009108895Z" level=info msg="connecting to shim 5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895" address="unix:///run/containerd/s/569736749cda54410fb5ecdd03d50c385c156f13fc4f394b4e1b6de12e4e15bf" protocol=ttrpc version=3 Sep 12 18:08:02.013847 kubelet[2354]: W0912 18:08:02.013744 2354 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://161.35.232.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 161.35.232.141:6443: connect: connection refused Sep 12 18:08:02.013847 kubelet[2354]: E0912 18:08:02.013826 2354 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://161.35.232.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 161.35.232.141:6443: connect: connection refused" logger="UnhandledError" Sep 12 18:08:02.029444 containerd[1562]: time="2025-09-12T18:08:02.029112140Z" level=info msg="Container 57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:02.036182 containerd[1562]: time="2025-09-12T18:08:02.036129985Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4426.1.0-c-cd41a2aea4,Uid:57f30eb7298f4b27fdb24c17107fca13,Namespace:kube-system,Attempt:0,} returns sandbox id \"117341ba7e2fdfa3c0136601be7267e7e9fc9f78b0a07adc3c049a88b39c206f\"" Sep 12 18:08:02.038652 kubelet[2354]: E0912 18:08:02.038604 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:02.042864 containerd[1562]: time="2025-09-12T18:08:02.042793324Z" level=info msg="CreateContainer within sandbox \"117341ba7e2fdfa3c0136601be7267e7e9fc9f78b0a07adc3c049a88b39c206f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 12 18:08:02.048471 containerd[1562]: time="2025-09-12T18:08:02.048400985Z" level=info msg="CreateContainer within sandbox \"94660ab30f85179ac7c659be434f4affb843a43caf0992f1852c2ad78b74f451\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b\"" Sep 12 18:08:02.049851 containerd[1562]: time="2025-09-12T18:08:02.049652916Z" level=info msg="StartContainer for \"57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b\"" Sep 12 18:08:02.051035 containerd[1562]: time="2025-09-12T18:08:02.050991608Z" level=info msg="connecting to shim 57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b" address="unix:///run/containerd/s/21a684ae0450a0703453fe82c56ff79e61e00bc62ef1e1c59936eea0858f037f" protocol=ttrpc version=3 Sep 12 18:08:02.060915 systemd[1]: Started cri-containerd-5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895.scope - libcontainer container 5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895. Sep 12 18:08:02.075547 containerd[1562]: time="2025-09-12T18:08:02.075455098Z" level=info msg="Container 7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:02.087880 containerd[1562]: time="2025-09-12T18:08:02.087819177Z" level=info msg="CreateContainer within sandbox \"117341ba7e2fdfa3c0136601be7267e7e9fc9f78b0a07adc3c049a88b39c206f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455\"" Sep 12 18:08:02.091860 containerd[1562]: time="2025-09-12T18:08:02.091809083Z" level=info msg="StartContainer for \"7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455\"" Sep 12 18:08:02.094331 containerd[1562]: time="2025-09-12T18:08:02.094271867Z" level=info msg="connecting to shim 7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455" address="unix:///run/containerd/s/319cc8bfb30cac887dbdd0d268da34d19d2b0f18f39039b92deaab73a47e2a98" protocol=ttrpc version=3 Sep 12 18:08:02.109812 systemd[1]: Started cri-containerd-57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b.scope - libcontainer container 57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b. Sep 12 18:08:02.159918 systemd[1]: Started cri-containerd-7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455.scope - libcontainer container 7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455. Sep 12 18:08:02.258451 containerd[1562]: time="2025-09-12T18:08:02.258116106Z" level=info msg="StartContainer for \"5fd94ee4a6f83370f4fed4d917ef0515a4c54de169f484bcd1c16eb84fb46895\" returns successfully" Sep 12 18:08:02.285627 containerd[1562]: time="2025-09-12T18:08:02.285558610Z" level=info msg="StartContainer for \"57d3ef6d9fb8685b3b330bb91c4bb55dac4c3a86e31ec6122f29ffbd026f092b\" returns successfully" Sep 12 18:08:02.367591 containerd[1562]: time="2025-09-12T18:08:02.367332506Z" level=info msg="StartContainer for \"7c02a93ccafe5101d5ee968b5ddca94a6e01f94fe80b910b81fb57a812098455\" returns successfully" Sep 12 18:08:02.668930 kubelet[2354]: I0912 18:08:02.668776 2354 kubelet_node_status.go:72] "Attempting to register node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:03.151013 kubelet[2354]: E0912 18:08:03.150968 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:03.155603 kubelet[2354]: E0912 18:08:03.155564 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:03.164511 kubelet[2354]: E0912 18:08:03.164041 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:04.164320 kubelet[2354]: E0912 18:08:04.164275 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:04.165138 kubelet[2354]: E0912 18:08:04.165001 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:04.165340 kubelet[2354]: E0912 18:08:04.165322 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:04.778596 kubelet[2354]: I0912 18:08:04.778510 2354 kubelet_node_status.go:75] "Successfully registered node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:04.778596 kubelet[2354]: E0912 18:08:04.778554 2354 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4426.1.0-c-cd41a2aea4\": node \"ci-4426.1.0-c-cd41a2aea4\" not found" Sep 12 18:08:05.039094 kubelet[2354]: I0912 18:08:05.038954 2354 apiserver.go:52] "Watching apiserver" Sep 12 18:08:05.058318 kubelet[2354]: I0912 18:08:05.058229 2354 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 12 18:08:05.176080 kubelet[2354]: E0912 18:08:05.175745 2354 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:05.176080 kubelet[2354]: E0912 18:08:05.175975 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:07.146756 systemd[1]: Reload requested from client PID 2620 ('systemctl') (unit session-7.scope)... Sep 12 18:08:07.146774 systemd[1]: Reloading... Sep 12 18:08:07.258576 zram_generator::config[2662]: No configuration found. Sep 12 18:08:07.584829 systemd[1]: Reloading finished in 437 ms. Sep 12 18:08:07.620516 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 18:08:07.644869 systemd[1]: kubelet.service: Deactivated successfully. Sep 12 18:08:07.645518 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:08:07.645593 systemd[1]: kubelet.service: Consumed 845ms CPU time, 123.2M memory peak. Sep 12 18:08:07.648523 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 18:08:07.821707 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 18:08:07.832572 (kubelet)[2714]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 18:08:07.899956 kubelet[2714]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 18:08:07.899956 kubelet[2714]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 12 18:08:07.899956 kubelet[2714]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 18:08:07.899956 kubelet[2714]: I0912 18:08:07.899308 2714 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 18:08:07.907403 kubelet[2714]: I0912 18:08:07.907355 2714 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 12 18:08:07.907403 kubelet[2714]: I0912 18:08:07.907385 2714 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 18:08:07.907720 kubelet[2714]: I0912 18:08:07.907700 2714 server.go:934] "Client rotation is on, will bootstrap in background" Sep 12 18:08:07.909323 kubelet[2714]: I0912 18:08:07.909234 2714 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 12 18:08:07.911739 kubelet[2714]: I0912 18:08:07.911496 2714 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 18:08:07.923864 kubelet[2714]: I0912 18:08:07.923836 2714 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 12 18:08:07.929559 kubelet[2714]: I0912 18:08:07.928656 2714 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 18:08:07.929559 kubelet[2714]: I0912 18:08:07.928900 2714 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 12 18:08:07.929559 kubelet[2714]: I0912 18:08:07.929049 2714 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 18:08:07.929559 kubelet[2714]: I0912 18:08:07.929078 2714 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4426.1.0-c-cd41a2aea4","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929455 2714 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929466 2714 container_manager_linux.go:300] "Creating device plugin manager" Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929527 2714 state_mem.go:36] "Initialized new in-memory state store" Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929677 2714 kubelet.go:408] "Attempting to sync node with API server" Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929691 2714 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929760 2714 kubelet.go:314] "Adding apiserver pod source" Sep 12 18:08:07.929855 kubelet[2714]: I0912 18:08:07.929771 2714 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 18:08:07.935344 kubelet[2714]: I0912 18:08:07.934468 2714 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 12 18:08:07.936657 kubelet[2714]: I0912 18:08:07.936633 2714 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 12 18:08:07.944450 kubelet[2714]: I0912 18:08:07.944421 2714 server.go:1274] "Started kubelet" Sep 12 18:08:07.950520 kubelet[2714]: I0912 18:08:07.949600 2714 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 18:08:07.950520 kubelet[2714]: I0912 18:08:07.950126 2714 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 18:08:07.950520 kubelet[2714]: I0912 18:08:07.950234 2714 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 18:08:07.952504 kubelet[2714]: I0912 18:08:07.951392 2714 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 18:08:07.954666 kubelet[2714]: I0912 18:08:07.954634 2714 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 12 18:08:07.954786 kubelet[2714]: I0912 18:08:07.951415 2714 server.go:449] "Adding debug handlers to kubelet server" Sep 12 18:08:07.956036 kubelet[2714]: I0912 18:08:07.956010 2714 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 12 18:08:07.956036 kubelet[2714]: I0912 18:08:07.951921 2714 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 18:08:07.956234 kubelet[2714]: I0912 18:08:07.956204 2714 reconciler.go:26] "Reconciler: start to sync state" Sep 12 18:08:07.959532 kubelet[2714]: E0912 18:08:07.958758 2714 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4426.1.0-c-cd41a2aea4\" not found" Sep 12 18:08:07.960069 kubelet[2714]: I0912 18:08:07.960041 2714 factory.go:221] Registration of the systemd container factory successfully Sep 12 18:08:07.960224 kubelet[2714]: I0912 18:08:07.960200 2714 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 18:08:07.964085 kubelet[2714]: E0912 18:08:07.963859 2714 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 18:08:07.970513 kubelet[2714]: I0912 18:08:07.970195 2714 factory.go:221] Registration of the containerd container factory successfully Sep 12 18:08:07.988973 kubelet[2714]: I0912 18:08:07.988904 2714 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 12 18:08:07.991725 kubelet[2714]: I0912 18:08:07.991692 2714 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 12 18:08:07.992514 kubelet[2714]: I0912 18:08:07.992484 2714 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 12 18:08:07.992514 kubelet[2714]: I0912 18:08:07.992511 2714 kubelet.go:2321] "Starting kubelet main sync loop" Sep 12 18:08:07.993159 kubelet[2714]: E0912 18:08:07.993114 2714 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 18:08:08.041520 kubelet[2714]: I0912 18:08:08.041470 2714 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 12 18:08:08.041520 kubelet[2714]: I0912 18:08:08.041509 2714 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 12 18:08:08.041520 kubelet[2714]: I0912 18:08:08.041536 2714 state_mem.go:36] "Initialized new in-memory state store" Sep 12 18:08:08.041775 kubelet[2714]: I0912 18:08:08.041748 2714 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 12 18:08:08.041826 kubelet[2714]: I0912 18:08:08.041763 2714 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 12 18:08:08.041826 kubelet[2714]: I0912 18:08:08.041793 2714 policy_none.go:49] "None policy: Start" Sep 12 18:08:08.042838 kubelet[2714]: I0912 18:08:08.042784 2714 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 12 18:08:08.042838 kubelet[2714]: I0912 18:08:08.042816 2714 state_mem.go:35] "Initializing new in-memory state store" Sep 12 18:08:08.043039 kubelet[2714]: I0912 18:08:08.043021 2714 state_mem.go:75] "Updated machine memory state" Sep 12 18:08:08.048845 kubelet[2714]: I0912 18:08:08.048797 2714 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 12 18:08:08.049085 kubelet[2714]: I0912 18:08:08.049042 2714 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 18:08:08.049172 kubelet[2714]: I0912 18:08:08.049065 2714 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 18:08:08.049654 kubelet[2714]: I0912 18:08:08.049621 2714 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 18:08:08.102598 kubelet[2714]: W0912 18:08:08.102371 2714 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 18:08:08.104833 kubelet[2714]: W0912 18:08:08.104783 2714 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 18:08:08.105052 kubelet[2714]: W0912 18:08:08.105031 2714 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 18:08:08.152250 kubelet[2714]: I0912 18:08:08.152120 2714 kubelet_node_status.go:72] "Attempting to register node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.158263 kubelet[2714]: I0912 18:08:08.158176 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.158263 kubelet[2714]: I0912 18:08:08.158226 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f791c201e896ce34810ec8b08354135b-k8s-certs\") pod \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" (UID: \"f791c201e896ce34810ec8b08354135b\") " pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.158263 kubelet[2714]: I0912 18:08:08.158245 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-ca-certs\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.158263 kubelet[2714]: I0912 18:08:08.158260 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-flexvolume-dir\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.159299 kubelet[2714]: I0912 18:08:08.158284 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-kubeconfig\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.159299 kubelet[2714]: I0912 18:08:08.158300 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/57f30eb7298f4b27fdb24c17107fca13-kubeconfig\") pod \"kube-scheduler-ci-4426.1.0-c-cd41a2aea4\" (UID: \"57f30eb7298f4b27fdb24c17107fca13\") " pod="kube-system/kube-scheduler-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.159299 kubelet[2714]: I0912 18:08:08.158317 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f791c201e896ce34810ec8b08354135b-ca-certs\") pod \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" (UID: \"f791c201e896ce34810ec8b08354135b\") " pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.159299 kubelet[2714]: I0912 18:08:08.158335 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f791c201e896ce34810ec8b08354135b-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" (UID: \"f791c201e896ce34810ec8b08354135b\") " pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.159299 kubelet[2714]: I0912 18:08:08.158351 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/340546bd497b3f71404b0f6daba1be5c-k8s-certs\") pod \"kube-controller-manager-ci-4426.1.0-c-cd41a2aea4\" (UID: \"340546bd497b3f71404b0f6daba1be5c\") " pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.162693 kubelet[2714]: I0912 18:08:08.162609 2714 kubelet_node_status.go:111] "Node was previously registered" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.163092 kubelet[2714]: I0912 18:08:08.162805 2714 kubelet_node_status.go:75] "Successfully registered node" node="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:08.404850 kubelet[2714]: E0912 18:08:08.404705 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:08.406699 kubelet[2714]: E0912 18:08:08.406294 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:08.410135 kubelet[2714]: E0912 18:08:08.407698 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:08.942767 kubelet[2714]: I0912 18:08:08.942671 2714 apiserver.go:52] "Watching apiserver" Sep 12 18:08:08.957144 kubelet[2714]: I0912 18:08:08.957098 2714 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 12 18:08:09.020503 kubelet[2714]: E0912 18:08:09.020452 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:09.022652 kubelet[2714]: E0912 18:08:09.022617 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:09.038047 kubelet[2714]: W0912 18:08:09.037904 2714 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 18:08:09.038870 kubelet[2714]: E0912 18:08:09.038163 2714 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4426.1.0-c-cd41a2aea4\" already exists" pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:09.039306 kubelet[2714]: E0912 18:08:09.039229 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:09.065123 kubelet[2714]: I0912 18:08:09.065009 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4426.1.0-c-cd41a2aea4" podStartSLOduration=1.064986798 podStartE2EDuration="1.064986798s" podCreationTimestamp="2025-09-12 18:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 18:08:09.062469334 +0000 UTC m=+1.223329629" watchObservedRunningTime="2025-09-12 18:08:09.064986798 +0000 UTC m=+1.225847099" Sep 12 18:08:09.081909 kubelet[2714]: I0912 18:08:09.081834 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4426.1.0-c-cd41a2aea4" podStartSLOduration=1.081813325 podStartE2EDuration="1.081813325s" podCreationTimestamp="2025-09-12 18:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 18:08:09.081776429 +0000 UTC m=+1.242636740" watchObservedRunningTime="2025-09-12 18:08:09.081813325 +0000 UTC m=+1.242673623" Sep 12 18:08:10.022557 kubelet[2714]: E0912 18:08:10.022503 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:11.779926 kubelet[2714]: I0912 18:08:11.779764 2714 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 12 18:08:11.780652 containerd[1562]: time="2025-09-12T18:08:11.780558382Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 12 18:08:11.781122 kubelet[2714]: I0912 18:08:11.780863 2714 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 12 18:08:12.948499 kubelet[2714]: I0912 18:08:12.948382 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4426.1.0-c-cd41a2aea4" podStartSLOduration=4.948350734 podStartE2EDuration="4.948350734s" podCreationTimestamp="2025-09-12 18:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 18:08:09.096106054 +0000 UTC m=+1.256966352" watchObservedRunningTime="2025-09-12 18:08:12.948350734 +0000 UTC m=+5.109211032" Sep 12 18:08:12.965706 systemd[1]: Created slice kubepods-besteffort-podf2d0dfb4_c0dd_43ae_8604_26d833545e25.slice - libcontainer container kubepods-besteffort-podf2d0dfb4_c0dd_43ae_8604_26d833545e25.slice. Sep 12 18:08:12.989406 kubelet[2714]: I0912 18:08:12.989341 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f2d0dfb4-c0dd-43ae-8604-26d833545e25-xtables-lock\") pod \"kube-proxy-4jrt8\" (UID: \"f2d0dfb4-c0dd-43ae-8604-26d833545e25\") " pod="kube-system/kube-proxy-4jrt8" Sep 12 18:08:12.989406 kubelet[2714]: I0912 18:08:12.989406 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crldk\" (UniqueName: \"kubernetes.io/projected/f2d0dfb4-c0dd-43ae-8604-26d833545e25-kube-api-access-crldk\") pod \"kube-proxy-4jrt8\" (UID: \"f2d0dfb4-c0dd-43ae-8604-26d833545e25\") " pod="kube-system/kube-proxy-4jrt8" Sep 12 18:08:12.991778 kubelet[2714]: I0912 18:08:12.989434 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f2d0dfb4-c0dd-43ae-8604-26d833545e25-kube-proxy\") pod \"kube-proxy-4jrt8\" (UID: \"f2d0dfb4-c0dd-43ae-8604-26d833545e25\") " pod="kube-system/kube-proxy-4jrt8" Sep 12 18:08:12.991778 kubelet[2714]: I0912 18:08:12.989457 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f2d0dfb4-c0dd-43ae-8604-26d833545e25-lib-modules\") pod \"kube-proxy-4jrt8\" (UID: \"f2d0dfb4-c0dd-43ae-8604-26d833545e25\") " pod="kube-system/kube-proxy-4jrt8" Sep 12 18:08:13.129410 systemd[1]: Created slice kubepods-besteffort-pod0df45a92_b465_4b22_9915_89e2a3189fa8.slice - libcontainer container kubepods-besteffort-pod0df45a92_b465_4b22_9915_89e2a3189fa8.slice. Sep 12 18:08:13.191078 kubelet[2714]: I0912 18:08:13.190976 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/0df45a92-b465-4b22-9915-89e2a3189fa8-var-lib-calico\") pod \"tigera-operator-58fc44c59b-vmfjd\" (UID: \"0df45a92-b465-4b22-9915-89e2a3189fa8\") " pod="tigera-operator/tigera-operator-58fc44c59b-vmfjd" Sep 12 18:08:13.191078 kubelet[2714]: I0912 18:08:13.191052 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrk55\" (UniqueName: \"kubernetes.io/projected/0df45a92-b465-4b22-9915-89e2a3189fa8-kube-api-access-vrk55\") pod \"tigera-operator-58fc44c59b-vmfjd\" (UID: \"0df45a92-b465-4b22-9915-89e2a3189fa8\") " pod="tigera-operator/tigera-operator-58fc44c59b-vmfjd" Sep 12 18:08:13.279726 kubelet[2714]: E0912 18:08:13.279667 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:13.281142 containerd[1562]: time="2025-09-12T18:08:13.280848899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4jrt8,Uid:f2d0dfb4-c0dd-43ae-8604-26d833545e25,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:13.317839 containerd[1562]: time="2025-09-12T18:08:13.317723763Z" level=info msg="connecting to shim 6fb43dc0cd41d651f6b28281198084da0eacc3317c5b44849effa2a4a51c6879" address="unix:///run/containerd/s/2448bd226a861e50d36a2bcf54a37cc0f87f579c73aeed45c9586b3c18e2da43" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:13.361810 systemd[1]: Started cri-containerd-6fb43dc0cd41d651f6b28281198084da0eacc3317c5b44849effa2a4a51c6879.scope - libcontainer container 6fb43dc0cd41d651f6b28281198084da0eacc3317c5b44849effa2a4a51c6879. Sep 12 18:08:13.400646 containerd[1562]: time="2025-09-12T18:08:13.400539090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4jrt8,Uid:f2d0dfb4-c0dd-43ae-8604-26d833545e25,Namespace:kube-system,Attempt:0,} returns sandbox id \"6fb43dc0cd41d651f6b28281198084da0eacc3317c5b44849effa2a4a51c6879\"" Sep 12 18:08:13.402014 kubelet[2714]: E0912 18:08:13.401978 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:13.406553 containerd[1562]: time="2025-09-12T18:08:13.406447335Z" level=info msg="CreateContainer within sandbox \"6fb43dc0cd41d651f6b28281198084da0eacc3317c5b44849effa2a4a51c6879\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 12 18:08:13.421504 containerd[1562]: time="2025-09-12T18:08:13.421415175Z" level=info msg="Container ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:13.433155 containerd[1562]: time="2025-09-12T18:08:13.433021692Z" level=info msg="CreateContainer within sandbox \"6fb43dc0cd41d651f6b28281198084da0eacc3317c5b44849effa2a4a51c6879\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419\"" Sep 12 18:08:13.435513 containerd[1562]: time="2025-09-12T18:08:13.434097170Z" level=info msg="StartContainer for \"ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419\"" Sep 12 18:08:13.438758 containerd[1562]: time="2025-09-12T18:08:13.438714730Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-58fc44c59b-vmfjd,Uid:0df45a92-b465-4b22-9915-89e2a3189fa8,Namespace:tigera-operator,Attempt:0,}" Sep 12 18:08:13.439638 containerd[1562]: time="2025-09-12T18:08:13.439593055Z" level=info msg="connecting to shim ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419" address="unix:///run/containerd/s/2448bd226a861e50d36a2bcf54a37cc0f87f579c73aeed45c9586b3c18e2da43" protocol=ttrpc version=3 Sep 12 18:08:13.467829 containerd[1562]: time="2025-09-12T18:08:13.467762921Z" level=info msg="connecting to shim 0ec0dc2d60394607ad1ba1ba83072d7c2be5c394762cc2d4ea277a0be11b6505" address="unix:///run/containerd/s/264617e382dab6b23d067aa2a334c83e4db23efdae4b2f06213dc8e42e6e4e46" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:13.471972 systemd[1]: Started cri-containerd-ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419.scope - libcontainer container ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419. Sep 12 18:08:13.484928 kubelet[2714]: E0912 18:08:13.484416 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:13.532787 systemd[1]: Started cri-containerd-0ec0dc2d60394607ad1ba1ba83072d7c2be5c394762cc2d4ea277a0be11b6505.scope - libcontainer container 0ec0dc2d60394607ad1ba1ba83072d7c2be5c394762cc2d4ea277a0be11b6505. Sep 12 18:08:13.584513 containerd[1562]: time="2025-09-12T18:08:13.584067873Z" level=info msg="StartContainer for \"ebc8468f160bfc82927b0cb3cd31b36205d39603201803a35d81d7b7d014c419\" returns successfully" Sep 12 18:08:13.628846 containerd[1562]: time="2025-09-12T18:08:13.628782085Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-58fc44c59b-vmfjd,Uid:0df45a92-b465-4b22-9915-89e2a3189fa8,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"0ec0dc2d60394607ad1ba1ba83072d7c2be5c394762cc2d4ea277a0be11b6505\"" Sep 12 18:08:13.632672 containerd[1562]: time="2025-09-12T18:08:13.632580627Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 12 18:08:13.637275 systemd-resolved[1404]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Sep 12 18:08:14.039402 kubelet[2714]: E0912 18:08:14.039331 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:14.040833 kubelet[2714]: E0912 18:08:14.040223 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:14.070959 kubelet[2714]: I0912 18:08:14.070887 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-4jrt8" podStartSLOduration=2.070862385 podStartE2EDuration="2.070862385s" podCreationTimestamp="2025-09-12 18:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 18:08:14.070679509 +0000 UTC m=+6.231539806" watchObservedRunningTime="2025-09-12 18:08:14.070862385 +0000 UTC m=+6.231722685" Sep 12 18:08:14.935429 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4174543212.mount: Deactivated successfully. Sep 12 18:08:15.517576 systemd-timesyncd[1421]: Contacted time server 51.81.226.229:123 (2.flatcar.pool.ntp.org). Sep 12 18:08:15.518861 systemd-timesyncd[1421]: Initial clock synchronization to Fri 2025-09-12 18:08:15.791809 UTC. Sep 12 18:08:15.628876 kubelet[2714]: E0912 18:08:15.628752 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:15.785838 containerd[1562]: time="2025-09-12T18:08:15.785040852Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:15.787048 containerd[1562]: time="2025-09-12T18:08:15.787006639Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Sep 12 18:08:15.787159 containerd[1562]: time="2025-09-12T18:08:15.787134184Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:15.789225 containerd[1562]: time="2025-09-12T18:08:15.789190137Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:15.790263 containerd[1562]: time="2025-09-12T18:08:15.790223593Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 2.157369657s" Sep 12 18:08:15.790263 containerd[1562]: time="2025-09-12T18:08:15.790260530Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Sep 12 18:08:15.794531 containerd[1562]: time="2025-09-12T18:08:15.793229839Z" level=info msg="CreateContainer within sandbox \"0ec0dc2d60394607ad1ba1ba83072d7c2be5c394762cc2d4ea277a0be11b6505\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 12 18:08:15.802994 containerd[1562]: time="2025-09-12T18:08:15.802936125Z" level=info msg="Container c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:15.810601 containerd[1562]: time="2025-09-12T18:08:15.810551817Z" level=info msg="CreateContainer within sandbox \"0ec0dc2d60394607ad1ba1ba83072d7c2be5c394762cc2d4ea277a0be11b6505\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce\"" Sep 12 18:08:15.811535 containerd[1562]: time="2025-09-12T18:08:15.811424712Z" level=info msg="StartContainer for \"c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce\"" Sep 12 18:08:15.814046 containerd[1562]: time="2025-09-12T18:08:15.814002921Z" level=info msg="connecting to shim c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce" address="unix:///run/containerd/s/264617e382dab6b23d067aa2a334c83e4db23efdae4b2f06213dc8e42e6e4e46" protocol=ttrpc version=3 Sep 12 18:08:15.848782 systemd[1]: Started cri-containerd-c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce.scope - libcontainer container c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce. Sep 12 18:08:15.894606 containerd[1562]: time="2025-09-12T18:08:15.894461826Z" level=info msg="StartContainer for \"c63aa014f9e0b9d6e1c080d55fc75277f7fbde683866e7ea336cb440105b03ce\" returns successfully" Sep 12 18:08:16.047928 kubelet[2714]: E0912 18:08:16.047223 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:16.067276 kubelet[2714]: I0912 18:08:16.067044 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-58fc44c59b-vmfjd" podStartSLOduration=0.906932403 podStartE2EDuration="3.067020454s" podCreationTimestamp="2025-09-12 18:08:13 +0000 UTC" firstStartedPulling="2025-09-12 18:08:13.631314157 +0000 UTC m=+5.792174440" lastFinishedPulling="2025-09-12 18:08:15.791402195 +0000 UTC m=+7.952262491" observedRunningTime="2025-09-12 18:08:16.066734047 +0000 UTC m=+8.227594345" watchObservedRunningTime="2025-09-12 18:08:16.067020454 +0000 UTC m=+8.227880754" Sep 12 18:08:17.322310 kubelet[2714]: E0912 18:08:17.322271 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:18.061504 kubelet[2714]: E0912 18:08:18.061444 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:22.447035 sudo[1796]: pam_unix(sudo:session): session closed for user root Sep 12 18:08:22.455449 sshd[1795]: Connection closed by 139.178.89.65 port 55096 Sep 12 18:08:22.458172 sshd-session[1792]: pam_unix(sshd:session): session closed for user core Sep 12 18:08:22.463181 systemd[1]: sshd@7-161.35.232.141:22-139.178.89.65:55096.service: Deactivated successfully. Sep 12 18:08:22.463637 systemd-logind[1528]: Session 7 logged out. Waiting for processes to exit. Sep 12 18:08:22.467794 systemd[1]: session-7.scope: Deactivated successfully. Sep 12 18:08:22.468121 systemd[1]: session-7.scope: Consumed 5.016s CPU time, 160.3M memory peak. Sep 12 18:08:22.474371 systemd-logind[1528]: Removed session 7. Sep 12 18:08:24.277337 update_engine[1530]: I20250912 18:08:24.276542 1530 update_attempter.cc:509] Updating boot flags... Sep 12 18:08:28.229003 systemd[1]: Created slice kubepods-besteffort-pod0d0fbec7_8c3a_401f_a94b_ccaac92acb8c.slice - libcontainer container kubepods-besteffort-pod0d0fbec7_8c3a_401f_a94b_ccaac92acb8c.slice. Sep 12 18:08:28.403182 kubelet[2714]: I0912 18:08:28.403118 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d0fbec7-8c3a-401f-a94b-ccaac92acb8c-tigera-ca-bundle\") pod \"calico-typha-5998cc644d-5ztgm\" (UID: \"0d0fbec7-8c3a-401f-a94b-ccaac92acb8c\") " pod="calico-system/calico-typha-5998cc644d-5ztgm" Sep 12 18:08:28.404115 kubelet[2714]: I0912 18:08:28.403437 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/0d0fbec7-8c3a-401f-a94b-ccaac92acb8c-typha-certs\") pod \"calico-typha-5998cc644d-5ztgm\" (UID: \"0d0fbec7-8c3a-401f-a94b-ccaac92acb8c\") " pod="calico-system/calico-typha-5998cc644d-5ztgm" Sep 12 18:08:28.404115 kubelet[2714]: I0912 18:08:28.403560 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ht2p\" (UniqueName: \"kubernetes.io/projected/0d0fbec7-8c3a-401f-a94b-ccaac92acb8c-kube-api-access-2ht2p\") pod \"calico-typha-5998cc644d-5ztgm\" (UID: \"0d0fbec7-8c3a-401f-a94b-ccaac92acb8c\") " pod="calico-system/calico-typha-5998cc644d-5ztgm" Sep 12 18:08:28.539074 kubelet[2714]: E0912 18:08:28.539021 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:28.540786 containerd[1562]: time="2025-09-12T18:08:28.540734733Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5998cc644d-5ztgm,Uid:0d0fbec7-8c3a-401f-a94b-ccaac92acb8c,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:28.578798 containerd[1562]: time="2025-09-12T18:08:28.578735883Z" level=info msg="connecting to shim 4b1a3e37e740e5f49a75204ac1436587ee6f85da38d3a88124e0c82155441aba" address="unix:///run/containerd/s/c634afc19a6adc25fa56abda295b7d5e6cdfa290e64497177ebfa9ca5e06004a" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:28.607587 kubelet[2714]: I0912 18:08:28.607542 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-var-lib-calico\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.617038 kubelet[2714]: I0912 18:08:28.616091 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44d766c7-de74-44eb-98c3-f19d735b7867-tigera-ca-bundle\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.617038 kubelet[2714]: I0912 18:08:28.616834 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-cni-log-dir\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.617038 kubelet[2714]: I0912 18:08:28.616864 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-lib-modules\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.617038 kubelet[2714]: I0912 18:08:28.616881 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49jfx\" (UniqueName: \"kubernetes.io/projected/44d766c7-de74-44eb-98c3-f19d735b7867-kube-api-access-49jfx\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.617614 kubelet[2714]: I0912 18:08:28.617551 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-cni-bin-dir\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.618056 systemd[1]: Created slice kubepods-besteffort-pod44d766c7_de74_44eb_98c3_f19d735b7867.slice - libcontainer container kubepods-besteffort-pod44d766c7_de74_44eb_98c3_f19d735b7867.slice. Sep 12 18:08:28.619864 kubelet[2714]: I0912 18:08:28.618728 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-flexvol-driver-host\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.619864 kubelet[2714]: I0912 18:08:28.619325 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/44d766c7-de74-44eb-98c3-f19d735b7867-node-certs\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.620627 kubelet[2714]: I0912 18:08:28.620064 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-cni-net-dir\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.620895 kubelet[2714]: I0912 18:08:28.620105 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-policysync\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.620895 kubelet[2714]: I0912 18:08:28.620792 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-var-run-calico\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.622404 kubelet[2714]: I0912 18:08:28.621981 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/44d766c7-de74-44eb-98c3-f19d735b7867-xtables-lock\") pod \"calico-node-m2tpt\" (UID: \"44d766c7-de74-44eb-98c3-f19d735b7867\") " pod="calico-system/calico-node-m2tpt" Sep 12 18:08:28.661816 systemd[1]: Started cri-containerd-4b1a3e37e740e5f49a75204ac1436587ee6f85da38d3a88124e0c82155441aba.scope - libcontainer container 4b1a3e37e740e5f49a75204ac1436587ee6f85da38d3a88124e0c82155441aba. Sep 12 18:08:28.728816 kubelet[2714]: E0912 18:08:28.728418 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.728816 kubelet[2714]: W0912 18:08:28.728677 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.731825 kubelet[2714]: E0912 18:08:28.731784 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.732113 kubelet[2714]: E0912 18:08:28.732083 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.732113 kubelet[2714]: W0912 18:08:28.732108 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.732197 kubelet[2714]: E0912 18:08:28.732130 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.746829 kubelet[2714]: E0912 18:08:28.746748 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.746829 kubelet[2714]: W0912 18:08:28.746776 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.746829 kubelet[2714]: E0912 18:08:28.746801 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.791513 containerd[1562]: time="2025-09-12T18:08:28.791046653Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5998cc644d-5ztgm,Uid:0d0fbec7-8c3a-401f-a94b-ccaac92acb8c,Namespace:calico-system,Attempt:0,} returns sandbox id \"4b1a3e37e740e5f49a75204ac1436587ee6f85da38d3a88124e0c82155441aba\"" Sep 12 18:08:28.796114 kubelet[2714]: E0912 18:08:28.795897 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:28.797369 containerd[1562]: time="2025-09-12T18:08:28.797325821Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 12 18:08:28.847392 kubelet[2714]: E0912 18:08:28.846847 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:28.910997 kubelet[2714]: E0912 18:08:28.910843 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.911597 kubelet[2714]: W0912 18:08:28.911453 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.912040 kubelet[2714]: E0912 18:08:28.911878 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.912722 kubelet[2714]: E0912 18:08:28.912686 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.912899 kubelet[2714]: W0912 18:08:28.912798 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.912899 kubelet[2714]: E0912 18:08:28.912829 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.913367 kubelet[2714]: E0912 18:08:28.913269 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.913367 kubelet[2714]: W0912 18:08:28.913291 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.913367 kubelet[2714]: E0912 18:08:28.913309 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.913786 kubelet[2714]: E0912 18:08:28.913774 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.913909 kubelet[2714]: W0912 18:08:28.913837 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.913909 kubelet[2714]: E0912 18:08:28.913854 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.914572 kubelet[2714]: E0912 18:08:28.914382 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.914572 kubelet[2714]: W0912 18:08:28.914409 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.914572 kubelet[2714]: E0912 18:08:28.914424 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.915153 kubelet[2714]: E0912 18:08:28.914968 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.915153 kubelet[2714]: W0912 18:08:28.914993 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.915153 kubelet[2714]: E0912 18:08:28.915012 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.915906 kubelet[2714]: E0912 18:08:28.915797 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.915906 kubelet[2714]: W0912 18:08:28.915811 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.916221 kubelet[2714]: E0912 18:08:28.915979 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.916801 kubelet[2714]: E0912 18:08:28.916670 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.916801 kubelet[2714]: W0912 18:08:28.916699 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.916801 kubelet[2714]: E0912 18:08:28.916712 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.917469 kubelet[2714]: E0912 18:08:28.917399 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.917469 kubelet[2714]: W0912 18:08:28.917412 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.917469 kubelet[2714]: E0912 18:08:28.917424 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.917810 kubelet[2714]: E0912 18:08:28.917791 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.917869 kubelet[2714]: W0912 18:08:28.917810 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.917869 kubelet[2714]: E0912 18:08:28.917825 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.918377 kubelet[2714]: E0912 18:08:28.918359 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.918377 kubelet[2714]: W0912 18:08:28.918375 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.918451 kubelet[2714]: E0912 18:08:28.918388 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.918611 kubelet[2714]: E0912 18:08:28.918583 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.918611 kubelet[2714]: W0912 18:08:28.918593 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.918611 kubelet[2714]: E0912 18:08:28.918602 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.918819 kubelet[2714]: E0912 18:08:28.918803 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.918819 kubelet[2714]: W0912 18:08:28.918818 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.918912 kubelet[2714]: E0912 18:08:28.918831 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.919194 kubelet[2714]: E0912 18:08:28.919172 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.919194 kubelet[2714]: W0912 18:08:28.919191 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.919308 kubelet[2714]: E0912 18:08:28.919205 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.919633 kubelet[2714]: E0912 18:08:28.919616 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.919674 kubelet[2714]: W0912 18:08:28.919634 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.919674 kubelet[2714]: E0912 18:08:28.919649 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.920223 kubelet[2714]: E0912 18:08:28.920203 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.920223 kubelet[2714]: W0912 18:08:28.920219 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.920345 kubelet[2714]: E0912 18:08:28.920234 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.921595 kubelet[2714]: E0912 18:08:28.921574 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.921595 kubelet[2714]: W0912 18:08:28.921591 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.921687 kubelet[2714]: E0912 18:08:28.921604 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.921786 kubelet[2714]: E0912 18:08:28.921774 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.921786 kubelet[2714]: W0912 18:08:28.921785 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.921991 kubelet[2714]: E0912 18:08:28.921793 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.922023 kubelet[2714]: E0912 18:08:28.922004 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.922023 kubelet[2714]: W0912 18:08:28.922016 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.922072 kubelet[2714]: E0912 18:08:28.922030 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.922261 kubelet[2714]: E0912 18:08:28.922246 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.922319 kubelet[2714]: W0912 18:08:28.922261 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.922319 kubelet[2714]: E0912 18:08:28.922275 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.923665 kubelet[2714]: E0912 18:08:28.923643 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.923747 kubelet[2714]: W0912 18:08:28.923677 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.923747 kubelet[2714]: E0912 18:08:28.923694 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.923747 kubelet[2714]: I0912 18:08:28.923724 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/b51a8f2d-d1c4-4004-80da-1576b8879e25-varrun\") pod \"csi-node-driver-8hj6c\" (UID: \"b51a8f2d-d1c4-4004-80da-1576b8879e25\") " pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:28.923962 kubelet[2714]: E0912 18:08:28.923948 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.924004 kubelet[2714]: W0912 18:08:28.923966 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.924004 kubelet[2714]: E0912 18:08:28.923992 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.924050 kubelet[2714]: I0912 18:08:28.924011 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b51a8f2d-d1c4-4004-80da-1576b8879e25-kubelet-dir\") pod \"csi-node-driver-8hj6c\" (UID: \"b51a8f2d-d1c4-4004-80da-1576b8879e25\") " pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:28.924257 kubelet[2714]: E0912 18:08:28.924242 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.924257 kubelet[2714]: W0912 18:08:28.924256 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.924326 kubelet[2714]: E0912 18:08:28.924270 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.924326 kubelet[2714]: I0912 18:08:28.924292 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b51a8f2d-d1c4-4004-80da-1576b8879e25-registration-dir\") pod \"csi-node-driver-8hj6c\" (UID: \"b51a8f2d-d1c4-4004-80da-1576b8879e25\") " pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:28.924559 kubelet[2714]: E0912 18:08:28.924541 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.924559 kubelet[2714]: W0912 18:08:28.924556 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.924671 kubelet[2714]: E0912 18:08:28.924581 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.924671 kubelet[2714]: I0912 18:08:28.924598 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b51a8f2d-d1c4-4004-80da-1576b8879e25-socket-dir\") pod \"csi-node-driver-8hj6c\" (UID: \"b51a8f2d-d1c4-4004-80da-1576b8879e25\") " pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:28.925660 kubelet[2714]: E0912 18:08:28.925631 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.925660 kubelet[2714]: W0912 18:08:28.925655 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.925765 kubelet[2714]: E0912 18:08:28.925673 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.925765 kubelet[2714]: I0912 18:08:28.925692 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nq4t\" (UniqueName: \"kubernetes.io/projected/b51a8f2d-d1c4-4004-80da-1576b8879e25-kube-api-access-5nq4t\") pod \"csi-node-driver-8hj6c\" (UID: \"b51a8f2d-d1c4-4004-80da-1576b8879e25\") " pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:28.925956 kubelet[2714]: E0912 18:08:28.925913 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.925956 kubelet[2714]: W0912 18:08:28.925923 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.926133 kubelet[2714]: E0912 18:08:28.925958 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.926171 kubelet[2714]: E0912 18:08:28.926144 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.926171 kubelet[2714]: W0912 18:08:28.926153 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.926247 kubelet[2714]: E0912 18:08:28.926232 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.927416 kubelet[2714]: E0912 18:08:28.927368 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.927416 kubelet[2714]: W0912 18:08:28.927388 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.927783 kubelet[2714]: E0912 18:08:28.927725 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.927783 kubelet[2714]: W0912 18:08:28.927736 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.928284 kubelet[2714]: E0912 18:08:28.927960 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.928284 kubelet[2714]: W0912 18:08:28.927975 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.928284 kubelet[2714]: E0912 18:08:28.927987 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.928284 kubelet[2714]: E0912 18:08:28.928009 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.928284 kubelet[2714]: E0912 18:08:28.928021 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.928646 kubelet[2714]: E0912 18:08:28.928271 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.928809 kubelet[2714]: W0912 18:08:28.928764 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.928972 kubelet[2714]: E0912 18:08:28.928881 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.929615 kubelet[2714]: E0912 18:08:28.929579 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.929615 kubelet[2714]: W0912 18:08:28.929611 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.929710 kubelet[2714]: E0912 18:08:28.929628 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.930682 kubelet[2714]: E0912 18:08:28.930286 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.930682 kubelet[2714]: W0912 18:08:28.930300 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.930682 kubelet[2714]: E0912 18:08:28.930313 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.931505 kubelet[2714]: E0912 18:08:28.931149 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.931579 containerd[1562]: time="2025-09-12T18:08:28.931262090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-m2tpt,Uid:44d766c7-de74-44eb-98c3-f19d735b7867,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:28.931672 kubelet[2714]: W0912 18:08:28.931654 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.931796 kubelet[2714]: E0912 18:08:28.931721 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.932103 kubelet[2714]: E0912 18:08:28.932088 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:28.932542 kubelet[2714]: W0912 18:08:28.932522 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:28.932643 kubelet[2714]: E0912 18:08:28.932627 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:28.965874 containerd[1562]: time="2025-09-12T18:08:28.965351326Z" level=info msg="connecting to shim e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df" address="unix:///run/containerd/s/850e0fd976ad36998fc0f8caa82b4a266ea028bc9bf9013ccc6c1ba4a65a8745" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:29.017758 systemd[1]: Started cri-containerd-e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df.scope - libcontainer container e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df. Sep 12 18:08:29.027299 kubelet[2714]: E0912 18:08:29.027258 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.027576 kubelet[2714]: W0912 18:08:29.027512 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.027921 kubelet[2714]: E0912 18:08:29.027834 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.029148 kubelet[2714]: E0912 18:08:29.029124 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.029267 kubelet[2714]: W0912 18:08:29.029253 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.029356 kubelet[2714]: E0912 18:08:29.029339 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.029895 kubelet[2714]: E0912 18:08:29.029871 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.029895 kubelet[2714]: W0912 18:08:29.029888 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.030190 kubelet[2714]: E0912 18:08:29.029914 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.030388 kubelet[2714]: E0912 18:08:29.030347 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.030388 kubelet[2714]: W0912 18:08:29.030368 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.030748 kubelet[2714]: E0912 18:08:29.030392 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.030748 kubelet[2714]: E0912 18:08:29.030681 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.030748 kubelet[2714]: W0912 18:08:29.030696 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.030748 kubelet[2714]: E0912 18:08:29.030721 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.031359 kubelet[2714]: E0912 18:08:29.031307 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.031359 kubelet[2714]: W0912 18:08:29.031328 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.031868 kubelet[2714]: E0912 18:08:29.031420 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.032132 kubelet[2714]: E0912 18:08:29.032110 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.032132 kubelet[2714]: W0912 18:08:29.032126 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.032442 kubelet[2714]: E0912 18:08:29.032193 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.032994 kubelet[2714]: E0912 18:08:29.032977 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.032994 kubelet[2714]: W0912 18:08:29.032992 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.033449 kubelet[2714]: E0912 18:08:29.033015 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.033790 kubelet[2714]: E0912 18:08:29.033719 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.033790 kubelet[2714]: W0912 18:08:29.033737 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.033790 kubelet[2714]: E0912 18:08:29.033758 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.035677 kubelet[2714]: E0912 18:08:29.035647 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.035677 kubelet[2714]: W0912 18:08:29.035672 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.035842 kubelet[2714]: E0912 18:08:29.035750 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.036008 kubelet[2714]: E0912 18:08:29.035984 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.036008 kubelet[2714]: W0912 18:08:29.036000 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.036008 kubelet[2714]: E0912 18:08:29.036027 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.036889 kubelet[2714]: E0912 18:08:29.036866 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.036889 kubelet[2714]: W0912 18:08:29.036884 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.037203 kubelet[2714]: E0912 18:08:29.037031 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.037564 kubelet[2714]: E0912 18:08:29.037540 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.037564 kubelet[2714]: W0912 18:08:29.037558 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.037850 kubelet[2714]: E0912 18:08:29.037587 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.038847 kubelet[2714]: E0912 18:08:29.038821 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.038847 kubelet[2714]: W0912 18:08:29.038842 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.038968 kubelet[2714]: E0912 18:08:29.038868 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.039415 kubelet[2714]: E0912 18:08:29.039399 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.039415 kubelet[2714]: W0912 18:08:29.039413 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.039671 kubelet[2714]: E0912 18:08:29.039444 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.039857 kubelet[2714]: E0912 18:08:29.039803 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.039857 kubelet[2714]: W0912 18:08:29.039818 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.040293 kubelet[2714]: E0912 18:08:29.040051 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.040293 kubelet[2714]: E0912 18:08:29.040129 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.040293 kubelet[2714]: W0912 18:08:29.040282 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.040909 kubelet[2714]: E0912 18:08:29.040393 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.040909 kubelet[2714]: E0912 18:08:29.040725 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.040909 kubelet[2714]: W0912 18:08:29.040737 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.040909 kubelet[2714]: E0912 18:08:29.040768 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.041401 kubelet[2714]: E0912 18:08:29.040953 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.041401 kubelet[2714]: W0912 18:08:29.040962 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.041401 kubelet[2714]: E0912 18:08:29.040975 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.043319 kubelet[2714]: E0912 18:08:29.041586 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.043319 kubelet[2714]: W0912 18:08:29.041599 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.043319 kubelet[2714]: E0912 18:08:29.041635 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.043319 kubelet[2714]: E0912 18:08:29.041993 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.043319 kubelet[2714]: W0912 18:08:29.042003 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.043319 kubelet[2714]: E0912 18:08:29.042075 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.043319 kubelet[2714]: E0912 18:08:29.042293 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.043319 kubelet[2714]: W0912 18:08:29.042303 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.043319 kubelet[2714]: E0912 18:08:29.042320 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.045077 kubelet[2714]: E0912 18:08:29.044612 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.045077 kubelet[2714]: W0912 18:08:29.044636 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.045418 kubelet[2714]: E0912 18:08:29.045342 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.045418 kubelet[2714]: E0912 18:08:29.045391 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.046149 kubelet[2714]: W0912 18:08:29.045581 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.046149 kubelet[2714]: E0912 18:08:29.045940 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.048251 kubelet[2714]: E0912 18:08:29.048215 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.048532 kubelet[2714]: W0912 18:08:29.048300 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.049622 kubelet[2714]: E0912 18:08:29.048651 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.070741 kubelet[2714]: E0912 18:08:29.070656 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:29.070741 kubelet[2714]: W0912 18:08:29.070685 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:29.070741 kubelet[2714]: E0912 18:08:29.070729 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:29.098815 containerd[1562]: time="2025-09-12T18:08:29.098466402Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-m2tpt,Uid:44d766c7-de74-44eb-98c3-f19d735b7867,Namespace:calico-system,Attempt:0,} returns sandbox id \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\"" Sep 12 18:08:29.993752 kubelet[2714]: E0912 18:08:29.993240 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:30.823189 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2147315684.mount: Deactivated successfully. Sep 12 18:08:31.993905 kubelet[2714]: E0912 18:08:31.993835 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:32.333548 containerd[1562]: time="2025-09-12T18:08:32.333415155Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:32.334600 containerd[1562]: time="2025-09-12T18:08:32.334453666Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Sep 12 18:08:32.335518 containerd[1562]: time="2025-09-12T18:08:32.335301654Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:32.338066 containerd[1562]: time="2025-09-12T18:08:32.337972336Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:32.339529 containerd[1562]: time="2025-09-12T18:08:32.338904218Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 3.541515416s" Sep 12 18:08:32.339529 containerd[1562]: time="2025-09-12T18:08:32.338956432Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Sep 12 18:08:32.341547 containerd[1562]: time="2025-09-12T18:08:32.341299875Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 12 18:08:32.370715 containerd[1562]: time="2025-09-12T18:08:32.369341657Z" level=info msg="CreateContainer within sandbox \"4b1a3e37e740e5f49a75204ac1436587ee6f85da38d3a88124e0c82155441aba\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 12 18:08:32.378942 containerd[1562]: time="2025-09-12T18:08:32.378888996Z" level=info msg="Container 3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:32.386952 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1324182656.mount: Deactivated successfully. Sep 12 18:08:32.395594 containerd[1562]: time="2025-09-12T18:08:32.395535886Z" level=info msg="CreateContainer within sandbox \"4b1a3e37e740e5f49a75204ac1436587ee6f85da38d3a88124e0c82155441aba\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff\"" Sep 12 18:08:32.398861 containerd[1562]: time="2025-09-12T18:08:32.396444795Z" level=info msg="StartContainer for \"3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff\"" Sep 12 18:08:32.399421 containerd[1562]: time="2025-09-12T18:08:32.399277571Z" level=info msg="connecting to shim 3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff" address="unix:///run/containerd/s/c634afc19a6adc25fa56abda295b7d5e6cdfa290e64497177ebfa9ca5e06004a" protocol=ttrpc version=3 Sep 12 18:08:32.435846 systemd[1]: Started cri-containerd-3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff.scope - libcontainer container 3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff. Sep 12 18:08:32.525022 containerd[1562]: time="2025-09-12T18:08:32.524654538Z" level=info msg="StartContainer for \"3804138d07b1eca316a49882cb89d1c8a8aa4de2ca1a9d878cf1f90de85c56ff\" returns successfully" Sep 12 18:08:33.106452 kubelet[2714]: E0912 18:08:33.106118 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:33.151550 kubelet[2714]: E0912 18:08:33.151467 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.151550 kubelet[2714]: W0912 18:08:33.151528 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.151550 kubelet[2714]: E0912 18:08:33.151558 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.151935 kubelet[2714]: E0912 18:08:33.151895 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.151935 kubelet[2714]: W0912 18:08:33.151914 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.152083 kubelet[2714]: E0912 18:08:33.151937 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.152225 kubelet[2714]: E0912 18:08:33.152210 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.152225 kubelet[2714]: W0912 18:08:33.152223 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.152348 kubelet[2714]: E0912 18:08:33.152238 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.152426 kubelet[2714]: E0912 18:08:33.152405 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.152426 kubelet[2714]: W0912 18:08:33.152413 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.152426 kubelet[2714]: E0912 18:08:33.152422 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.152715 kubelet[2714]: E0912 18:08:33.152698 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.152715 kubelet[2714]: W0912 18:08:33.152714 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.152878 kubelet[2714]: E0912 18:08:33.152730 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.152959 kubelet[2714]: E0912 18:08:33.152930 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.152959 kubelet[2714]: W0912 18:08:33.152942 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.152959 kubelet[2714]: E0912 18:08:33.152954 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.153172 kubelet[2714]: E0912 18:08:33.153097 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.153172 kubelet[2714]: W0912 18:08:33.153104 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.153172 kubelet[2714]: E0912 18:08:33.153112 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.153363 kubelet[2714]: E0912 18:08:33.153243 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.153363 kubelet[2714]: W0912 18:08:33.153282 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.153363 kubelet[2714]: E0912 18:08:33.153295 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.153590 kubelet[2714]: E0912 18:08:33.153461 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.153590 kubelet[2714]: W0912 18:08:33.153468 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.153590 kubelet[2714]: E0912 18:08:33.153491 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.153782 kubelet[2714]: E0912 18:08:33.153644 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.153782 kubelet[2714]: W0912 18:08:33.153654 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.153782 kubelet[2714]: E0912 18:08:33.153666 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.154011 kubelet[2714]: E0912 18:08:33.153839 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.154011 kubelet[2714]: W0912 18:08:33.153847 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.154011 kubelet[2714]: E0912 18:08:33.153856 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.154011 kubelet[2714]: E0912 18:08:33.153987 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.154011 kubelet[2714]: W0912 18:08:33.153993 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.154011 kubelet[2714]: E0912 18:08:33.154001 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.154393 kubelet[2714]: E0912 18:08:33.154132 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.154393 kubelet[2714]: W0912 18:08:33.154138 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.154393 kubelet[2714]: E0912 18:08:33.154147 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.154393 kubelet[2714]: E0912 18:08:33.154281 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.154393 kubelet[2714]: W0912 18:08:33.154287 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.154393 kubelet[2714]: E0912 18:08:33.154295 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.154695 kubelet[2714]: E0912 18:08:33.154415 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.154695 kubelet[2714]: W0912 18:08:33.154426 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.154695 kubelet[2714]: E0912 18:08:33.154436 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.170135 kubelet[2714]: E0912 18:08:33.170071 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.170135 kubelet[2714]: W0912 18:08:33.170103 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.170857 kubelet[2714]: E0912 18:08:33.170507 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.171328 kubelet[2714]: E0912 18:08:33.171265 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.171328 kubelet[2714]: W0912 18:08:33.171303 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.171927 kubelet[2714]: E0912 18:08:33.171561 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.172197 kubelet[2714]: E0912 18:08:33.172090 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.172197 kubelet[2714]: W0912 18:08:33.172108 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.172197 kubelet[2714]: E0912 18:08:33.172137 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.172826 kubelet[2714]: E0912 18:08:33.172792 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.173291 kubelet[2714]: W0912 18:08:33.173180 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.173291 kubelet[2714]: E0912 18:08:33.173220 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.173844 kubelet[2714]: E0912 18:08:33.173823 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.174110 kubelet[2714]: W0912 18:08:33.173964 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.174110 kubelet[2714]: E0912 18:08:33.174010 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.174591 kubelet[2714]: E0912 18:08:33.174373 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.174591 kubelet[2714]: W0912 18:08:33.174396 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.174591 kubelet[2714]: E0912 18:08:33.174461 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.175220 kubelet[2714]: E0912 18:08:33.174750 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.175220 kubelet[2714]: W0912 18:08:33.174966 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.176197 kubelet[2714]: E0912 18:08:33.176093 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.177159 kubelet[2714]: E0912 18:08:33.176613 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.177159 kubelet[2714]: W0912 18:08:33.176630 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.177159 kubelet[2714]: E0912 18:08:33.176657 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.177159 kubelet[2714]: E0912 18:08:33.177020 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.177159 kubelet[2714]: W0912 18:08:33.177033 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.177159 kubelet[2714]: E0912 18:08:33.177066 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.177614 kubelet[2714]: E0912 18:08:33.177329 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.177614 kubelet[2714]: W0912 18:08:33.177338 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.177614 kubelet[2714]: E0912 18:08:33.177350 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.177614 kubelet[2714]: E0912 18:08:33.177598 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.177614 kubelet[2714]: W0912 18:08:33.177609 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.177990 kubelet[2714]: E0912 18:08:33.177627 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.177990 kubelet[2714]: E0912 18:08:33.177971 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.178089 kubelet[2714]: W0912 18:08:33.178002 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.178089 kubelet[2714]: E0912 18:08:33.178014 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.178977 kubelet[2714]: E0912 18:08:33.178955 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.178977 kubelet[2714]: W0912 18:08:33.178973 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.179214 kubelet[2714]: E0912 18:08:33.179049 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.179864 kubelet[2714]: E0912 18:08:33.179550 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.179864 kubelet[2714]: W0912 18:08:33.179566 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.179864 kubelet[2714]: E0912 18:08:33.179832 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.180058 kubelet[2714]: E0912 18:08:33.179922 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.180058 kubelet[2714]: W0912 18:08:33.179933 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.180058 kubelet[2714]: E0912 18:08:33.179951 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.180332 kubelet[2714]: E0912 18:08:33.180317 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.180332 kubelet[2714]: W0912 18:08:33.180330 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.180616 kubelet[2714]: E0912 18:08:33.180342 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.180868 kubelet[2714]: E0912 18:08:33.180852 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.180943 kubelet[2714]: W0912 18:08:33.180879 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.180943 kubelet[2714]: E0912 18:08:33.180891 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.181459 kubelet[2714]: E0912 18:08:33.181440 2714 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 18:08:33.181459 kubelet[2714]: W0912 18:08:33.181457 2714 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 18:08:33.181680 kubelet[2714]: E0912 18:08:33.181570 2714 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 18:08:33.738141 containerd[1562]: time="2025-09-12T18:08:33.737559581Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:33.739589 containerd[1562]: time="2025-09-12T18:08:33.739357529Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Sep 12 18:08:33.740279 containerd[1562]: time="2025-09-12T18:08:33.740216041Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:33.744647 containerd[1562]: time="2025-09-12T18:08:33.744593378Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:33.746881 containerd[1562]: time="2025-09-12T18:08:33.746710253Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 1.405358643s" Sep 12 18:08:33.746881 containerd[1562]: time="2025-09-12T18:08:33.746767488Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Sep 12 18:08:33.750528 containerd[1562]: time="2025-09-12T18:08:33.750301411Z" level=info msg="CreateContainer within sandbox \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 12 18:08:33.764514 containerd[1562]: time="2025-09-12T18:08:33.762697459Z" level=info msg="Container 30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:33.766923 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1965432451.mount: Deactivated successfully. Sep 12 18:08:33.775812 containerd[1562]: time="2025-09-12T18:08:33.775743773Z" level=info msg="CreateContainer within sandbox \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\"" Sep 12 18:08:33.777230 containerd[1562]: time="2025-09-12T18:08:33.777189854Z" level=info msg="StartContainer for \"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\"" Sep 12 18:08:33.779606 containerd[1562]: time="2025-09-12T18:08:33.779541788Z" level=info msg="connecting to shim 30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7" address="unix:///run/containerd/s/850e0fd976ad36998fc0f8caa82b4a266ea028bc9bf9013ccc6c1ba4a65a8745" protocol=ttrpc version=3 Sep 12 18:08:33.827810 systemd[1]: Started cri-containerd-30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7.scope - libcontainer container 30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7. Sep 12 18:08:33.892828 containerd[1562]: time="2025-09-12T18:08:33.892771690Z" level=info msg="StartContainer for \"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\" returns successfully" Sep 12 18:08:33.906431 systemd[1]: cri-containerd-30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7.scope: Deactivated successfully. Sep 12 18:08:33.931379 containerd[1562]: time="2025-09-12T18:08:33.931247600Z" level=info msg="TaskExit event in podsandbox handler container_id:\"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\" id:\"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\" pid:3400 exited_at:{seconds:1757700513 nanos:915297085}" Sep 12 18:08:33.931379 containerd[1562]: time="2025-09-12T18:08:33.931322022Z" level=info msg="received exit event container_id:\"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\" id:\"30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7\" pid:3400 exited_at:{seconds:1757700513 nanos:915297085}" Sep 12 18:08:33.972849 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-30e6c728fccd9be2336fd1bbcde9da66d57f144fa0402c6ccb1af3bb07b301f7-rootfs.mount: Deactivated successfully. Sep 12 18:08:33.994875 kubelet[2714]: E0912 18:08:33.993700 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:34.112637 containerd[1562]: time="2025-09-12T18:08:34.112278171Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 12 18:08:34.113716 kubelet[2714]: I0912 18:08:34.113645 2714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 18:08:34.116577 kubelet[2714]: E0912 18:08:34.116541 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:34.133378 kubelet[2714]: I0912 18:08:34.132748 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5998cc644d-5ztgm" podStartSLOduration=2.58881669 podStartE2EDuration="6.132723549s" podCreationTimestamp="2025-09-12 18:08:28 +0000 UTC" firstStartedPulling="2025-09-12 18:08:28.796851497 +0000 UTC m=+20.957711773" lastFinishedPulling="2025-09-12 18:08:32.340758335 +0000 UTC m=+24.501618632" observedRunningTime="2025-09-12 18:08:33.123669999 +0000 UTC m=+25.284530324" watchObservedRunningTime="2025-09-12 18:08:34.132723549 +0000 UTC m=+26.293583883" Sep 12 18:08:35.994510 kubelet[2714]: E0912 18:08:35.994387 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:37.043410 kubelet[2714]: I0912 18:08:37.042796 2714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 18:08:37.043988 kubelet[2714]: E0912 18:08:37.043538 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:37.122244 kubelet[2714]: E0912 18:08:37.121586 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:37.996731 kubelet[2714]: E0912 18:08:37.996671 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:38.208573 containerd[1562]: time="2025-09-12T18:08:38.207663325Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:38.209393 containerd[1562]: time="2025-09-12T18:08:38.209354470Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Sep 12 18:08:38.209689 containerd[1562]: time="2025-09-12T18:08:38.209662892Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:38.211771 containerd[1562]: time="2025-09-12T18:08:38.211733896Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:38.212716 containerd[1562]: time="2025-09-12T18:08:38.212678717Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 4.100350093s" Sep 12 18:08:38.212716 containerd[1562]: time="2025-09-12T18:08:38.212718506Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Sep 12 18:08:38.217614 containerd[1562]: time="2025-09-12T18:08:38.217433866Z" level=info msg="CreateContainer within sandbox \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 12 18:08:38.228862 containerd[1562]: time="2025-09-12T18:08:38.228806431Z" level=info msg="Container 0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:38.234698 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2072620922.mount: Deactivated successfully. Sep 12 18:08:38.246236 containerd[1562]: time="2025-09-12T18:08:38.246159153Z" level=info msg="CreateContainer within sandbox \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\"" Sep 12 18:08:38.248572 containerd[1562]: time="2025-09-12T18:08:38.247450014Z" level=info msg="StartContainer for \"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\"" Sep 12 18:08:38.250547 containerd[1562]: time="2025-09-12T18:08:38.250459990Z" level=info msg="connecting to shim 0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182" address="unix:///run/containerd/s/850e0fd976ad36998fc0f8caa82b4a266ea028bc9bf9013ccc6c1ba4a65a8745" protocol=ttrpc version=3 Sep 12 18:08:38.284813 systemd[1]: Started cri-containerd-0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182.scope - libcontainer container 0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182. Sep 12 18:08:38.392660 containerd[1562]: time="2025-09-12T18:08:38.392610922Z" level=info msg="StartContainer for \"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\" returns successfully" Sep 12 18:08:39.041672 systemd[1]: cri-containerd-0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182.scope: Deactivated successfully. Sep 12 18:08:39.041960 systemd[1]: cri-containerd-0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182.scope: Consumed 597ms CPU time, 153.3M memory peak, 7.7M read from disk, 171.3M written to disk. Sep 12 18:08:39.046512 containerd[1562]: time="2025-09-12T18:08:39.046381287Z" level=info msg="received exit event container_id:\"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\" id:\"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\" pid:3460 exited_at:{seconds:1757700519 nanos:46033387}" Sep 12 18:08:39.047594 containerd[1562]: time="2025-09-12T18:08:39.047557246Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\" id:\"0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182\" pid:3460 exited_at:{seconds:1757700519 nanos:46033387}" Sep 12 18:08:39.089229 kubelet[2714]: I0912 18:08:39.087747 2714 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 12 18:08:39.094682 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0d05f9513cdf1fe0531118a98f4a27d7acad2654465033c470eb6e98d55a8182-rootfs.mount: Deactivated successfully. Sep 12 18:08:39.141347 containerd[1562]: time="2025-09-12T18:08:39.140645370Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 12 18:08:39.159343 systemd[1]: Created slice kubepods-burstable-podaf0fd070_cb87_4036_9733_c0c1bf7001d7.slice - libcontainer container kubepods-burstable-podaf0fd070_cb87_4036_9733_c0c1bf7001d7.slice. Sep 12 18:08:39.185076 kubelet[2714]: W0912 18:08:39.183945 2714 reflector.go:561] object-"kube-system"/"coredns": failed to list *v1.ConfigMap: configmaps "coredns" is forbidden: User "system:node:ci-4426.1.0-c-cd41a2aea4" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4426.1.0-c-cd41a2aea4' and this object Sep 12 18:08:39.185076 kubelet[2714]: E0912 18:08:39.183996 2714 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"coredns\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"coredns\" is forbidden: User \"system:node:ci-4426.1.0-c-cd41a2aea4\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4426.1.0-c-cd41a2aea4' and this object" logger="UnhandledError" Sep 12 18:08:39.198599 systemd[1]: Created slice kubepods-besteffort-pod5b5d2ce6_01c4_4e9f_921d_be7a721ef2d1.slice - libcontainer container kubepods-besteffort-pod5b5d2ce6_01c4_4e9f_921d_be7a721ef2d1.slice. Sep 12 18:08:39.210888 systemd[1]: Created slice kubepods-burstable-pod7dcd9f5e_1b7c_4c5f_90ce_0fb0ba14636f.slice - libcontainer container kubepods-burstable-pod7dcd9f5e_1b7c_4c5f_90ce_0fb0ba14636f.slice. Sep 12 18:08:39.221700 systemd[1]: Created slice kubepods-besteffort-pod03ef677f_8c88_4982_8f89_bcb65e1d571b.slice - libcontainer container kubepods-besteffort-pod03ef677f_8c88_4982_8f89_bcb65e1d571b.slice. Sep 12 18:08:39.224527 kubelet[2714]: I0912 18:08:39.223691 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/b88db574-df8a-411c-a754-0d8df69ed036-calico-apiserver-certs\") pod \"calico-apiserver-6f9c78995-2bd6k\" (UID: \"b88db574-df8a-411c-a754-0d8df69ed036\") " pod="calico-apiserver/calico-apiserver-6f9c78995-2bd6k" Sep 12 18:08:39.225797 kubelet[2714]: I0912 18:08:39.225768 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-backend-key-pair\") pod \"whisker-6c75bf8899-qknpp\" (UID: \"03ef677f-8c88-4982-8f89-bcb65e1d571b\") " pod="calico-system/whisker-6c75bf8899-qknpp" Sep 12 18:08:39.226378 kubelet[2714]: I0912 18:08:39.225818 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68n9d\" (UniqueName: \"kubernetes.io/projected/5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1-kube-api-access-68n9d\") pod \"calico-kube-controllers-855d7c9bb7-bjgr8\" (UID: \"5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1\") " pod="calico-system/calico-kube-controllers-855d7c9bb7-bjgr8" Sep 12 18:08:39.226378 kubelet[2714]: I0912 18:08:39.225845 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08f7765b-0da5-414e-bf81-140d051267cd-goldmane-ca-bundle\") pod \"goldmane-7988f88666-dsgd8\" (UID: \"08f7765b-0da5-414e-bf81-140d051267cd\") " pod="calico-system/goldmane-7988f88666-dsgd8" Sep 12 18:08:39.226378 kubelet[2714]: I0912 18:08:39.225906 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/8428e997-ec0b-47c0-8e5c-deb3401be527-calico-apiserver-certs\") pod \"calico-apiserver-6f9c78995-44njx\" (UID: \"8428e997-ec0b-47c0-8e5c-deb3401be527\") " pod="calico-apiserver/calico-apiserver-6f9c78995-44njx" Sep 12 18:08:39.226378 kubelet[2714]: I0912 18:08:39.225928 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npdhs\" (UniqueName: \"kubernetes.io/projected/8428e997-ec0b-47c0-8e5c-deb3401be527-kube-api-access-npdhs\") pod \"calico-apiserver-6f9c78995-44njx\" (UID: \"8428e997-ec0b-47c0-8e5c-deb3401be527\") " pod="calico-apiserver/calico-apiserver-6f9c78995-44njx" Sep 12 18:08:39.226378 kubelet[2714]: I0912 18:08:39.225969 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-ca-bundle\") pod \"whisker-6c75bf8899-qknpp\" (UID: \"03ef677f-8c88-4982-8f89-bcb65e1d571b\") " pod="calico-system/whisker-6c75bf8899-qknpp" Sep 12 18:08:39.226549 kubelet[2714]: I0912 18:08:39.225986 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rftk7\" (UniqueName: \"kubernetes.io/projected/03ef677f-8c88-4982-8f89-bcb65e1d571b-kube-api-access-rftk7\") pod \"whisker-6c75bf8899-qknpp\" (UID: \"03ef677f-8c88-4982-8f89-bcb65e1d571b\") " pod="calico-system/whisker-6c75bf8899-qknpp" Sep 12 18:08:39.226549 kubelet[2714]: I0912 18:08:39.226004 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f-config-volume\") pod \"coredns-7c65d6cfc9-zncc7\" (UID: \"7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f\") " pod="kube-system/coredns-7c65d6cfc9-zncc7" Sep 12 18:08:39.226549 kubelet[2714]: I0912 18:08:39.226019 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1-tigera-ca-bundle\") pod \"calico-kube-controllers-855d7c9bb7-bjgr8\" (UID: \"5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1\") " pod="calico-system/calico-kube-controllers-855d7c9bb7-bjgr8" Sep 12 18:08:39.226549 kubelet[2714]: I0912 18:08:39.226044 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmmrh\" (UniqueName: \"kubernetes.io/projected/af0fd070-cb87-4036-9733-c0c1bf7001d7-kube-api-access-jmmrh\") pod \"coredns-7c65d6cfc9-5wflb\" (UID: \"af0fd070-cb87-4036-9733-c0c1bf7001d7\") " pod="kube-system/coredns-7c65d6cfc9-5wflb" Sep 12 18:08:39.226549 kubelet[2714]: I0912 18:08:39.226063 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6nxx\" (UniqueName: \"kubernetes.io/projected/08f7765b-0da5-414e-bf81-140d051267cd-kube-api-access-d6nxx\") pod \"goldmane-7988f88666-dsgd8\" (UID: \"08f7765b-0da5-414e-bf81-140d051267cd\") " pod="calico-system/goldmane-7988f88666-dsgd8" Sep 12 18:08:39.226684 kubelet[2714]: I0912 18:08:39.226092 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8nqp\" (UniqueName: \"kubernetes.io/projected/7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f-kube-api-access-c8nqp\") pod \"coredns-7c65d6cfc9-zncc7\" (UID: \"7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f\") " pod="kube-system/coredns-7c65d6cfc9-zncc7" Sep 12 18:08:39.226684 kubelet[2714]: I0912 18:08:39.226117 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08f7765b-0da5-414e-bf81-140d051267cd-config\") pod \"goldmane-7988f88666-dsgd8\" (UID: \"08f7765b-0da5-414e-bf81-140d051267cd\") " pod="calico-system/goldmane-7988f88666-dsgd8" Sep 12 18:08:39.226684 kubelet[2714]: I0912 18:08:39.226134 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxmq9\" (UniqueName: \"kubernetes.io/projected/b88db574-df8a-411c-a754-0d8df69ed036-kube-api-access-zxmq9\") pod \"calico-apiserver-6f9c78995-2bd6k\" (UID: \"b88db574-df8a-411c-a754-0d8df69ed036\") " pod="calico-apiserver/calico-apiserver-6f9c78995-2bd6k" Sep 12 18:08:39.226684 kubelet[2714]: I0912 18:08:39.226154 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0fd070-cb87-4036-9733-c0c1bf7001d7-config-volume\") pod \"coredns-7c65d6cfc9-5wflb\" (UID: \"af0fd070-cb87-4036-9733-c0c1bf7001d7\") " pod="kube-system/coredns-7c65d6cfc9-5wflb" Sep 12 18:08:39.226684 kubelet[2714]: I0912 18:08:39.226172 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/08f7765b-0da5-414e-bf81-140d051267cd-goldmane-key-pair\") pod \"goldmane-7988f88666-dsgd8\" (UID: \"08f7765b-0da5-414e-bf81-140d051267cd\") " pod="calico-system/goldmane-7988f88666-dsgd8" Sep 12 18:08:39.233342 systemd[1]: Created slice kubepods-besteffort-pod8428e997_ec0b_47c0_8e5c_deb3401be527.slice - libcontainer container kubepods-besteffort-pod8428e997_ec0b_47c0_8e5c_deb3401be527.slice. Sep 12 18:08:39.242783 systemd[1]: Created slice kubepods-besteffort-podb88db574_df8a_411c_a754_0d8df69ed036.slice - libcontainer container kubepods-besteffort-podb88db574_df8a_411c_a754_0d8df69ed036.slice. Sep 12 18:08:39.253427 systemd[1]: Created slice kubepods-besteffort-pod08f7765b_0da5_414e_bf81_140d051267cd.slice - libcontainer container kubepods-besteffort-pod08f7765b_0da5_414e_bf81_140d051267cd.slice. Sep 12 18:08:39.505526 containerd[1562]: time="2025-09-12T18:08:39.505089169Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-855d7c9bb7-bjgr8,Uid:5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:39.531071 containerd[1562]: time="2025-09-12T18:08:39.530793249Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6c75bf8899-qknpp,Uid:03ef677f-8c88-4982-8f89-bcb65e1d571b,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:39.543555 containerd[1562]: time="2025-09-12T18:08:39.543450673Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-44njx,Uid:8428e997-ec0b-47c0-8e5c-deb3401be527,Namespace:calico-apiserver,Attempt:0,}" Sep 12 18:08:39.552618 containerd[1562]: time="2025-09-12T18:08:39.552383445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-2bd6k,Uid:b88db574-df8a-411c-a754-0d8df69ed036,Namespace:calico-apiserver,Attempt:0,}" Sep 12 18:08:39.570101 containerd[1562]: time="2025-09-12T18:08:39.570053896Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-dsgd8,Uid:08f7765b-0da5-414e-bf81-140d051267cd,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:39.831123 containerd[1562]: time="2025-09-12T18:08:39.830906159Z" level=error msg="Failed to destroy network for sandbox \"40551576b7120aa577be46988734d4bcc44bdc1e181f41e0120c35326429ae95\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.837442 containerd[1562]: time="2025-09-12T18:08:39.837379583Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-dsgd8,Uid:08f7765b-0da5-414e-bf81-140d051267cd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"40551576b7120aa577be46988734d4bcc44bdc1e181f41e0120c35326429ae95\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.844529 kubelet[2714]: E0912 18:08:39.844417 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"40551576b7120aa577be46988734d4bcc44bdc1e181f41e0120c35326429ae95\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.844727 kubelet[2714]: E0912 18:08:39.844585 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"40551576b7120aa577be46988734d4bcc44bdc1e181f41e0120c35326429ae95\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7988f88666-dsgd8" Sep 12 18:08:39.846690 kubelet[2714]: E0912 18:08:39.846620 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"40551576b7120aa577be46988734d4bcc44bdc1e181f41e0120c35326429ae95\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7988f88666-dsgd8" Sep 12 18:08:39.846816 kubelet[2714]: E0912 18:08:39.846749 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7988f88666-dsgd8_calico-system(08f7765b-0da5-414e-bf81-140d051267cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7988f88666-dsgd8_calico-system(08f7765b-0da5-414e-bf81-140d051267cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"40551576b7120aa577be46988734d4bcc44bdc1e181f41e0120c35326429ae95\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7988f88666-dsgd8" podUID="08f7765b-0da5-414e-bf81-140d051267cd" Sep 12 18:08:39.847055 containerd[1562]: time="2025-09-12T18:08:39.847004221Z" level=error msg="Failed to destroy network for sandbox \"871b151946b850be60201d5c516e617f284e03e26bee2cdebb7d3eee8b4a3323\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.851545 containerd[1562]: time="2025-09-12T18:08:39.851493042Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-44njx,Uid:8428e997-ec0b-47c0-8e5c-deb3401be527,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"871b151946b850be60201d5c516e617f284e03e26bee2cdebb7d3eee8b4a3323\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.852718 kubelet[2714]: E0912 18:08:39.852620 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"871b151946b850be60201d5c516e617f284e03e26bee2cdebb7d3eee8b4a3323\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.853569 kubelet[2714]: E0912 18:08:39.852853 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"871b151946b850be60201d5c516e617f284e03e26bee2cdebb7d3eee8b4a3323\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6f9c78995-44njx" Sep 12 18:08:39.853569 kubelet[2714]: E0912 18:08:39.852896 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"871b151946b850be60201d5c516e617f284e03e26bee2cdebb7d3eee8b4a3323\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6f9c78995-44njx" Sep 12 18:08:39.853569 kubelet[2714]: E0912 18:08:39.852941 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6f9c78995-44njx_calico-apiserver(8428e997-ec0b-47c0-8e5c-deb3401be527)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6f9c78995-44njx_calico-apiserver(8428e997-ec0b-47c0-8e5c-deb3401be527)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"871b151946b850be60201d5c516e617f284e03e26bee2cdebb7d3eee8b4a3323\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6f9c78995-44njx" podUID="8428e997-ec0b-47c0-8e5c-deb3401be527" Sep 12 18:08:39.883193 containerd[1562]: time="2025-09-12T18:08:39.883144270Z" level=error msg="Failed to destroy network for sandbox \"d4c288de13a09033f1fd4abbb593e3a35083fb24a0313fabbeab384c793b8418\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.884909 containerd[1562]: time="2025-09-12T18:08:39.884860593Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6c75bf8899-qknpp,Uid:03ef677f-8c88-4982-8f89-bcb65e1d571b,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4c288de13a09033f1fd4abbb593e3a35083fb24a0313fabbeab384c793b8418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.885769 kubelet[2714]: E0912 18:08:39.885729 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4c288de13a09033f1fd4abbb593e3a35083fb24a0313fabbeab384c793b8418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.885869 kubelet[2714]: E0912 18:08:39.885791 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4c288de13a09033f1fd4abbb593e3a35083fb24a0313fabbeab384c793b8418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6c75bf8899-qknpp" Sep 12 18:08:39.885869 kubelet[2714]: E0912 18:08:39.885839 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4c288de13a09033f1fd4abbb593e3a35083fb24a0313fabbeab384c793b8418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6c75bf8899-qknpp" Sep 12 18:08:39.885937 kubelet[2714]: E0912 18:08:39.885888 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6c75bf8899-qknpp_calico-system(03ef677f-8c88-4982-8f89-bcb65e1d571b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6c75bf8899-qknpp_calico-system(03ef677f-8c88-4982-8f89-bcb65e1d571b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d4c288de13a09033f1fd4abbb593e3a35083fb24a0313fabbeab384c793b8418\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6c75bf8899-qknpp" podUID="03ef677f-8c88-4982-8f89-bcb65e1d571b" Sep 12 18:08:39.887032 containerd[1562]: time="2025-09-12T18:08:39.886879489Z" level=error msg="Failed to destroy network for sandbox \"b7be43d4f35ed58c4c357695afc6d93398127e88a01c17d97e23156e33246cd9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.889802 containerd[1562]: time="2025-09-12T18:08:39.889680193Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-2bd6k,Uid:b88db574-df8a-411c-a754-0d8df69ed036,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7be43d4f35ed58c4c357695afc6d93398127e88a01c17d97e23156e33246cd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.890429 kubelet[2714]: E0912 18:08:39.890383 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7be43d4f35ed58c4c357695afc6d93398127e88a01c17d97e23156e33246cd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.890541 kubelet[2714]: E0912 18:08:39.890459 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7be43d4f35ed58c4c357695afc6d93398127e88a01c17d97e23156e33246cd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6f9c78995-2bd6k" Sep 12 18:08:39.890541 kubelet[2714]: E0912 18:08:39.890494 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7be43d4f35ed58c4c357695afc6d93398127e88a01c17d97e23156e33246cd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6f9c78995-2bd6k" Sep 12 18:08:39.890612 kubelet[2714]: E0912 18:08:39.890545 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6f9c78995-2bd6k_calico-apiserver(b88db574-df8a-411c-a754-0d8df69ed036)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6f9c78995-2bd6k_calico-apiserver(b88db574-df8a-411c-a754-0d8df69ed036)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b7be43d4f35ed58c4c357695afc6d93398127e88a01c17d97e23156e33246cd9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6f9c78995-2bd6k" podUID="b88db574-df8a-411c-a754-0d8df69ed036" Sep 12 18:08:39.898234 containerd[1562]: time="2025-09-12T18:08:39.898110293Z" level=error msg="Failed to destroy network for sandbox \"2e74331c701101205a488de808401a12e0770b1534573e7eb092d76cd81aeffa\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.918298 containerd[1562]: time="2025-09-12T18:08:39.918158308Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-855d7c9bb7-bjgr8,Uid:5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2e74331c701101205a488de808401a12e0770b1534573e7eb092d76cd81aeffa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.918742 kubelet[2714]: E0912 18:08:39.918705 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2e74331c701101205a488de808401a12e0770b1534573e7eb092d76cd81aeffa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:39.918817 kubelet[2714]: E0912 18:08:39.918766 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2e74331c701101205a488de808401a12e0770b1534573e7eb092d76cd81aeffa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-855d7c9bb7-bjgr8" Sep 12 18:08:39.918817 kubelet[2714]: E0912 18:08:39.918785 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2e74331c701101205a488de808401a12e0770b1534573e7eb092d76cd81aeffa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-855d7c9bb7-bjgr8" Sep 12 18:08:39.918875 kubelet[2714]: E0912 18:08:39.918831 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-855d7c9bb7-bjgr8_calico-system(5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-855d7c9bb7-bjgr8_calico-system(5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2e74331c701101205a488de808401a12e0770b1534573e7eb092d76cd81aeffa\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-855d7c9bb7-bjgr8" podUID="5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1" Sep 12 18:08:40.002026 systemd[1]: Created slice kubepods-besteffort-podb51a8f2d_d1c4_4004_80da_1576b8879e25.slice - libcontainer container kubepods-besteffort-podb51a8f2d_d1c4_4004_80da_1576b8879e25.slice. Sep 12 18:08:40.009687 containerd[1562]: time="2025-09-12T18:08:40.009591676Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-8hj6c,Uid:b51a8f2d-d1c4-4004-80da-1576b8879e25,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:40.072622 containerd[1562]: time="2025-09-12T18:08:40.072457398Z" level=error msg="Failed to destroy network for sandbox \"db58b3ac01ffc7524d218f266feb854d14c175fb6cc0776d5e249b61babe5601\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:40.074139 containerd[1562]: time="2025-09-12T18:08:40.073520864Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-8hj6c,Uid:b51a8f2d-d1c4-4004-80da-1576b8879e25,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"db58b3ac01ffc7524d218f266feb854d14c175fb6cc0776d5e249b61babe5601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:40.074303 kubelet[2714]: E0912 18:08:40.073744 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db58b3ac01ffc7524d218f266feb854d14c175fb6cc0776d5e249b61babe5601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:40.074303 kubelet[2714]: E0912 18:08:40.073808 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db58b3ac01ffc7524d218f266feb854d14c175fb6cc0776d5e249b61babe5601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:40.074303 kubelet[2714]: E0912 18:08:40.073830 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db58b3ac01ffc7524d218f266feb854d14c175fb6cc0776d5e249b61babe5601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-8hj6c" Sep 12 18:08:40.074414 kubelet[2714]: E0912 18:08:40.073870 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-8hj6c_calico-system(b51a8f2d-d1c4-4004-80da-1576b8879e25)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-8hj6c_calico-system(b51a8f2d-d1c4-4004-80da-1576b8879e25)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"db58b3ac01ffc7524d218f266feb854d14c175fb6cc0776d5e249b61babe5601\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-8hj6c" podUID="b51a8f2d-d1c4-4004-80da-1576b8879e25" Sep 12 18:08:40.336853 kubelet[2714]: E0912 18:08:40.336777 2714 configmap.go:193] Couldn't get configMap kube-system/coredns: failed to sync configmap cache: timed out waiting for the condition Sep 12 18:08:40.337295 kubelet[2714]: E0912 18:08:40.336908 2714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af0fd070-cb87-4036-9733-c0c1bf7001d7-config-volume podName:af0fd070-cb87-4036-9733-c0c1bf7001d7 nodeName:}" failed. No retries permitted until 2025-09-12 18:08:40.836886724 +0000 UTC m=+32.997747000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/af0fd070-cb87-4036-9733-c0c1bf7001d7-config-volume") pod "coredns-7c65d6cfc9-5wflb" (UID: "af0fd070-cb87-4036-9733-c0c1bf7001d7") : failed to sync configmap cache: timed out waiting for the condition Sep 12 18:08:40.338037 kubelet[2714]: E0912 18:08:40.337920 2714 configmap.go:193] Couldn't get configMap kube-system/coredns: failed to sync configmap cache: timed out waiting for the condition Sep 12 18:08:40.338037 kubelet[2714]: E0912 18:08:40.338009 2714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f-config-volume podName:7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f nodeName:}" failed. No retries permitted until 2025-09-12 18:08:40.837989882 +0000 UTC m=+32.998850171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f-config-volume") pod "coredns-7c65d6cfc9-zncc7" (UID: "7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f") : failed to sync configmap cache: timed out waiting for the condition Sep 12 18:08:40.998603 kubelet[2714]: E0912 18:08:40.998401 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:41.008063 containerd[1562]: time="2025-09-12T18:08:41.008019842Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-5wflb,Uid:af0fd070-cb87-4036-9733-c0c1bf7001d7,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:41.020664 kubelet[2714]: E0912 18:08:41.020535 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:41.028680 containerd[1562]: time="2025-09-12T18:08:41.027427014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-zncc7,Uid:7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:41.211001 containerd[1562]: time="2025-09-12T18:08:41.210872535Z" level=error msg="Failed to destroy network for sandbox \"6f3f5f629258f53adca47973f8c961d6856c4996f2b8a45851e61adea6fad429\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:41.214879 systemd[1]: run-netns-cni\x2dd7db1d95\x2d8164\x2df2f2\x2de959\x2d0d4b0a405504.mount: Deactivated successfully. Sep 12 18:08:41.215808 containerd[1562]: time="2025-09-12T18:08:41.215041896Z" level=error msg="Failed to destroy network for sandbox \"5c13f3586c7b8c7b329c284ce2fffc12e9afdf9c79c01a27cf8b52c7f00caf5f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:41.221115 systemd[1]: run-netns-cni\x2d34bb655c\x2d52ef\x2de5e1\x2d3f78\x2d4f2ca252b0f0.mount: Deactivated successfully. Sep 12 18:08:41.235840 containerd[1562]: time="2025-09-12T18:08:41.235769405Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-5wflb,Uid:af0fd070-cb87-4036-9733-c0c1bf7001d7,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3f5f629258f53adca47973f8c961d6856c4996f2b8a45851e61adea6fad429\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:41.236252 containerd[1562]: time="2025-09-12T18:08:41.236224102Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-zncc7,Uid:7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c13f3586c7b8c7b329c284ce2fffc12e9afdf9c79c01a27cf8b52c7f00caf5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:41.236490 kubelet[2714]: E0912 18:08:41.236415 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c13f3586c7b8c7b329c284ce2fffc12e9afdf9c79c01a27cf8b52c7f00caf5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:41.236553 kubelet[2714]: E0912 18:08:41.236503 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c13f3586c7b8c7b329c284ce2fffc12e9afdf9c79c01a27cf8b52c7f00caf5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-zncc7" Sep 12 18:08:41.236553 kubelet[2714]: E0912 18:08:41.236531 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c13f3586c7b8c7b329c284ce2fffc12e9afdf9c79c01a27cf8b52c7f00caf5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-zncc7" Sep 12 18:08:41.236621 kubelet[2714]: E0912 18:08:41.236576 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-zncc7_kube-system(7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-zncc7_kube-system(7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5c13f3586c7b8c7b329c284ce2fffc12e9afdf9c79c01a27cf8b52c7f00caf5f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-zncc7" podUID="7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f" Sep 12 18:08:41.238315 kubelet[2714]: E0912 18:08:41.236420 2714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3f5f629258f53adca47973f8c961d6856c4996f2b8a45851e61adea6fad429\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 18:08:41.238315 kubelet[2714]: E0912 18:08:41.236723 2714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3f5f629258f53adca47973f8c961d6856c4996f2b8a45851e61adea6fad429\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-5wflb" Sep 12 18:08:41.238315 kubelet[2714]: E0912 18:08:41.236775 2714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6f3f5f629258f53adca47973f8c961d6856c4996f2b8a45851e61adea6fad429\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-5wflb" Sep 12 18:08:41.238558 kubelet[2714]: E0912 18:08:41.236811 2714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-5wflb_kube-system(af0fd070-cb87-4036-9733-c0c1bf7001d7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-5wflb_kube-system(af0fd070-cb87-4036-9733-c0c1bf7001d7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6f3f5f629258f53adca47973f8c961d6856c4996f2b8a45851e61adea6fad429\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-5wflb" podUID="af0fd070-cb87-4036-9733-c0c1bf7001d7" Sep 12 18:08:45.251049 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3172368233.mount: Deactivated successfully. Sep 12 18:08:45.301091 containerd[1562]: time="2025-09-12T18:08:45.301029974Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:45.302701 containerd[1562]: time="2025-09-12T18:08:45.302655853Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Sep 12 18:08:45.303533 containerd[1562]: time="2025-09-12T18:08:45.302917596Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:45.304703 containerd[1562]: time="2025-09-12T18:08:45.304666086Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:45.305357 containerd[1562]: time="2025-09-12T18:08:45.305327885Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 6.164636881s" Sep 12 18:08:45.305502 containerd[1562]: time="2025-09-12T18:08:45.305457351Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Sep 12 18:08:45.345841 containerd[1562]: time="2025-09-12T18:08:45.345801152Z" level=info msg="CreateContainer within sandbox \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 12 18:08:45.373615 containerd[1562]: time="2025-09-12T18:08:45.372923589Z" level=info msg="Container 1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:45.388748 containerd[1562]: time="2025-09-12T18:08:45.388692914Z" level=info msg="CreateContainer within sandbox \"e321547d580a7b2e4cbe01a0b7506c30431b9c8c7a6f22a56d1e9439f9afb3df\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\"" Sep 12 18:08:45.390312 containerd[1562]: time="2025-09-12T18:08:45.389815001Z" level=info msg="StartContainer for \"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\"" Sep 12 18:08:45.394678 containerd[1562]: time="2025-09-12T18:08:45.394642559Z" level=info msg="connecting to shim 1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e" address="unix:///run/containerd/s/850e0fd976ad36998fc0f8caa82b4a266ea028bc9bf9013ccc6c1ba4a65a8745" protocol=ttrpc version=3 Sep 12 18:08:45.569720 systemd[1]: Started cri-containerd-1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e.scope - libcontainer container 1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e. Sep 12 18:08:45.642108 containerd[1562]: time="2025-09-12T18:08:45.642046140Z" level=info msg="StartContainer for \"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\" returns successfully" Sep 12 18:08:45.770247 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 12 18:08:45.770391 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 12 18:08:45.997472 kubelet[2714]: I0912 18:08:45.997211 2714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-backend-key-pair\") pod \"03ef677f-8c88-4982-8f89-bcb65e1d571b\" (UID: \"03ef677f-8c88-4982-8f89-bcb65e1d571b\") " Sep 12 18:08:45.999835 kubelet[2714]: I0912 18:08:45.999010 2714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rftk7\" (UniqueName: \"kubernetes.io/projected/03ef677f-8c88-4982-8f89-bcb65e1d571b-kube-api-access-rftk7\") pod \"03ef677f-8c88-4982-8f89-bcb65e1d571b\" (UID: \"03ef677f-8c88-4982-8f89-bcb65e1d571b\") " Sep 12 18:08:45.999835 kubelet[2714]: I0912 18:08:45.999065 2714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-ca-bundle\") pod \"03ef677f-8c88-4982-8f89-bcb65e1d571b\" (UID: \"03ef677f-8c88-4982-8f89-bcb65e1d571b\") " Sep 12 18:08:46.000657 kubelet[2714]: I0912 18:08:45.999457 2714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "03ef677f-8c88-4982-8f89-bcb65e1d571b" (UID: "03ef677f-8c88-4982-8f89-bcb65e1d571b"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 12 18:08:46.004044 kubelet[2714]: I0912 18:08:46.004010 2714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ef677f-8c88-4982-8f89-bcb65e1d571b-kube-api-access-rftk7" (OuterVolumeSpecName: "kube-api-access-rftk7") pod "03ef677f-8c88-4982-8f89-bcb65e1d571b" (UID: "03ef677f-8c88-4982-8f89-bcb65e1d571b"). InnerVolumeSpecName "kube-api-access-rftk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 12 18:08:46.010306 kubelet[2714]: I0912 18:08:46.010196 2714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "03ef677f-8c88-4982-8f89-bcb65e1d571b" (UID: "03ef677f-8c88-4982-8f89-bcb65e1d571b"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 12 18:08:46.100081 kubelet[2714]: I0912 18:08:46.099910 2714 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-ca-bundle\") on node \"ci-4426.1.0-c-cd41a2aea4\" DevicePath \"\"" Sep 12 18:08:46.100081 kubelet[2714]: I0912 18:08:46.099948 2714 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/03ef677f-8c88-4982-8f89-bcb65e1d571b-whisker-backend-key-pair\") on node \"ci-4426.1.0-c-cd41a2aea4\" DevicePath \"\"" Sep 12 18:08:46.100081 kubelet[2714]: I0912 18:08:46.099958 2714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rftk7\" (UniqueName: \"kubernetes.io/projected/03ef677f-8c88-4982-8f89-bcb65e1d571b-kube-api-access-rftk7\") on node \"ci-4426.1.0-c-cd41a2aea4\" DevicePath \"\"" Sep 12 18:08:46.180818 systemd[1]: Removed slice kubepods-besteffort-pod03ef677f_8c88_4982_8f89_bcb65e1d571b.slice - libcontainer container kubepods-besteffort-pod03ef677f_8c88_4982_8f89_bcb65e1d571b.slice. Sep 12 18:08:46.224510 kubelet[2714]: I0912 18:08:46.224086 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-m2tpt" podStartSLOduration=2.011873729 podStartE2EDuration="18.224062293s" podCreationTimestamp="2025-09-12 18:08:28 +0000 UTC" firstStartedPulling="2025-09-12 18:08:29.100902476 +0000 UTC m=+21.261762765" lastFinishedPulling="2025-09-12 18:08:45.313091049 +0000 UTC m=+37.473951329" observedRunningTime="2025-09-12 18:08:46.202727075 +0000 UTC m=+38.363587377" watchObservedRunningTime="2025-09-12 18:08:46.224062293 +0000 UTC m=+38.384922584" Sep 12 18:08:46.251777 systemd[1]: var-lib-kubelet-pods-03ef677f\x2d8c88\x2d4982\x2d8f89\x2dbcb65e1d571b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2drftk7.mount: Deactivated successfully. Sep 12 18:08:46.254077 systemd[1]: var-lib-kubelet-pods-03ef677f\x2d8c88\x2d4982\x2d8f89\x2dbcb65e1d571b-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 12 18:08:46.309724 systemd[1]: Created slice kubepods-besteffort-pode5c7a50b_dd30_4a73_b3dc_53c1d3f12335.slice - libcontainer container kubepods-besteffort-pode5c7a50b_dd30_4a73_b3dc_53c1d3f12335.slice. Sep 12 18:08:46.404958 kubelet[2714]: I0912 18:08:46.404893 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5c7a50b-dd30-4a73-b3dc-53c1d3f12335-whisker-ca-bundle\") pod \"whisker-5885c4d4d6-vl2xj\" (UID: \"e5c7a50b-dd30-4a73-b3dc-53c1d3f12335\") " pod="calico-system/whisker-5885c4d4d6-vl2xj" Sep 12 18:08:46.404958 kubelet[2714]: I0912 18:08:46.404961 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz7wp\" (UniqueName: \"kubernetes.io/projected/e5c7a50b-dd30-4a73-b3dc-53c1d3f12335-kube-api-access-mz7wp\") pod \"whisker-5885c4d4d6-vl2xj\" (UID: \"e5c7a50b-dd30-4a73-b3dc-53c1d3f12335\") " pod="calico-system/whisker-5885c4d4d6-vl2xj" Sep 12 18:08:46.405231 kubelet[2714]: I0912 18:08:46.404998 2714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e5c7a50b-dd30-4a73-b3dc-53c1d3f12335-whisker-backend-key-pair\") pod \"whisker-5885c4d4d6-vl2xj\" (UID: \"e5c7a50b-dd30-4a73-b3dc-53c1d3f12335\") " pod="calico-system/whisker-5885c4d4d6-vl2xj" Sep 12 18:08:46.421577 containerd[1562]: time="2025-09-12T18:08:46.421533473Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\" id:\"f61fd0abd09062f91b0aa398607dd235404e01bbf7cafb3a45fa545d6953948a\" pid:3783 exit_status:1 exited_at:{seconds:1757700526 nanos:420931121}" Sep 12 18:08:46.615808 containerd[1562]: time="2025-09-12T18:08:46.615651140Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5885c4d4d6-vl2xj,Uid:e5c7a50b-dd30-4a73-b3dc-53c1d3f12335,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:46.940568 systemd-networkd[1450]: calideea7767c59: Link UP Sep 12 18:08:46.941049 systemd-networkd[1450]: calideea7767c59: Gained carrier Sep 12 18:08:46.963663 containerd[1562]: 2025-09-12 18:08:46.663 [INFO][3808] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 12 18:08:46.963663 containerd[1562]: 2025-09-12 18:08:46.701 [INFO][3808] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0 whisker-5885c4d4d6- calico-system e5c7a50b-dd30-4a73-b3dc-53c1d3f12335 891 0 2025-09-12 18:08:46 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5885c4d4d6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 whisker-5885c4d4d6-vl2xj eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calideea7767c59 [] [] }} ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-" Sep 12 18:08:46.963663 containerd[1562]: 2025-09-12 18:08:46.701 [INFO][3808] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:46.963663 containerd[1562]: 2025-09-12 18:08:46.855 [INFO][3819] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" HandleID="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.857 [INFO][3819] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" HandleID="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001024b0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"whisker-5885c4d4d6-vl2xj", "timestamp":"2025-09-12 18:08:46.855034614 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.857 [INFO][3819] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.858 [INFO][3819] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.858 [INFO][3819] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.876 [INFO][3819] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.885 [INFO][3819] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.892 [INFO][3819] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.894 [INFO][3819] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.964079 containerd[1562]: 2025-09-12 18:08:46.897 [INFO][3819] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.897 [INFO][3819] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.899 [INFO][3819] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.910 [INFO][3819] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.917 [INFO][3819] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.1/26] block=192.168.6.0/26 handle="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.918 [INFO][3819] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.1/26] handle="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.918 [INFO][3819] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:46.967615 containerd[1562]: 2025-09-12 18:08:46.918 [INFO][3819] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.1/26] IPv6=[] ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" HandleID="k8s-pod-network.ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:46.967936 containerd[1562]: 2025-09-12 18:08:46.923 [INFO][3808] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0", GenerateName:"whisker-5885c4d4d6-", Namespace:"calico-system", SelfLink:"", UID:"e5c7a50b-dd30-4a73-b3dc-53c1d3f12335", ResourceVersion:"891", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5885c4d4d6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"whisker-5885c4d4d6-vl2xj", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.6.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calideea7767c59", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:46.967936 containerd[1562]: 2025-09-12 18:08:46.923 [INFO][3808] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.1/32] ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:46.968093 containerd[1562]: 2025-09-12 18:08:46.923 [INFO][3808] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calideea7767c59 ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:46.968093 containerd[1562]: 2025-09-12 18:08:46.938 [INFO][3808] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:46.968182 containerd[1562]: 2025-09-12 18:08:46.941 [INFO][3808] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0", GenerateName:"whisker-5885c4d4d6-", Namespace:"calico-system", SelfLink:"", UID:"e5c7a50b-dd30-4a73-b3dc-53c1d3f12335", ResourceVersion:"891", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5885c4d4d6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c", Pod:"whisker-5885c4d4d6-vl2xj", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.6.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calideea7767c59", MAC:"82:42:49:ab:93:0f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:46.968281 containerd[1562]: 2025-09-12 18:08:46.957 [INFO][3808] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" Namespace="calico-system" Pod="whisker-5885c4d4d6-vl2xj" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-whisker--5885c4d4d6--vl2xj-eth0" Sep 12 18:08:47.113827 containerd[1562]: time="2025-09-12T18:08:47.113780975Z" level=info msg="connecting to shim ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c" address="unix:///run/containerd/s/c7c8b8da15298b3d1e96399fd8b8c9b21d1ea1c8ce1909c89f781cd17293b27a" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:47.141752 systemd[1]: Started cri-containerd-ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c.scope - libcontainer container ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c. Sep 12 18:08:47.237074 containerd[1562]: time="2025-09-12T18:08:47.236926104Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5885c4d4d6-vl2xj,Uid:e5c7a50b-dd30-4a73-b3dc-53c1d3f12335,Namespace:calico-system,Attempt:0,} returns sandbox id \"ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c\"" Sep 12 18:08:47.242618 containerd[1562]: time="2025-09-12T18:08:47.242581482Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 12 18:08:47.304259 containerd[1562]: time="2025-09-12T18:08:47.304201752Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\" id:\"4285ca83a7154ba057f4636f862871f79a50d733c65deb4b4a301830856c6f93\" pid:3881 exit_status:1 exited_at:{seconds:1757700527 nanos:303115228}" Sep 12 18:08:47.998168 kubelet[2714]: I0912 18:08:47.998089 2714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ef677f-8c88-4982-8f89-bcb65e1d571b" path="/var/lib/kubelet/pods/03ef677f-8c88-4982-8f89-bcb65e1d571b/volumes" Sep 12 18:08:48.188105 systemd-networkd[1450]: vxlan.calico: Link UP Sep 12 18:08:48.188115 systemd-networkd[1450]: vxlan.calico: Gained carrier Sep 12 18:08:48.550229 systemd-networkd[1450]: calideea7767c59: Gained IPv6LL Sep 12 18:08:49.011788 containerd[1562]: time="2025-09-12T18:08:49.011677986Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:49.013299 containerd[1562]: time="2025-09-12T18:08:49.012500418Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Sep 12 18:08:49.013299 containerd[1562]: time="2025-09-12T18:08:49.013233782Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:49.028641 containerd[1562]: time="2025-09-12T18:08:49.028552572Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:49.029408 containerd[1562]: time="2025-09-12T18:08:49.029365268Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 1.785676292s" Sep 12 18:08:49.029529 containerd[1562]: time="2025-09-12T18:08:49.029412342Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Sep 12 18:08:49.033710 containerd[1562]: time="2025-09-12T18:08:49.033658620Z" level=info msg="CreateContainer within sandbox \"ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 12 18:08:49.041532 containerd[1562]: time="2025-09-12T18:08:49.040635020Z" level=info msg="Container 82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:49.054698 containerd[1562]: time="2025-09-12T18:08:49.054654074Z" level=info msg="CreateContainer within sandbox \"ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68\"" Sep 12 18:08:49.056242 containerd[1562]: time="2025-09-12T18:08:49.055431740Z" level=info msg="StartContainer for \"82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68\"" Sep 12 18:08:49.058142 containerd[1562]: time="2025-09-12T18:08:49.058099467Z" level=info msg="connecting to shim 82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68" address="unix:///run/containerd/s/c7c8b8da15298b3d1e96399fd8b8c9b21d1ea1c8ce1909c89f781cd17293b27a" protocol=ttrpc version=3 Sep 12 18:08:49.093721 systemd[1]: Started cri-containerd-82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68.scope - libcontainer container 82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68. Sep 12 18:08:49.162442 containerd[1562]: time="2025-09-12T18:08:49.162315561Z" level=info msg="StartContainer for \"82cbae73d8937bfb123648d2c57243556c83065323338ea80b78ac971de9cb68\" returns successfully" Sep 12 18:08:49.165512 containerd[1562]: time="2025-09-12T18:08:49.165450357Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 12 18:08:49.829807 systemd-networkd[1450]: vxlan.calico: Gained IPv6LL Sep 12 18:08:51.260029 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3579169722.mount: Deactivated successfully. Sep 12 18:08:51.283161 containerd[1562]: time="2025-09-12T18:08:51.283103874Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:51.284899 containerd[1562]: time="2025-09-12T18:08:51.284811751Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Sep 12 18:08:51.285773 containerd[1562]: time="2025-09-12T18:08:51.285733634Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:51.288844 containerd[1562]: time="2025-09-12T18:08:51.288774444Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:51.295287 containerd[1562]: time="2025-09-12T18:08:51.294547857Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 2.128885219s" Sep 12 18:08:51.295287 containerd[1562]: time="2025-09-12T18:08:51.294612621Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Sep 12 18:08:51.302954 containerd[1562]: time="2025-09-12T18:08:51.302799121Z" level=info msg="CreateContainer within sandbox \"ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 12 18:08:51.314704 containerd[1562]: time="2025-09-12T18:08:51.313676001Z" level=info msg="Container 4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:51.323507 containerd[1562]: time="2025-09-12T18:08:51.323414141Z" level=info msg="CreateContainer within sandbox \"ad932c210125dc4217eecdf98f846ba5c08d1ad631526bddf82416b2ae4c2b4c\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3\"" Sep 12 18:08:51.326591 containerd[1562]: time="2025-09-12T18:08:51.326020080Z" level=info msg="StartContainer for \"4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3\"" Sep 12 18:08:51.329083 containerd[1562]: time="2025-09-12T18:08:51.329030222Z" level=info msg="connecting to shim 4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3" address="unix:///run/containerd/s/c7c8b8da15298b3d1e96399fd8b8c9b21d1ea1c8ce1909c89f781cd17293b27a" protocol=ttrpc version=3 Sep 12 18:08:51.361726 systemd[1]: Started cri-containerd-4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3.scope - libcontainer container 4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3. Sep 12 18:08:51.422768 containerd[1562]: time="2025-09-12T18:08:51.422699529Z" level=info msg="StartContainer for \"4fe4491f2b492eea4921d8c793554cc72f907e54f960e0248ee02cfa613094f3\" returns successfully" Sep 12 18:08:51.995728 containerd[1562]: time="2025-09-12T18:08:51.995671102Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-8hj6c,Uid:b51a8f2d-d1c4-4004-80da-1576b8879e25,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:52.157990 systemd-networkd[1450]: cali2a438d4f454: Link UP Sep 12 18:08:52.160466 systemd-networkd[1450]: cali2a438d4f454: Gained carrier Sep 12 18:08:52.179155 containerd[1562]: 2025-09-12 18:08:52.047 [INFO][4167] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0 csi-node-driver- calico-system b51a8f2d-d1c4-4004-80da-1576b8879e25 708 0 2025-09-12 18:08:28 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:856c6b598f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 csi-node-driver-8hj6c eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali2a438d4f454 [] [] }} ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-" Sep 12 18:08:52.179155 containerd[1562]: 2025-09-12 18:08:52.048 [INFO][4167] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.179155 containerd[1562]: 2025-09-12 18:08:52.091 [INFO][4178] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" HandleID="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.092 [INFO][4178] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" HandleID="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5640), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"csi-node-driver-8hj6c", "timestamp":"2025-09-12 18:08:52.091927344 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.092 [INFO][4178] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.092 [INFO][4178] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.092 [INFO][4178] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.103 [INFO][4178] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.113 [INFO][4178] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.122 [INFO][4178] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.125 [INFO][4178] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179526 containerd[1562]: 2025-09-12 18:08:52.129 [INFO][4178] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.129 [INFO][4178] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.132 [INFO][4178] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907 Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.138 [INFO][4178] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.146 [INFO][4178] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.2/26] block=192.168.6.0/26 handle="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.146 [INFO][4178] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.2/26] handle="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.146 [INFO][4178] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:52.179943 containerd[1562]: 2025-09-12 18:08:52.146 [INFO][4178] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.2/26] IPv6=[] ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" HandleID="k8s-pod-network.2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.180285 containerd[1562]: 2025-09-12 18:08:52.151 [INFO][4167] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b51a8f2d-d1c4-4004-80da-1576b8879e25", ResourceVersion:"708", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"856c6b598f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"csi-node-driver-8hj6c", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.6.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2a438d4f454", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:52.180415 containerd[1562]: 2025-09-12 18:08:52.152 [INFO][4167] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.2/32] ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.180415 containerd[1562]: 2025-09-12 18:08:52.152 [INFO][4167] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2a438d4f454 ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.180415 containerd[1562]: 2025-09-12 18:08:52.160 [INFO][4167] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.180586 containerd[1562]: 2025-09-12 18:08:52.162 [INFO][4167] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b51a8f2d-d1c4-4004-80da-1576b8879e25", ResourceVersion:"708", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"856c6b598f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907", Pod:"csi-node-driver-8hj6c", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.6.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2a438d4f454", MAC:"da:1f:aa:ab:2d:00", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:52.182155 containerd[1562]: 2025-09-12 18:08:52.174 [INFO][4167] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" Namespace="calico-system" Pod="csi-node-driver-8hj6c" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-csi--node--driver--8hj6c-eth0" Sep 12 18:08:52.219921 containerd[1562]: time="2025-09-12T18:08:52.219870233Z" level=info msg="connecting to shim 2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907" address="unix:///run/containerd/s/68c56be90e7162b8eaf44ad985c8d9a5fae52f37d66ad8695b6c0c591472c0ca" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:52.263690 systemd[1]: Started cri-containerd-2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907.scope - libcontainer container 2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907. Sep 12 18:08:52.269305 kubelet[2714]: I0912 18:08:52.269217 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-5885c4d4d6-vl2xj" podStartSLOduration=2.212796151 podStartE2EDuration="6.269196851s" podCreationTimestamp="2025-09-12 18:08:46 +0000 UTC" firstStartedPulling="2025-09-12 18:08:47.241282345 +0000 UTC m=+39.402142634" lastFinishedPulling="2025-09-12 18:08:51.297683036 +0000 UTC m=+43.458543334" observedRunningTime="2025-09-12 18:08:52.269036992 +0000 UTC m=+44.429897289" watchObservedRunningTime="2025-09-12 18:08:52.269196851 +0000 UTC m=+44.430057147" Sep 12 18:08:52.353196 containerd[1562]: time="2025-09-12T18:08:52.353138128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-8hj6c,Uid:b51a8f2d-d1c4-4004-80da-1576b8879e25,Namespace:calico-system,Attempt:0,} returns sandbox id \"2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907\"" Sep 12 18:08:52.355390 containerd[1562]: time="2025-09-12T18:08:52.355270099Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 12 18:08:52.994502 containerd[1562]: time="2025-09-12T18:08:52.994432553Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-2bd6k,Uid:b88db574-df8a-411c-a754-0d8df69ed036,Namespace:calico-apiserver,Attempt:0,}" Sep 12 18:08:52.995001 containerd[1562]: time="2025-09-12T18:08:52.994466078Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-44njx,Uid:8428e997-ec0b-47c0-8e5c-deb3401be527,Namespace:calico-apiserver,Attempt:0,}" Sep 12 18:08:53.217729 systemd-networkd[1450]: calib24ddbc2437: Link UP Sep 12 18:08:53.222195 systemd-networkd[1450]: calib24ddbc2437: Gained carrier Sep 12 18:08:53.253095 containerd[1562]: 2025-09-12 18:08:53.094 [INFO][4252] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0 calico-apiserver-6f9c78995- calico-apiserver b88db574-df8a-411c-a754-0d8df69ed036 828 0 2025-09-12 18:08:23 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6f9c78995 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 calico-apiserver-6f9c78995-2bd6k eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calib24ddbc2437 [] [] }} ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-" Sep 12 18:08:53.253095 containerd[1562]: 2025-09-12 18:08:53.095 [INFO][4252] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.253095 containerd[1562]: 2025-09-12 18:08:53.153 [INFO][4272] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" HandleID="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.153 [INFO][4272] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" HandleID="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000305b90), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"calico-apiserver-6f9c78995-2bd6k", "timestamp":"2025-09-12 18:08:53.152934523 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.153 [INFO][4272] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.153 [INFO][4272] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.153 [INFO][4272] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.165 [INFO][4272] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.173 [INFO][4272] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.180 [INFO][4272] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.183 [INFO][4272] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.253439 containerd[1562]: 2025-09-12 18:08:53.187 [INFO][4272] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.187 [INFO][4272] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.190 [INFO][4272] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49 Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.199 [INFO][4272] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.207 [INFO][4272] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.3/26] block=192.168.6.0/26 handle="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.208 [INFO][4272] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.3/26] handle="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.208 [INFO][4272] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:53.254962 containerd[1562]: 2025-09-12 18:08:53.208 [INFO][4272] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.3/26] IPv6=[] ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" HandleID="k8s-pod-network.afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.255207 containerd[1562]: 2025-09-12 18:08:53.212 [INFO][4252] cni-plugin/k8s.go 418: Populated endpoint ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0", GenerateName:"calico-apiserver-6f9c78995-", Namespace:"calico-apiserver", SelfLink:"", UID:"b88db574-df8a-411c-a754-0d8df69ed036", ResourceVersion:"828", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6f9c78995", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"calico-apiserver-6f9c78995-2bd6k", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.6.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calib24ddbc2437", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:53.255289 containerd[1562]: 2025-09-12 18:08:53.212 [INFO][4252] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.3/32] ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.255289 containerd[1562]: 2025-09-12 18:08:53.212 [INFO][4252] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib24ddbc2437 ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.255289 containerd[1562]: 2025-09-12 18:08:53.228 [INFO][4252] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.255362 containerd[1562]: 2025-09-12 18:08:53.229 [INFO][4252] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0", GenerateName:"calico-apiserver-6f9c78995-", Namespace:"calico-apiserver", SelfLink:"", UID:"b88db574-df8a-411c-a754-0d8df69ed036", ResourceVersion:"828", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6f9c78995", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49", Pod:"calico-apiserver-6f9c78995-2bd6k", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.6.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calib24ddbc2437", MAC:"ce:0b:c7:ea:7d:6f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:53.255842 containerd[1562]: 2025-09-12 18:08:53.247 [INFO][4252] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-2bd6k" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--2bd6k-eth0" Sep 12 18:08:53.339862 containerd[1562]: time="2025-09-12T18:08:53.339790206Z" level=info msg="connecting to shim afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49" address="unix:///run/containerd/s/72da7491329d11f0717360e9ea043692ce901bfaca0620ed2998d6f47b840685" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:53.395032 systemd[1]: Started cri-containerd-afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49.scope - libcontainer container afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49. Sep 12 18:08:53.400300 systemd-networkd[1450]: cali04975e03855: Link UP Sep 12 18:08:53.403437 systemd-networkd[1450]: cali04975e03855: Gained carrier Sep 12 18:08:53.442636 containerd[1562]: 2025-09-12 18:08:53.100 [INFO][4244] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0 calico-apiserver-6f9c78995- calico-apiserver 8428e997-ec0b-47c0-8e5c-deb3401be527 826 0 2025-09-12 18:08:23 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6f9c78995 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 calico-apiserver-6f9c78995-44njx eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali04975e03855 [] [] }} ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-" Sep 12 18:08:53.442636 containerd[1562]: 2025-09-12 18:08:53.101 [INFO][4244] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.442636 containerd[1562]: 2025-09-12 18:08:53.168 [INFO][4277] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" HandleID="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.168 [INFO][4277] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" HandleID="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5950), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"calico-apiserver-6f9c78995-44njx", "timestamp":"2025-09-12 18:08:53.168363856 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.168 [INFO][4277] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.208 [INFO][4277] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.208 [INFO][4277] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.269 [INFO][4277] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.294 [INFO][4277] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.323 [INFO][4277] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.341 [INFO][4277] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.443380 containerd[1562]: 2025-09-12 18:08:53.347 [INFO][4277] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.347 [INFO][4277] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.350 [INFO][4277] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46 Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.364 [INFO][4277] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.381 [INFO][4277] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.4/26] block=192.168.6.0/26 handle="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.381 [INFO][4277] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.4/26] handle="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.381 [INFO][4277] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:53.444585 containerd[1562]: 2025-09-12 18:08:53.381 [INFO][4277] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.4/26] IPv6=[] ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" HandleID="k8s-pod-network.2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.445090 containerd[1562]: 2025-09-12 18:08:53.387 [INFO][4244] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0", GenerateName:"calico-apiserver-6f9c78995-", Namespace:"calico-apiserver", SelfLink:"", UID:"8428e997-ec0b-47c0-8e5c-deb3401be527", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6f9c78995", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"calico-apiserver-6f9c78995-44njx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.6.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali04975e03855", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:53.445269 containerd[1562]: 2025-09-12 18:08:53.389 [INFO][4244] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.4/32] ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.445269 containerd[1562]: 2025-09-12 18:08:53.390 [INFO][4244] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali04975e03855 ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.445269 containerd[1562]: 2025-09-12 18:08:53.404 [INFO][4244] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.446791 containerd[1562]: 2025-09-12 18:08:53.405 [INFO][4244] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0", GenerateName:"calico-apiserver-6f9c78995-", Namespace:"calico-apiserver", SelfLink:"", UID:"8428e997-ec0b-47c0-8e5c-deb3401be527", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6f9c78995", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46", Pod:"calico-apiserver-6f9c78995-44njx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.6.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali04975e03855", MAC:"02:b1:34:a9:80:89", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:53.448858 containerd[1562]: 2025-09-12 18:08:53.430 [INFO][4244] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" Namespace="calico-apiserver" Pod="calico-apiserver-6f9c78995-44njx" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--apiserver--6f9c78995--44njx-eth0" Sep 12 18:08:53.483293 containerd[1562]: time="2025-09-12T18:08:53.483225982Z" level=info msg="connecting to shim 2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46" address="unix:///run/containerd/s/69cdfa2435154fde25eded6370db6eb4996004032deec5dfee7f7af3ab65873d" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:53.525806 systemd[1]: Started cri-containerd-2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46.scope - libcontainer container 2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46. Sep 12 18:08:53.553234 containerd[1562]: time="2025-09-12T18:08:53.553153679Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-2bd6k,Uid:b88db574-df8a-411c-a754-0d8df69ed036,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49\"" Sep 12 18:08:53.610724 containerd[1562]: time="2025-09-12T18:08:53.610583609Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6f9c78995-44njx,Uid:8428e997-ec0b-47c0-8e5c-deb3401be527,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46\"" Sep 12 18:08:53.733772 systemd-networkd[1450]: cali2a438d4f454: Gained IPv6LL Sep 12 18:08:53.996571 kubelet[2714]: E0912 18:08:53.995202 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:54.003044 containerd[1562]: time="2025-09-12T18:08:54.002980896Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-855d7c9bb7-bjgr8,Uid:5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:54.007327 containerd[1562]: time="2025-09-12T18:08:54.005977596Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-zncc7,Uid:7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:54.407822 containerd[1562]: time="2025-09-12T18:08:54.407685347Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:54.410196 containerd[1562]: time="2025-09-12T18:08:54.410112889Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Sep 12 18:08:54.419093 containerd[1562]: time="2025-09-12T18:08:54.418007686Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:54.435424 containerd[1562]: time="2025-09-12T18:08:54.435352495Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:54.440247 containerd[1562]: time="2025-09-12T18:08:54.440176273Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 2.084856856s" Sep 12 18:08:54.440247 containerd[1562]: time="2025-09-12T18:08:54.440244621Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Sep 12 18:08:54.447084 containerd[1562]: time="2025-09-12T18:08:54.447011400Z" level=info msg="CreateContainer within sandbox \"2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 12 18:08:54.449276 containerd[1562]: time="2025-09-12T18:08:54.448936111Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 12 18:08:54.462971 systemd-networkd[1450]: cali7ce76fe3512: Link UP Sep 12 18:08:54.469246 systemd-networkd[1450]: cali7ce76fe3512: Gained carrier Sep 12 18:08:54.502529 containerd[1562]: time="2025-09-12T18:08:54.501800198Z" level=info msg="Container 90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:54.516075 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3903904059.mount: Deactivated successfully. Sep 12 18:08:54.531031 containerd[1562]: 2025-09-12 18:08:54.188 [INFO][4407] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0 coredns-7c65d6cfc9- kube-system 7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f 825 0 2025-09-12 18:08:13 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 coredns-7c65d6cfc9-zncc7 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali7ce76fe3512 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-" Sep 12 18:08:54.531031 containerd[1562]: 2025-09-12 18:08:54.189 [INFO][4407] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.531031 containerd[1562]: 2025-09-12 18:08:54.302 [INFO][4432] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" HandleID="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.302 [INFO][4432] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" HandleID="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5220), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"coredns-7c65d6cfc9-zncc7", "timestamp":"2025-09-12 18:08:54.302226213 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.302 [INFO][4432] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.302 [INFO][4432] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.302 [INFO][4432] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.334 [INFO][4432] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.352 [INFO][4432] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.371 [INFO][4432] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.378 [INFO][4432] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.531659 containerd[1562]: 2025-09-12 18:08:54.384 [INFO][4432] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.384 [INFO][4432] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.390 [INFO][4432] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2 Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.402 [INFO][4432] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.433 [INFO][4432] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.5/26] block=192.168.6.0/26 handle="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.433 [INFO][4432] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.5/26] handle="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.433 [INFO][4432] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:54.534899 containerd[1562]: 2025-09-12 18:08:54.433 [INFO][4432] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.5/26] IPv6=[] ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" HandleID="k8s-pod-network.b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.535610 containerd[1562]: 2025-09-12 18:08:54.450 [INFO][4407] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"coredns-7c65d6cfc9-zncc7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.6.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7ce76fe3512", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:54.535610 containerd[1562]: 2025-09-12 18:08:54.453 [INFO][4407] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.5/32] ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.535610 containerd[1562]: 2025-09-12 18:08:54.454 [INFO][4407] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7ce76fe3512 ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.535610 containerd[1562]: 2025-09-12 18:08:54.473 [INFO][4407] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.535610 containerd[1562]: 2025-09-12 18:08:54.481 [INFO][4407] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2", Pod:"coredns-7c65d6cfc9-zncc7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.6.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7ce76fe3512", MAC:"9e:80:3e:99:33:38", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:54.535610 containerd[1562]: 2025-09-12 18:08:54.515 [INFO][4407] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-zncc7" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--zncc7-eth0" Sep 12 18:08:54.538318 containerd[1562]: time="2025-09-12T18:08:54.538017579Z" level=info msg="CreateContainer within sandbox \"2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251\"" Sep 12 18:08:54.540350 containerd[1562]: time="2025-09-12T18:08:54.540185343Z" level=info msg="StartContainer for \"90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251\"" Sep 12 18:08:54.546054 containerd[1562]: time="2025-09-12T18:08:54.545996222Z" level=info msg="connecting to shim 90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251" address="unix:///run/containerd/s/68c56be90e7162b8eaf44ad985c8d9a5fae52f37d66ad8695b6c0c591472c0ca" protocol=ttrpc version=3 Sep 12 18:08:54.629886 systemd[1]: Started cri-containerd-90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251.scope - libcontainer container 90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251. Sep 12 18:08:54.669139 systemd-networkd[1450]: calid8ef394a159: Link UP Sep 12 18:08:54.676672 containerd[1562]: time="2025-09-12T18:08:54.674671211Z" level=info msg="connecting to shim b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2" address="unix:///run/containerd/s/634c78eb6a14f71b06f833101e4d1ebcc6d664236ac8e7908822bfc0023e5173" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:54.674744 systemd-networkd[1450]: calid8ef394a159: Gained carrier Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.191 [INFO][4406] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0 calico-kube-controllers-855d7c9bb7- calico-system 5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1 827 0 2025-09-12 18:08:28 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:855d7c9bb7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 calico-kube-controllers-855d7c9bb7-bjgr8 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calid8ef394a159 [] [] }} ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.193 [INFO][4406] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.367 [INFO][4433] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" HandleID="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.367 [INFO][4433] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" HandleID="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000333510), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"calico-kube-controllers-855d7c9bb7-bjgr8", "timestamp":"2025-09-12 18:08:54.36719251 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.367 [INFO][4433] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.433 [INFO][4433] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.434 [INFO][4433] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.468 [INFO][4433] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.495 [INFO][4433] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.539 [INFO][4433] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.549 [INFO][4433] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.575 [INFO][4433] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.575 [INFO][4433] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.586 [INFO][4433] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00 Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.599 [INFO][4433] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.629 [INFO][4433] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.6/26] block=192.168.6.0/26 handle="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.631 [INFO][4433] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.6/26] handle="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.631 [INFO][4433] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:54.745830 containerd[1562]: 2025-09-12 18:08:54.631 [INFO][4433] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.6/26] IPv6=[] ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" HandleID="k8s-pod-network.1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.749463 containerd[1562]: 2025-09-12 18:08:54.645 [INFO][4406] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0", GenerateName:"calico-kube-controllers-855d7c9bb7-", Namespace:"calico-system", SelfLink:"", UID:"5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1", ResourceVersion:"827", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"855d7c9bb7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"calico-kube-controllers-855d7c9bb7-bjgr8", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.6.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calid8ef394a159", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:54.749463 containerd[1562]: 2025-09-12 18:08:54.648 [INFO][4406] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.6/32] ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.749463 containerd[1562]: 2025-09-12 18:08:54.650 [INFO][4406] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid8ef394a159 ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.749463 containerd[1562]: 2025-09-12 18:08:54.677 [INFO][4406] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.749463 containerd[1562]: 2025-09-12 18:08:54.680 [INFO][4406] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0", GenerateName:"calico-kube-controllers-855d7c9bb7-", Namespace:"calico-system", SelfLink:"", UID:"5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1", ResourceVersion:"827", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"855d7c9bb7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00", Pod:"calico-kube-controllers-855d7c9bb7-bjgr8", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.6.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calid8ef394a159", MAC:"82:b2:b0:ea:f6:e1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:54.749463 containerd[1562]: 2025-09-12 18:08:54.725 [INFO][4406] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" Namespace="calico-system" Pod="calico-kube-controllers-855d7c9bb7-bjgr8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-calico--kube--controllers--855d7c9bb7--bjgr8-eth0" Sep 12 18:08:54.800128 systemd[1]: Started cri-containerd-b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2.scope - libcontainer container b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2. Sep 12 18:08:54.809041 containerd[1562]: time="2025-09-12T18:08:54.808976080Z" level=info msg="connecting to shim 1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00" address="unix:///run/containerd/s/4165bd43384c8f7e6d7fcbf0b859412051cfd72ebb71909d06a3065809cd1c47" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:54.821788 systemd-networkd[1450]: calib24ddbc2437: Gained IPv6LL Sep 12 18:08:54.822333 systemd-networkd[1450]: cali04975e03855: Gained IPv6LL Sep 12 18:08:54.884353 systemd[1]: Started cri-containerd-1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00.scope - libcontainer container 1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00. Sep 12 18:08:54.950041 containerd[1562]: time="2025-09-12T18:08:54.949866772Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-zncc7,Uid:7dcd9f5e-1b7c-4c5f-90ce-0fb0ba14636f,Namespace:kube-system,Attempt:0,} returns sandbox id \"b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2\"" Sep 12 18:08:54.951768 kubelet[2714]: E0912 18:08:54.951715 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:54.956578 containerd[1562]: time="2025-09-12T18:08:54.956516995Z" level=info msg="CreateContainer within sandbox \"b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 18:08:54.969461 containerd[1562]: time="2025-09-12T18:08:54.969319730Z" level=info msg="Container debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:54.976512 containerd[1562]: time="2025-09-12T18:08:54.976309685Z" level=info msg="CreateContainer within sandbox \"b766bacfb424d0d53ee622fd1ffd13b669420c18c8e06077c09962f0c5aad9a2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88\"" Sep 12 18:08:54.978874 containerd[1562]: time="2025-09-12T18:08:54.978815348Z" level=info msg="StartContainer for \"debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88\"" Sep 12 18:08:54.980255 containerd[1562]: time="2025-09-12T18:08:54.980216225Z" level=info msg="connecting to shim debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88" address="unix:///run/containerd/s/634c78eb6a14f71b06f833101e4d1ebcc6d664236ac8e7908822bfc0023e5173" protocol=ttrpc version=3 Sep 12 18:08:54.996511 containerd[1562]: time="2025-09-12T18:08:54.996437943Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-dsgd8,Uid:08f7765b-0da5-414e-bf81-140d051267cd,Namespace:calico-system,Attempt:0,}" Sep 12 18:08:54.997657 kubelet[2714]: E0912 18:08:54.996736 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:54.999607 containerd[1562]: time="2025-09-12T18:08:54.999553915Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-5wflb,Uid:af0fd070-cb87-4036-9733-c0c1bf7001d7,Namespace:kube-system,Attempt:0,}" Sep 12 18:08:55.125761 containerd[1562]: time="2025-09-12T18:08:55.125293034Z" level=info msg="StartContainer for \"90c8adae0702c40c5c2c879f59f6a223013ddf52e9312b521a63933350484251\" returns successfully" Sep 12 18:08:55.163344 systemd[1]: Started cri-containerd-debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88.scope - libcontainer container debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88. Sep 12 18:08:55.212762 containerd[1562]: time="2025-09-12T18:08:55.211559020Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-855d7c9bb7-bjgr8,Uid:5b5d2ce6-01c4-4e9f-921d-be7a721ef2d1,Namespace:calico-system,Attempt:0,} returns sandbox id \"1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00\"" Sep 12 18:08:55.312643 containerd[1562]: time="2025-09-12T18:08:55.312599949Z" level=info msg="StartContainer for \"debeef6c6043a6c080b3a1155bb2e826fb7d63a83d345a274b51da4d00033f88\" returns successfully" Sep 12 18:08:55.490248 systemd-networkd[1450]: calif7a6b934b3d: Link UP Sep 12 18:08:55.495512 systemd-networkd[1450]: calif7a6b934b3d: Gained carrier Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.273 [INFO][4584] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0 coredns-7c65d6cfc9- kube-system af0fd070-cb87-4036-9733-c0c1bf7001d7 829 0 2025-09-12 18:08:13 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 coredns-7c65d6cfc9-5wflb eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calif7a6b934b3d [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.278 [INFO][4584] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.374 [INFO][4641] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" HandleID="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.377 [INFO][4641] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" HandleID="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003239c0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"coredns-7c65d6cfc9-5wflb", "timestamp":"2025-09-12 18:08:55.374792805 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.377 [INFO][4641] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.377 [INFO][4641] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.377 [INFO][4641] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.398 [INFO][4641] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.410 [INFO][4641] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.431 [INFO][4641] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.439 [INFO][4641] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.444 [INFO][4641] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.444 [INFO][4641] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.451 [INFO][4641] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3 Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.460 [INFO][4641] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.475 [INFO][4641] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.7/26] block=192.168.6.0/26 handle="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.475 [INFO][4641] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.7/26] handle="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.475 [INFO][4641] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:55.526386 containerd[1562]: 2025-09-12 18:08:55.475 [INFO][4641] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.7/26] IPv6=[] ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" HandleID="k8s-pod-network.4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.528040 containerd[1562]: 2025-09-12 18:08:55.481 [INFO][4584] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"af0fd070-cb87-4036-9733-c0c1bf7001d7", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"coredns-7c65d6cfc9-5wflb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.6.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif7a6b934b3d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:55.528040 containerd[1562]: 2025-09-12 18:08:55.482 [INFO][4584] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.7/32] ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.528040 containerd[1562]: 2025-09-12 18:08:55.482 [INFO][4584] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif7a6b934b3d ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.528040 containerd[1562]: 2025-09-12 18:08:55.487 [INFO][4584] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.528040 containerd[1562]: 2025-09-12 18:08:55.488 [INFO][4584] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"af0fd070-cb87-4036-9733-c0c1bf7001d7", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3", Pod:"coredns-7c65d6cfc9-5wflb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.6.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif7a6b934b3d", MAC:"2a:f0:ef:ce:fe:7a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:55.528040 containerd[1562]: 2025-09-12 18:08:55.518 [INFO][4584] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" Namespace="kube-system" Pod="coredns-7c65d6cfc9-5wflb" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-coredns--7c65d6cfc9--5wflb-eth0" Sep 12 18:08:55.624515 containerd[1562]: time="2025-09-12T18:08:55.624405829Z" level=info msg="connecting to shim 4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3" address="unix:///run/containerd/s/861ad05dd0a180ebfad382f41eb6a98c52d902ba26a89bf09b1561197a370757" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:55.685970 systemd[1]: Started cri-containerd-4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3.scope - libcontainer container 4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3. Sep 12 18:08:55.695230 systemd-networkd[1450]: cali3ace9c0b086: Link UP Sep 12 18:08:55.697573 systemd-networkd[1450]: cali3ace9c0b086: Gained carrier Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.316 [INFO][4575] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0 goldmane-7988f88666- calico-system 08f7765b-0da5-414e-bf81-140d051267cd 830 0 2025-09-12 18:08:27 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:7988f88666 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4426.1.0-c-cd41a2aea4 goldmane-7988f88666-dsgd8 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali3ace9c0b086 [] [] }} ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.322 [INFO][4575] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.386 [INFO][4650] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" HandleID="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.387 [INFO][4650] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" HandleID="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5140), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4426.1.0-c-cd41a2aea4", "pod":"goldmane-7988f88666-dsgd8", "timestamp":"2025-09-12 18:08:55.386957061 +0000 UTC"}, Hostname:"ci-4426.1.0-c-cd41a2aea4", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.387 [INFO][4650] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.475 [INFO][4650] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.476 [INFO][4650] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4426.1.0-c-cd41a2aea4' Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.503 [INFO][4650] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.519 [INFO][4650] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.549 [INFO][4650] ipam/ipam.go 511: Trying affinity for 192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.559 [INFO][4650] ipam/ipam.go 158: Attempting to load block cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.580 [INFO][4650] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.6.0/26 host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.580 [INFO][4650] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.6.0/26 handle="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.599 [INFO][4650] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.632 [INFO][4650] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.6.0/26 handle="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.666 [INFO][4650] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.6.8/26] block=192.168.6.0/26 handle="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.667 [INFO][4650] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.6.8/26] handle="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" host="ci-4426.1.0-c-cd41a2aea4" Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.667 [INFO][4650] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 18:08:55.723600 containerd[1562]: 2025-09-12 18:08:55.668 [INFO][4650] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.6.8/26] IPv6=[] ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" HandleID="k8s-pod-network.a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Workload="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.724291 containerd[1562]: 2025-09-12 18:08:55.685 [INFO][4575] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0", GenerateName:"goldmane-7988f88666-", Namespace:"calico-system", SelfLink:"", UID:"08f7765b-0da5-414e-bf81-140d051267cd", ResourceVersion:"830", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 27, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7988f88666", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"", Pod:"goldmane-7988f88666-dsgd8", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.6.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3ace9c0b086", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:55.724291 containerd[1562]: 2025-09-12 18:08:55.686 [INFO][4575] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.6.8/32] ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.724291 containerd[1562]: 2025-09-12 18:08:55.686 [INFO][4575] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3ace9c0b086 ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.724291 containerd[1562]: 2025-09-12 18:08:55.698 [INFO][4575] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.724291 containerd[1562]: 2025-09-12 18:08:55.699 [INFO][4575] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0", GenerateName:"goldmane-7988f88666-", Namespace:"calico-system", SelfLink:"", UID:"08f7765b-0da5-414e-bf81-140d051267cd", ResourceVersion:"830", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 18, 8, 27, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7988f88666", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4426.1.0-c-cd41a2aea4", ContainerID:"a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb", Pod:"goldmane-7988f88666-dsgd8", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.6.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3ace9c0b086", MAC:"be:fb:e0:a7:50:fa", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 18:08:55.724291 containerd[1562]: 2025-09-12 18:08:55.717 [INFO][4575] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" Namespace="calico-system" Pod="goldmane-7988f88666-dsgd8" WorkloadEndpoint="ci--4426.1.0--c--cd41a2aea4-k8s-goldmane--7988f88666--dsgd8-eth0" Sep 12 18:08:55.774502 containerd[1562]: time="2025-09-12T18:08:55.773547955Z" level=info msg="connecting to shim a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb" address="unix:///run/containerd/s/db21a1dd861e010fcda65d5392fb2a5227cb1913ef98a65b105f869b88a16f5e" namespace=k8s.io protocol=ttrpc version=3 Sep 12 18:08:55.835770 systemd[1]: Started cri-containerd-a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb.scope - libcontainer container a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb. Sep 12 18:08:55.847866 systemd-networkd[1450]: calid8ef394a159: Gained IPv6LL Sep 12 18:08:55.926554 containerd[1562]: time="2025-09-12T18:08:55.925188604Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-5wflb,Uid:af0fd070-cb87-4036-9733-c0c1bf7001d7,Namespace:kube-system,Attempt:0,} returns sandbox id \"4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3\"" Sep 12 18:08:55.934207 kubelet[2714]: E0912 18:08:55.934161 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:55.955682 containerd[1562]: time="2025-09-12T18:08:55.955442565Z" level=info msg="CreateContainer within sandbox \"4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 18:08:55.967470 containerd[1562]: time="2025-09-12T18:08:55.966340110Z" level=info msg="Container 2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:56.024744 containerd[1562]: time="2025-09-12T18:08:56.024559337Z" level=info msg="CreateContainer within sandbox \"4bf2996aeac1d16f2b536cb71ab1d07e1fc69b67d87203e7a58ebc0bd05a77c3\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26\"" Sep 12 18:08:56.030824 containerd[1562]: time="2025-09-12T18:08:56.030121677Z" level=info msg="StartContainer for \"2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26\"" Sep 12 18:08:56.034501 containerd[1562]: time="2025-09-12T18:08:56.033973129Z" level=info msg="connecting to shim 2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26" address="unix:///run/containerd/s/861ad05dd0a180ebfad382f41eb6a98c52d902ba26a89bf09b1561197a370757" protocol=ttrpc version=3 Sep 12 18:08:56.065950 containerd[1562]: time="2025-09-12T18:08:56.065884952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-dsgd8,Uid:08f7765b-0da5-414e-bf81-140d051267cd,Namespace:calico-system,Attempt:0,} returns sandbox id \"a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb\"" Sep 12 18:08:56.090758 systemd[1]: Started cri-containerd-2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26.scope - libcontainer container 2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26. Sep 12 18:08:56.160125 containerd[1562]: time="2025-09-12T18:08:56.160051294Z" level=info msg="StartContainer for \"2718f4acc5ae2e4d75303d1f801ff239c8bec921e0e98a3555c0f3acdd75ee26\" returns successfully" Sep 12 18:08:56.229968 systemd-networkd[1450]: cali7ce76fe3512: Gained IPv6LL Sep 12 18:08:56.334994 kubelet[2714]: E0912 18:08:56.334226 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:56.347481 kubelet[2714]: E0912 18:08:56.347071 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:56.387785 kubelet[2714]: I0912 18:08:56.387694 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-zncc7" podStartSLOduration=43.387660403 podStartE2EDuration="43.387660403s" podCreationTimestamp="2025-09-12 18:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 18:08:56.380209478 +0000 UTC m=+48.541069777" watchObservedRunningTime="2025-09-12 18:08:56.387660403 +0000 UTC m=+48.548520732" Sep 12 18:08:56.399883 kubelet[2714]: I0912 18:08:56.399515 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-5wflb" podStartSLOduration=43.399491595 podStartE2EDuration="43.399491595s" podCreationTimestamp="2025-09-12 18:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 18:08:56.399292634 +0000 UTC m=+48.560152948" watchObservedRunningTime="2025-09-12 18:08:56.399491595 +0000 UTC m=+48.560351884" Sep 12 18:08:56.805796 systemd-networkd[1450]: calif7a6b934b3d: Gained IPv6LL Sep 12 18:08:56.999326 systemd-networkd[1450]: cali3ace9c0b086: Gained IPv6LL Sep 12 18:08:57.347678 kubelet[2714]: E0912 18:08:57.347505 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:57.348674 kubelet[2714]: E0912 18:08:57.348320 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:57.644871 systemd[1]: Started sshd@8-161.35.232.141:22-45.79.181.251:11686.service - OpenSSH per-connection server daemon (45.79.181.251:11686). Sep 12 18:08:57.752398 containerd[1562]: time="2025-09-12T18:08:57.752334502Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:57.753658 containerd[1562]: time="2025-09-12T18:08:57.752782583Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Sep 12 18:08:57.754522 containerd[1562]: time="2025-09-12T18:08:57.754263652Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:57.757611 containerd[1562]: time="2025-09-12T18:08:57.756678485Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:57.757611 containerd[1562]: time="2025-09-12T18:08:57.757380830Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 3.308135025s" Sep 12 18:08:57.757611 containerd[1562]: time="2025-09-12T18:08:57.757442993Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 12 18:08:57.759238 containerd[1562]: time="2025-09-12T18:08:57.759211393Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 12 18:08:57.763608 containerd[1562]: time="2025-09-12T18:08:57.763563797Z" level=info msg="CreateContainer within sandbox \"afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 12 18:08:57.771911 containerd[1562]: time="2025-09-12T18:08:57.771857091Z" level=info msg="Container 0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:57.787447 containerd[1562]: time="2025-09-12T18:08:57.787375830Z" level=info msg="CreateContainer within sandbox \"afa40bac1bac2d3ffc24c1dcd5d3bac7705bb6d4919caf457f3a5b327f946e49\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0\"" Sep 12 18:08:57.789865 containerd[1562]: time="2025-09-12T18:08:57.789809981Z" level=info msg="StartContainer for \"0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0\"" Sep 12 18:08:57.792949 containerd[1562]: time="2025-09-12T18:08:57.792843801Z" level=info msg="connecting to shim 0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0" address="unix:///run/containerd/s/72da7491329d11f0717360e9ea043692ce901bfaca0620ed2998d6f47b840685" protocol=ttrpc version=3 Sep 12 18:08:57.831874 systemd[1]: Started cri-containerd-0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0.scope - libcontainer container 0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0. Sep 12 18:08:57.909106 containerd[1562]: time="2025-09-12T18:08:57.908565436Z" level=info msg="StartContainer for \"0f69ec793c45816d1b76d1049182ed8b58776a96cbf61f09456e64e0a4b810b0\" returns successfully" Sep 12 18:08:58.176188 containerd[1562]: time="2025-09-12T18:08:58.175999611Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:08:58.177551 containerd[1562]: time="2025-09-12T18:08:58.177521310Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Sep 12 18:08:58.178800 containerd[1562]: time="2025-09-12T18:08:58.178672986Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 419.149543ms" Sep 12 18:08:58.178800 containerd[1562]: time="2025-09-12T18:08:58.178763791Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 12 18:08:58.180027 containerd[1562]: time="2025-09-12T18:08:58.179857503Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 12 18:08:58.189558 containerd[1562]: time="2025-09-12T18:08:58.189177356Z" level=info msg="CreateContainer within sandbox \"2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 12 18:08:58.197348 containerd[1562]: time="2025-09-12T18:08:58.196824553Z" level=info msg="Container a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:08:58.206046 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount991634325.mount: Deactivated successfully. Sep 12 18:08:58.211960 containerd[1562]: time="2025-09-12T18:08:58.211860049Z" level=info msg="CreateContainer within sandbox \"2d18db00cbe64e22c4a5e42847dd86e003508a669dfb2296a6e5f8d2c5a4aa46\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7\"" Sep 12 18:08:58.213636 containerd[1562]: time="2025-09-12T18:08:58.213596903Z" level=info msg="StartContainer for \"a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7\"" Sep 12 18:08:58.215714 containerd[1562]: time="2025-09-12T18:08:58.215629259Z" level=info msg="connecting to shim a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7" address="unix:///run/containerd/s/69cdfa2435154fde25eded6370db6eb4996004032deec5dfee7f7af3ab65873d" protocol=ttrpc version=3 Sep 12 18:08:58.243717 systemd[1]: Started cri-containerd-a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7.scope - libcontainer container a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7. Sep 12 18:08:58.339181 containerd[1562]: time="2025-09-12T18:08:58.338452943Z" level=info msg="StartContainer for \"a7ed2d4e13f02dedc17fe44ef4dbd0ed610364eceb2b1eccac4c4f5d858714a7\" returns successfully" Sep 12 18:08:58.362552 kubelet[2714]: E0912 18:08:58.361930 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:58.364501 kubelet[2714]: E0912 18:08:58.363724 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:08:58.410044 kubelet[2714]: I0912 18:08:58.409879 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6f9c78995-44njx" podStartSLOduration=30.846030154 podStartE2EDuration="35.40985801s" podCreationTimestamp="2025-09-12 18:08:23 +0000 UTC" firstStartedPulling="2025-09-12 18:08:53.615679068 +0000 UTC m=+45.776539360" lastFinishedPulling="2025-09-12 18:08:58.179506931 +0000 UTC m=+50.340367216" observedRunningTime="2025-09-12 18:08:58.383366757 +0000 UTC m=+50.544227054" watchObservedRunningTime="2025-09-12 18:08:58.40985801 +0000 UTC m=+50.570718307" Sep 12 18:08:58.411381 kubelet[2714]: I0912 18:08:58.411326 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6f9c78995-2bd6k" podStartSLOduration=31.209083888 podStartE2EDuration="35.411308875s" podCreationTimestamp="2025-09-12 18:08:23 +0000 UTC" firstStartedPulling="2025-09-12 18:08:53.556599093 +0000 UTC m=+45.717459383" lastFinishedPulling="2025-09-12 18:08:57.758824093 +0000 UTC m=+49.919684370" observedRunningTime="2025-09-12 18:08:58.411129031 +0000 UTC m=+50.571989320" watchObservedRunningTime="2025-09-12 18:08:58.411308875 +0000 UTC m=+50.572169172" Sep 12 18:08:58.488582 sshd[4821]: Connection closed by 45.79.181.251 port 11686 [preauth] Sep 12 18:08:58.494150 systemd[1]: sshd@8-161.35.232.141:22-45.79.181.251:11686.service: Deactivated successfully. Sep 12 18:08:58.589593 systemd[1]: Started sshd@9-161.35.232.141:22-45.79.181.251:11702.service - OpenSSH per-connection server daemon (45.79.181.251:11702). Sep 12 18:08:59.376503 kubelet[2714]: I0912 18:08:59.374966 2714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 18:08:59.376503 kubelet[2714]: I0912 18:08:59.374966 2714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 18:08:59.577833 sshd[4902]: Connection closed by 45.79.181.251 port 11702 [preauth] Sep 12 18:08:59.581933 systemd[1]: sshd@9-161.35.232.141:22-45.79.181.251:11702.service: Deactivated successfully. Sep 12 18:08:59.667469 systemd[1]: Started sshd@10-161.35.232.141:22-45.79.181.251:43220.service - OpenSSH per-connection server daemon (45.79.181.251:43220). Sep 12 18:09:00.468096 sshd[4920]: Connection closed by 45.79.181.251 port 43220 [preauth] Sep 12 18:09:00.472206 systemd[1]: sshd@10-161.35.232.141:22-45.79.181.251:43220.service: Deactivated successfully. Sep 12 18:09:00.536853 containerd[1562]: time="2025-09-12T18:09:00.535955731Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:00.537532 containerd[1562]: time="2025-09-12T18:09:00.537489181Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Sep 12 18:09:00.540549 containerd[1562]: time="2025-09-12T18:09:00.540504609Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:00.545167 containerd[1562]: time="2025-09-12T18:09:00.545115122Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:00.546308 containerd[1562]: time="2025-09-12T18:09:00.546174278Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 2.366141633s" Sep 12 18:09:00.546308 containerd[1562]: time="2025-09-12T18:09:00.546213467Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Sep 12 18:09:00.550813 containerd[1562]: time="2025-09-12T18:09:00.550760611Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 12 18:09:00.554187 containerd[1562]: time="2025-09-12T18:09:00.554063636Z" level=info msg="CreateContainer within sandbox \"2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 12 18:09:00.574645 containerd[1562]: time="2025-09-12T18:09:00.574595417Z" level=info msg="Container 3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:09:00.591552 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2341736599.mount: Deactivated successfully. Sep 12 18:09:00.614498 containerd[1562]: time="2025-09-12T18:09:00.613336208Z" level=info msg="CreateContainer within sandbox \"2a49d6a077f2e562beb8038f682cfe39cfa31c4a319946a0c218b1f92f1c6907\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f\"" Sep 12 18:09:00.616428 containerd[1562]: time="2025-09-12T18:09:00.616243056Z" level=info msg="StartContainer for \"3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f\"" Sep 12 18:09:00.621517 containerd[1562]: time="2025-09-12T18:09:00.621217797Z" level=info msg="connecting to shim 3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f" address="unix:///run/containerd/s/68c56be90e7162b8eaf44ad985c8d9a5fae52f37d66ad8695b6c0c591472c0ca" protocol=ttrpc version=3 Sep 12 18:09:00.682837 systemd[1]: Started cri-containerd-3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f.scope - libcontainer container 3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f. Sep 12 18:09:00.746741 containerd[1562]: time="2025-09-12T18:09:00.745175021Z" level=info msg="StartContainer for \"3ed00875a40298252258d72d00e7954e3497ac3dab6a7aaa5b70da32af3e646f\" returns successfully" Sep 12 18:09:01.229114 kubelet[2714]: I0912 18:09:01.229060 2714 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 12 18:09:01.232101 kubelet[2714]: I0912 18:09:01.232053 2714 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 12 18:09:04.726952 containerd[1562]: time="2025-09-12T18:09:04.726848768Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:04.728496 containerd[1562]: time="2025-09-12T18:09:04.728423065Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Sep 12 18:09:04.729690 containerd[1562]: time="2025-09-12T18:09:04.729635330Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:04.733187 containerd[1562]: time="2025-09-12T18:09:04.733136134Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:04.734903 containerd[1562]: time="2025-09-12T18:09:04.734733967Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 4.181310581s" Sep 12 18:09:04.734903 containerd[1562]: time="2025-09-12T18:09:04.734780102Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Sep 12 18:09:04.736138 containerd[1562]: time="2025-09-12T18:09:04.736098188Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 12 18:09:04.788178 containerd[1562]: time="2025-09-12T18:09:04.787222411Z" level=info msg="CreateContainer within sandbox \"1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 12 18:09:04.801513 containerd[1562]: time="2025-09-12T18:09:04.798722425Z" level=info msg="Container 2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:09:04.830177 containerd[1562]: time="2025-09-12T18:09:04.830107097Z" level=info msg="CreateContainer within sandbox \"1eddbee59b20ae5ee75e4d18120630e44efae52528b19ccea0cf2a76a4c05b00\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\"" Sep 12 18:09:04.832024 containerd[1562]: time="2025-09-12T18:09:04.831976837Z" level=info msg="StartContainer for \"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\"" Sep 12 18:09:04.833273 containerd[1562]: time="2025-09-12T18:09:04.833229851Z" level=info msg="connecting to shim 2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1" address="unix:///run/containerd/s/4165bd43384c8f7e6d7fcbf0b859412051cfd72ebb71909d06a3065809cd1c47" protocol=ttrpc version=3 Sep 12 18:09:04.902756 systemd[1]: Started cri-containerd-2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1.scope - libcontainer container 2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1. Sep 12 18:09:05.061223 containerd[1562]: time="2025-09-12T18:09:05.061182362Z" level=info msg="StartContainer for \"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\" returns successfully" Sep 12 18:09:05.457152 kubelet[2714]: I0912 18:09:05.456883 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-8hj6c" podStartSLOduration=29.252140324 podStartE2EDuration="37.447015163s" podCreationTimestamp="2025-09-12 18:08:28 +0000 UTC" firstStartedPulling="2025-09-12 18:08:52.354879969 +0000 UTC m=+44.515740247" lastFinishedPulling="2025-09-12 18:09:00.549754792 +0000 UTC m=+52.710615086" observedRunningTime="2025-09-12 18:09:01.412774012 +0000 UTC m=+53.573634309" watchObservedRunningTime="2025-09-12 18:09:05.447015163 +0000 UTC m=+57.607875462" Sep 12 18:09:05.458883 kubelet[2714]: I0912 18:09:05.457588 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-855d7c9bb7-bjgr8" podStartSLOduration=27.936412996 podStartE2EDuration="37.457569261s" podCreationTimestamp="2025-09-12 18:08:28 +0000 UTC" firstStartedPulling="2025-09-12 18:08:55.214725615 +0000 UTC m=+47.375585899" lastFinishedPulling="2025-09-12 18:09:04.735881865 +0000 UTC m=+56.896742164" observedRunningTime="2025-09-12 18:09:05.446173973 +0000 UTC m=+57.607034271" watchObservedRunningTime="2025-09-12 18:09:05.457569261 +0000 UTC m=+57.618429555" Sep 12 18:09:05.647288 containerd[1562]: time="2025-09-12T18:09:05.647143647Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\" id:\"9c9014c77ad80ecc16ccebb81fa28b8249a714a5b4ccef6e56643a2b7588ec6d\" pid:5021 exited_at:{seconds:1757700545 nanos:515572804}" Sep 12 18:09:05.649646 containerd[1562]: time="2025-09-12T18:09:05.647353327Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\" id:\"04ab6fe6479d37ba7dd3cf5c71035e8828c01cfb404094326463c5a2c0850c51\" pid:5043 exited_at:{seconds:1757700545 nanos:575305537}" Sep 12 18:09:07.534179 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1680746382.mount: Deactivated successfully. Sep 12 18:09:08.367938 containerd[1562]: time="2025-09-12T18:09:08.367860435Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:08.379501 containerd[1562]: time="2025-09-12T18:09:08.369066220Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Sep 12 18:09:08.379501 containerd[1562]: time="2025-09-12T18:09:08.373683349Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:08.381624 containerd[1562]: time="2025-09-12T18:09:08.381580117Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 3.645444899s" Sep 12 18:09:08.381624 containerd[1562]: time="2025-09-12T18:09:08.381622710Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Sep 12 18:09:08.382147 containerd[1562]: time="2025-09-12T18:09:08.382126330Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 18:09:08.387766 containerd[1562]: time="2025-09-12T18:09:08.387731117Z" level=info msg="CreateContainer within sandbox \"a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 12 18:09:08.412319 containerd[1562]: time="2025-09-12T18:09:08.412264575Z" level=info msg="Container 09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e: CDI devices from CRI Config.CDIDevices: []" Sep 12 18:09:08.423958 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4042583613.mount: Deactivated successfully. Sep 12 18:09:08.428262 containerd[1562]: time="2025-09-12T18:09:08.428095513Z" level=info msg="CreateContainer within sandbox \"a2eb9b7f6aadbcf8223cd8b9ac198ea782ec2a55c95206f60179d868844e35eb\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\"" Sep 12 18:09:08.430369 containerd[1562]: time="2025-09-12T18:09:08.430325198Z" level=info msg="StartContainer for \"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\"" Sep 12 18:09:08.433582 containerd[1562]: time="2025-09-12T18:09:08.433524956Z" level=info msg="connecting to shim 09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e" address="unix:///run/containerd/s/db21a1dd861e010fcda65d5392fb2a5227cb1913ef98a65b105f869b88a16f5e" protocol=ttrpc version=3 Sep 12 18:09:08.468774 systemd[1]: Started cri-containerd-09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e.scope - libcontainer container 09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e. Sep 12 18:09:08.564702 containerd[1562]: time="2025-09-12T18:09:08.564576540Z" level=info msg="StartContainer for \"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" returns successfully" Sep 12 18:09:08.980159 systemd[1]: Started sshd@11-161.35.232.141:22-139.178.89.65:54206.service - OpenSSH per-connection server daemon (139.178.89.65:54206). Sep 12 18:09:09.151645 sshd[5110]: Accepted publickey for core from 139.178.89.65 port 54206 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:09.153812 sshd-session[5110]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:09.162964 systemd-logind[1528]: New session 8 of user core. Sep 12 18:09:09.168692 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 12 18:09:09.555505 kubelet[2714]: I0912 18:09:09.554244 2714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-7988f88666-dsgd8" podStartSLOduration=30.238099899 podStartE2EDuration="42.554222927s" podCreationTimestamp="2025-09-12 18:08:27 +0000 UTC" firstStartedPulling="2025-09-12 18:08:56.068940557 +0000 UTC m=+48.229800833" lastFinishedPulling="2025-09-12 18:09:08.38506357 +0000 UTC m=+60.545923861" observedRunningTime="2025-09-12 18:09:09.553104916 +0000 UTC m=+61.713965214" watchObservedRunningTime="2025-09-12 18:09:09.554222927 +0000 UTC m=+61.715083237" Sep 12 18:09:09.763910 containerd[1562]: time="2025-09-12T18:09:09.763302392Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\" id:\"2ad752f84e8ba6a2db410747e845dd58023a9db6a96348f41913f0e35c38b635\" pid:5158 exited_at:{seconds:1757700549 nanos:762823929}" Sep 12 18:09:09.828638 containerd[1562]: time="2025-09-12T18:09:09.828515784Z" level=info msg="TaskExit event in podsandbox handler container_id:\"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" id:\"6c337616efadb407b02656b238f71444fa10195dda7eedd2da7299c3eece0459\" pid:5148 exit_status:1 exited_at:{seconds:1757700549 nanos:827987071}" Sep 12 18:09:09.922265 sshd[5119]: Connection closed by 139.178.89.65 port 54206 Sep 12 18:09:09.923253 sshd-session[5110]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:09.933410 systemd[1]: sshd@11-161.35.232.141:22-139.178.89.65:54206.service: Deactivated successfully. Sep 12 18:09:09.936975 systemd[1]: session-8.scope: Deactivated successfully. Sep 12 18:09:09.939110 systemd-logind[1528]: Session 8 logged out. Waiting for processes to exit. Sep 12 18:09:09.941288 systemd-logind[1528]: Removed session 8. Sep 12 18:09:09.983871 containerd[1562]: time="2025-09-12T18:09:09.983782929Z" level=info msg="TaskExit event in podsandbox handler container_id:\"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" id:\"818ba5680e042705f8e13365c7462b05d55bd37f525f23f884fa18db45bf85bc\" pid:5185 exit_status:1 exited_at:{seconds:1757700549 nanos:983230128}" Sep 12 18:09:10.608393 containerd[1562]: time="2025-09-12T18:09:10.608329445Z" level=info msg="TaskExit event in podsandbox handler container_id:\"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" id:\"fba5b281c5bff4d8db28dd3007f58f63b678740417ec9016e22ea66f3e5cf582\" pid:5211 exit_status:1 exited_at:{seconds:1757700550 nanos:607775648}" Sep 12 18:09:14.938536 systemd[1]: Started sshd@12-161.35.232.141:22-139.178.89.65:37190.service - OpenSSH per-connection server daemon (139.178.89.65:37190). Sep 12 18:09:15.027749 sshd[5224]: Accepted publickey for core from 139.178.89.65 port 37190 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:15.030260 sshd-session[5224]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:15.038504 systemd-logind[1528]: New session 9 of user core. Sep 12 18:09:15.046828 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 12 18:09:15.241972 sshd[5227]: Connection closed by 139.178.89.65 port 37190 Sep 12 18:09:15.242925 sshd-session[5224]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:15.250412 systemd[1]: sshd@12-161.35.232.141:22-139.178.89.65:37190.service: Deactivated successfully. Sep 12 18:09:15.254384 systemd[1]: session-9.scope: Deactivated successfully. Sep 12 18:09:15.256552 systemd-logind[1528]: Session 9 logged out. Waiting for processes to exit. Sep 12 18:09:15.258395 systemd-logind[1528]: Removed session 9. Sep 12 18:09:15.335949 containerd[1562]: time="2025-09-12T18:09:15.335850839Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\" id:\"66721db1310075e7b87f243274f1270788cda2725007d14a1d7fa43bc80b7956\" pid:5251 exited_at:{seconds:1757700555 nanos:335378977}" Sep 12 18:09:20.260092 systemd[1]: Started sshd@13-161.35.232.141:22-139.178.89.65:45532.service - OpenSSH per-connection server daemon (139.178.89.65:45532). Sep 12 18:09:20.332854 sshd[5265]: Accepted publickey for core from 139.178.89.65 port 45532 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:20.334582 sshd-session[5265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:20.343798 systemd-logind[1528]: New session 10 of user core. Sep 12 18:09:20.350755 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 12 18:09:20.508816 sshd[5268]: Connection closed by 139.178.89.65 port 45532 Sep 12 18:09:20.509625 sshd-session[5265]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:20.521043 systemd[1]: sshd@13-161.35.232.141:22-139.178.89.65:45532.service: Deactivated successfully. Sep 12 18:09:20.523641 systemd[1]: session-10.scope: Deactivated successfully. Sep 12 18:09:20.525005 systemd-logind[1528]: Session 10 logged out. Waiting for processes to exit. Sep 12 18:09:20.529262 systemd[1]: Started sshd@14-161.35.232.141:22-139.178.89.65:45536.service - OpenSSH per-connection server daemon (139.178.89.65:45536). Sep 12 18:09:20.531427 systemd-logind[1528]: Removed session 10. Sep 12 18:09:20.618058 sshd[5281]: Accepted publickey for core from 139.178.89.65 port 45536 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:20.619595 sshd-session[5281]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:20.626603 systemd-logind[1528]: New session 11 of user core. Sep 12 18:09:20.632832 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 12 18:09:20.707512 kubelet[2714]: I0912 18:09:20.706019 2714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 18:09:20.935763 sshd[5284]: Connection closed by 139.178.89.65 port 45536 Sep 12 18:09:20.937170 sshd-session[5281]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:20.952807 systemd[1]: sshd@14-161.35.232.141:22-139.178.89.65:45536.service: Deactivated successfully. Sep 12 18:09:20.957072 systemd[1]: session-11.scope: Deactivated successfully. Sep 12 18:09:20.959758 systemd-logind[1528]: Session 11 logged out. Waiting for processes to exit. Sep 12 18:09:20.966693 systemd[1]: Started sshd@15-161.35.232.141:22-139.178.89.65:45546.service - OpenSSH per-connection server daemon (139.178.89.65:45546). Sep 12 18:09:20.972545 systemd-logind[1528]: Removed session 11. Sep 12 18:09:21.084256 sshd[5296]: Accepted publickey for core from 139.178.89.65 port 45546 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:21.086938 sshd-session[5296]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:21.096554 systemd-logind[1528]: New session 12 of user core. Sep 12 18:09:21.101838 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 12 18:09:21.261399 sshd[5299]: Connection closed by 139.178.89.65 port 45546 Sep 12 18:09:21.261262 sshd-session[5296]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:21.268565 systemd[1]: sshd@15-161.35.232.141:22-139.178.89.65:45546.service: Deactivated successfully. Sep 12 18:09:21.274367 systemd[1]: session-12.scope: Deactivated successfully. Sep 12 18:09:21.276726 systemd-logind[1528]: Session 12 logged out. Waiting for processes to exit. Sep 12 18:09:21.280849 systemd-logind[1528]: Removed session 12. Sep 12 18:09:21.996290 kubelet[2714]: E0912 18:09:21.993760 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:09:23.994465 kubelet[2714]: E0912 18:09:23.993571 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:09:24.604271 kubelet[2714]: I0912 18:09:24.604193 2714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 18:09:26.277739 systemd[1]: Started sshd@16-161.35.232.141:22-139.178.89.65:45554.service - OpenSSH per-connection server daemon (139.178.89.65:45554). Sep 12 18:09:26.367644 sshd[5316]: Accepted publickey for core from 139.178.89.65 port 45554 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:26.369775 sshd-session[5316]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:26.375636 systemd-logind[1528]: New session 13 of user core. Sep 12 18:09:26.381722 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 12 18:09:26.551620 sshd[5319]: Connection closed by 139.178.89.65 port 45554 Sep 12 18:09:26.552524 sshd-session[5316]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:26.562273 systemd[1]: sshd@16-161.35.232.141:22-139.178.89.65:45554.service: Deactivated successfully. Sep 12 18:09:26.565777 systemd[1]: session-13.scope: Deactivated successfully. Sep 12 18:09:26.570110 systemd-logind[1528]: Session 13 logged out. Waiting for processes to exit. Sep 12 18:09:26.571786 systemd-logind[1528]: Removed session 13. Sep 12 18:09:26.994864 kubelet[2714]: E0912 18:09:26.994754 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:09:28.153856 containerd[1562]: time="2025-09-12T18:09:28.153800459Z" level=info msg="TaskExit event in podsandbox handler container_id:\"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" id:\"1ea73ef63f5201e80add840331e5f4cf36561658fc724649b554553289880854\" pid:5346 exited_at:{seconds:1757700568 nanos:152237257}" Sep 12 18:09:31.566286 systemd[1]: Started sshd@17-161.35.232.141:22-139.178.89.65:59272.service - OpenSSH per-connection server daemon (139.178.89.65:59272). Sep 12 18:09:31.700941 sshd[5363]: Accepted publickey for core from 139.178.89.65 port 59272 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:31.703097 sshd-session[5363]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:31.709360 systemd-logind[1528]: New session 14 of user core. Sep 12 18:09:31.713709 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 12 18:09:31.897671 sshd[5366]: Connection closed by 139.178.89.65 port 59272 Sep 12 18:09:31.898446 sshd-session[5363]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:31.904614 systemd-logind[1528]: Session 14 logged out. Waiting for processes to exit. Sep 12 18:09:31.904853 systemd[1]: sshd@17-161.35.232.141:22-139.178.89.65:59272.service: Deactivated successfully. Sep 12 18:09:31.907121 systemd[1]: session-14.scope: Deactivated successfully. Sep 12 18:09:31.909296 systemd-logind[1528]: Removed session 14. Sep 12 18:09:35.400243 containerd[1562]: time="2025-09-12T18:09:35.400169463Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\" id:\"8a0bfe61bfcc08cb66ffc1eb9f9f66547562b2d02802e0ce66536757ed2e3928\" pid:5388 exited_at:{seconds:1757700575 nanos:399791695}" Sep 12 18:09:36.916299 systemd[1]: Started sshd@18-161.35.232.141:22-139.178.89.65:59278.service - OpenSSH per-connection server daemon (139.178.89.65:59278). Sep 12 18:09:37.030934 sshd[5402]: Accepted publickey for core from 139.178.89.65 port 59278 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:37.033109 sshd-session[5402]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:37.040029 systemd-logind[1528]: New session 15 of user core. Sep 12 18:09:37.048738 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 12 18:09:37.226219 sshd[5405]: Connection closed by 139.178.89.65 port 59278 Sep 12 18:09:37.226630 sshd-session[5402]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:37.233132 systemd-logind[1528]: Session 15 logged out. Waiting for processes to exit. Sep 12 18:09:37.233677 systemd[1]: sshd@18-161.35.232.141:22-139.178.89.65:59278.service: Deactivated successfully. Sep 12 18:09:37.236353 systemd[1]: session-15.scope: Deactivated successfully. Sep 12 18:09:37.239309 systemd-logind[1528]: Removed session 15. Sep 12 18:09:39.557329 containerd[1562]: time="2025-09-12T18:09:39.557188331Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\" id:\"08c4b86038d60057f28de26c61574e15386c030afc57b7481501e4768a873c1a\" pid:5427 exited_at:{seconds:1757700579 nanos:556791915}" Sep 12 18:09:39.669951 containerd[1562]: time="2025-09-12T18:09:39.669896599Z" level=info msg="TaskExit event in podsandbox handler container_id:\"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" id:\"88fda026da70d23cee14c5fb536857f926177223778382a097de2cb37734bf34\" pid:5449 exited_at:{seconds:1757700579 nanos:669320226}" Sep 12 18:09:40.010359 kubelet[2714]: E0912 18:09:40.010204 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:09:42.254062 systemd[1]: Started sshd@19-161.35.232.141:22-139.178.89.65:56526.service - OpenSSH per-connection server daemon (139.178.89.65:56526). Sep 12 18:09:42.408502 sshd[5462]: Accepted publickey for core from 139.178.89.65 port 56526 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:42.414438 sshd-session[5462]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:42.425144 systemd-logind[1528]: New session 16 of user core. Sep 12 18:09:42.431769 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 12 18:09:42.876167 sshd[5465]: Connection closed by 139.178.89.65 port 56526 Sep 12 18:09:42.880500 sshd-session[5462]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:42.891326 systemd[1]: Started sshd@20-161.35.232.141:22-139.178.89.65:56528.service - OpenSSH per-connection server daemon (139.178.89.65:56528). Sep 12 18:09:42.897865 systemd[1]: sshd@19-161.35.232.141:22-139.178.89.65:56526.service: Deactivated successfully. Sep 12 18:09:42.905310 systemd[1]: session-16.scope: Deactivated successfully. Sep 12 18:09:42.916807 systemd-logind[1528]: Session 16 logged out. Waiting for processes to exit. Sep 12 18:09:42.920202 systemd-logind[1528]: Removed session 16. Sep 12 18:09:43.012007 sshd[5473]: Accepted publickey for core from 139.178.89.65 port 56528 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:43.013996 sshd-session[5473]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:43.023655 systemd-logind[1528]: New session 17 of user core. Sep 12 18:09:43.032755 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 12 18:09:43.448549 sshd[5479]: Connection closed by 139.178.89.65 port 56528 Sep 12 18:09:43.463863 sshd-session[5473]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:43.481123 systemd[1]: Started sshd@21-161.35.232.141:22-139.178.89.65:56534.service - OpenSSH per-connection server daemon (139.178.89.65:56534). Sep 12 18:09:43.484370 systemd[1]: sshd@20-161.35.232.141:22-139.178.89.65:56528.service: Deactivated successfully. Sep 12 18:09:43.494495 systemd[1]: session-17.scope: Deactivated successfully. Sep 12 18:09:43.506563 systemd-logind[1528]: Session 17 logged out. Waiting for processes to exit. Sep 12 18:09:43.511070 systemd-logind[1528]: Removed session 17. Sep 12 18:09:43.610782 sshd[5486]: Accepted publickey for core from 139.178.89.65 port 56534 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:43.612685 sshd-session[5486]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:43.621400 systemd-logind[1528]: New session 18 of user core. Sep 12 18:09:43.626817 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 12 18:09:46.699086 sshd[5492]: Connection closed by 139.178.89.65 port 56534 Sep 12 18:09:46.729154 sshd-session[5486]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:46.766114 systemd[1]: Started sshd@22-161.35.232.141:22-139.178.89.65:56544.service - OpenSSH per-connection server daemon (139.178.89.65:56544). Sep 12 18:09:46.770271 systemd[1]: sshd@21-161.35.232.141:22-139.178.89.65:56534.service: Deactivated successfully. Sep 12 18:09:46.785301 systemd[1]: session-18.scope: Deactivated successfully. Sep 12 18:09:46.791881 systemd[1]: session-18.scope: Consumed 652ms CPU time, 77.9M memory peak. Sep 12 18:09:46.794629 systemd-logind[1528]: Session 18 logged out. Waiting for processes to exit. Sep 12 18:09:46.799922 systemd-logind[1528]: Removed session 18. Sep 12 18:09:46.940832 sshd[5508]: Accepted publickey for core from 139.178.89.65 port 56544 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:46.943823 sshd-session[5508]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:46.954829 systemd-logind[1528]: New session 19 of user core. Sep 12 18:09:46.962761 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 12 18:09:48.021630 sshd[5515]: Connection closed by 139.178.89.65 port 56544 Sep 12 18:09:48.020871 sshd-session[5508]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:48.040408 systemd[1]: sshd@22-161.35.232.141:22-139.178.89.65:56544.service: Deactivated successfully. Sep 12 18:09:48.045380 systemd[1]: session-19.scope: Deactivated successfully. Sep 12 18:09:48.050389 systemd-logind[1528]: Session 19 logged out. Waiting for processes to exit. Sep 12 18:09:48.055841 systemd-logind[1528]: Removed session 19. Sep 12 18:09:48.059255 systemd[1]: Started sshd@23-161.35.232.141:22-139.178.89.65:56548.service - OpenSSH per-connection server daemon (139.178.89.65:56548). Sep 12 18:09:48.179229 sshd[5525]: Accepted publickey for core from 139.178.89.65 port 56548 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:48.183207 sshd-session[5525]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:48.206663 systemd-logind[1528]: New session 20 of user core. Sep 12 18:09:48.209422 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 12 18:09:48.469085 sshd[5528]: Connection closed by 139.178.89.65 port 56548 Sep 12 18:09:48.469530 sshd-session[5525]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:48.481631 systemd[1]: sshd@23-161.35.232.141:22-139.178.89.65:56548.service: Deactivated successfully. Sep 12 18:09:48.490858 systemd[1]: session-20.scope: Deactivated successfully. Sep 12 18:09:48.494385 systemd-logind[1528]: Session 20 logged out. Waiting for processes to exit. Sep 12 18:09:48.497524 systemd-logind[1528]: Removed session 20. Sep 12 18:09:53.491277 systemd[1]: Started sshd@24-161.35.232.141:22-139.178.89.65:40200.service - OpenSSH per-connection server daemon (139.178.89.65:40200). Sep 12 18:09:53.617079 sshd[5544]: Accepted publickey for core from 139.178.89.65 port 40200 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:53.619900 sshd-session[5544]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:53.628569 systemd-logind[1528]: New session 21 of user core. Sep 12 18:09:53.634773 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 12 18:09:53.864502 sshd[5547]: Connection closed by 139.178.89.65 port 40200 Sep 12 18:09:53.865946 sshd-session[5544]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:53.872215 systemd[1]: sshd@24-161.35.232.141:22-139.178.89.65:40200.service: Deactivated successfully. Sep 12 18:09:53.875252 systemd[1]: session-21.scope: Deactivated successfully. Sep 12 18:09:53.884124 systemd-logind[1528]: Session 21 logged out. Waiting for processes to exit. Sep 12 18:09:53.885432 systemd-logind[1528]: Removed session 21. Sep 12 18:09:54.088433 kubelet[2714]: E0912 18:09:54.088236 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:09:58.878922 systemd[1]: Started sshd@25-161.35.232.141:22-139.178.89.65:40216.service - OpenSSH per-connection server daemon (139.178.89.65:40216). Sep 12 18:09:59.018227 sshd[5559]: Accepted publickey for core from 139.178.89.65 port 40216 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:09:59.021058 sshd-session[5559]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:09:59.029560 systemd-logind[1528]: New session 22 of user core. Sep 12 18:09:59.035762 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 12 18:09:59.363267 sshd[5562]: Connection closed by 139.178.89.65 port 40216 Sep 12 18:09:59.363972 sshd-session[5559]: pam_unix(sshd:session): session closed for user core Sep 12 18:09:59.370369 systemd[1]: sshd@25-161.35.232.141:22-139.178.89.65:40216.service: Deactivated successfully. Sep 12 18:09:59.370428 systemd-logind[1528]: Session 22 logged out. Waiting for processes to exit. Sep 12 18:09:59.375701 systemd[1]: session-22.scope: Deactivated successfully. Sep 12 18:09:59.380957 systemd-logind[1528]: Removed session 22. Sep 12 18:10:02.995783 kubelet[2714]: E0912 18:10:02.994937 2714 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 12 18:10:04.382814 systemd[1]: Started sshd@26-161.35.232.141:22-139.178.89.65:41474.service - OpenSSH per-connection server daemon (139.178.89.65:41474). Sep 12 18:10:04.489000 sshd[5573]: Accepted publickey for core from 139.178.89.65 port 41474 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:10:04.491952 sshd-session[5573]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:10:04.500373 systemd-logind[1528]: New session 23 of user core. Sep 12 18:10:04.509647 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 12 18:10:04.820661 sshd[5576]: Connection closed by 139.178.89.65 port 41474 Sep 12 18:10:04.823186 sshd-session[5573]: pam_unix(sshd:session): session closed for user core Sep 12 18:10:04.836224 systemd[1]: sshd@26-161.35.232.141:22-139.178.89.65:41474.service: Deactivated successfully. Sep 12 18:10:04.836732 systemd-logind[1528]: Session 23 logged out. Waiting for processes to exit. Sep 12 18:10:04.844632 systemd[1]: session-23.scope: Deactivated successfully. Sep 12 18:10:04.851285 systemd-logind[1528]: Removed session 23. Sep 12 18:10:05.764602 containerd[1562]: time="2025-09-12T18:10:05.764520074Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1139ed185527bc6b48b8f066e3a736429d8859776e9e05a24325226106c8ae0e\" id:\"8fa969795e3be048297ebc28ffbf63e3cd6165737ab1ea5fed6a08abc40b2fd1\" pid:5600 exited_at:{seconds:1757700605 nanos:759760195}" Sep 12 18:10:09.647901 containerd[1562]: time="2025-09-12T18:10:09.647787749Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2599abc1d52d312fcf3d7e857f9bfbc4c84397baba82597691b7ed0da1ca25a1\" id:\"98109977bd05b4ee9afc0303fd80a8532501b0dd9a2819572a9f6863ccaabf5e\" pid:5634 exited_at:{seconds:1757700609 nanos:647140131}" Sep 12 18:10:09.849409 systemd[1]: Started sshd@27-161.35.232.141:22-139.178.89.65:41486.service - OpenSSH per-connection server daemon (139.178.89.65:41486). Sep 12 18:10:09.882506 containerd[1562]: time="2025-09-12T18:10:09.859653597Z" level=info msg="TaskExit event in podsandbox handler container_id:\"09fa6243a8f85622d5a113f7bcad698d38b9d60ccbe391d3a456d8424bcef83e\" id:\"949161f0f21984dd709cf69c54e1800b3f2cea6cf1233150dfa70eea9753a9a6\" pid:5652 exited_at:{seconds:1757700609 nanos:857050259}" Sep 12 18:10:10.089513 sshd[5666]: Accepted publickey for core from 139.178.89.65 port 41486 ssh2: RSA SHA256:rgM4CCKqcUK6ImSFkPmxEROhKavbkgyEegeKnVmOeSQ Sep 12 18:10:10.091983 sshd-session[5666]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 18:10:10.103581 systemd-logind[1528]: New session 24 of user core. Sep 12 18:10:10.110613 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 12 18:10:11.004911 sshd[5669]: Connection closed by 139.178.89.65 port 41486 Sep 12 18:10:11.004691 sshd-session[5666]: pam_unix(sshd:session): session closed for user core Sep 12 18:10:11.022052 systemd[1]: sshd@27-161.35.232.141:22-139.178.89.65:41486.service: Deactivated successfully. Sep 12 18:10:11.025838 systemd[1]: session-24.scope: Deactivated successfully. Sep 12 18:10:11.030730 systemd-logind[1528]: Session 24 logged out. Waiting for processes to exit. Sep 12 18:10:11.035854 systemd-logind[1528]: Removed session 24.