Sep 9 21:16:58.750197 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Sep 9 21:16:58.750218 kernel: Linux version 6.12.45-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Tue Sep 9 19:54:20 -00 2025 Sep 9 21:16:58.750228 kernel: KASLR enabled Sep 9 21:16:58.750234 kernel: efi: EFI v2.7 by EDK II Sep 9 21:16:58.750239 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb832018 ACPI 2.0=0xdbfd0018 RNG=0xdbfd0a18 MEMRESERVE=0xdb838218 Sep 9 21:16:58.750244 kernel: random: crng init done Sep 9 21:16:58.750251 kernel: secureboot: Secure boot disabled Sep 9 21:16:58.750257 kernel: ACPI: Early table checksum verification disabled Sep 9 21:16:58.750263 kernel: ACPI: RSDP 0x00000000DBFD0018 000024 (v02 BOCHS ) Sep 9 21:16:58.750269 kernel: ACPI: XSDT 0x00000000DBFD0F18 000064 (v01 BOCHS BXPC 00000001 01000013) Sep 9 21:16:58.750275 kernel: ACPI: FACP 0x00000000DBFD0B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750281 kernel: ACPI: DSDT 0x00000000DBF0E018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750287 kernel: ACPI: APIC 0x00000000DBFD0C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750293 kernel: ACPI: PPTT 0x00000000DBFD0098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750300 kernel: ACPI: GTDT 0x00000000DBFD0818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750307 kernel: ACPI: MCFG 0x00000000DBFD0A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750313 kernel: ACPI: SPCR 0x00000000DBFD0918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750320 kernel: ACPI: DBG2 0x00000000DBFD0998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750326 kernel: ACPI: IORT 0x00000000DBFD0198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 21:16:58.750332 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Sep 9 21:16:58.750338 kernel: ACPI: Use ACPI SPCR as default console: No Sep 9 21:16:58.750344 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Sep 9 21:16:58.750350 kernel: NODE_DATA(0) allocated [mem 0xdc965a00-0xdc96cfff] Sep 9 21:16:58.750356 kernel: Zone ranges: Sep 9 21:16:58.750362 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Sep 9 21:16:58.750369 kernel: DMA32 empty Sep 9 21:16:58.750375 kernel: Normal empty Sep 9 21:16:58.750381 kernel: Device empty Sep 9 21:16:58.750387 kernel: Movable zone start for each node Sep 9 21:16:58.750393 kernel: Early memory node ranges Sep 9 21:16:58.750399 kernel: node 0: [mem 0x0000000040000000-0x00000000db81ffff] Sep 9 21:16:58.750405 kernel: node 0: [mem 0x00000000db820000-0x00000000db82ffff] Sep 9 21:16:58.750411 kernel: node 0: [mem 0x00000000db830000-0x00000000dc09ffff] Sep 9 21:16:58.750417 kernel: node 0: [mem 0x00000000dc0a0000-0x00000000dc2dffff] Sep 9 21:16:58.750423 kernel: node 0: [mem 0x00000000dc2e0000-0x00000000dc36ffff] Sep 9 21:16:58.750429 kernel: node 0: [mem 0x00000000dc370000-0x00000000dc45ffff] Sep 9 21:16:58.750435 kernel: node 0: [mem 0x00000000dc460000-0x00000000dc52ffff] Sep 9 21:16:58.750443 kernel: node 0: [mem 0x00000000dc530000-0x00000000dc5cffff] Sep 9 21:16:58.750449 kernel: node 0: [mem 0x00000000dc5d0000-0x00000000dce1ffff] Sep 9 21:16:58.750455 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Sep 9 21:16:58.750464 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Sep 9 21:16:58.750470 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Sep 9 21:16:58.750477 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Sep 9 21:16:58.750484 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Sep 9 21:16:58.750491 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Sep 9 21:16:58.750497 kernel: cma: Reserved 16 MiB at 0x00000000d8000000 on node -1 Sep 9 21:16:58.750503 kernel: psci: probing for conduit method from ACPI. Sep 9 21:16:58.750510 kernel: psci: PSCIv1.1 detected in firmware. Sep 9 21:16:58.750516 kernel: psci: Using standard PSCI v0.2 function IDs Sep 9 21:16:58.750522 kernel: psci: Trusted OS migration not required Sep 9 21:16:58.750529 kernel: psci: SMC Calling Convention v1.1 Sep 9 21:16:58.750535 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Sep 9 21:16:58.750542 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Sep 9 21:16:58.750550 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Sep 9 21:16:58.750556 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Sep 9 21:16:58.750563 kernel: Detected PIPT I-cache on CPU0 Sep 9 21:16:58.750569 kernel: CPU features: detected: GIC system register CPU interface Sep 9 21:16:58.750576 kernel: CPU features: detected: Spectre-v4 Sep 9 21:16:58.750582 kernel: CPU features: detected: Spectre-BHB Sep 9 21:16:58.750600 kernel: CPU features: kernel page table isolation forced ON by KASLR Sep 9 21:16:58.750607 kernel: CPU features: detected: Kernel page table isolation (KPTI) Sep 9 21:16:58.750614 kernel: CPU features: detected: ARM erratum 1418040 Sep 9 21:16:58.750620 kernel: CPU features: detected: SSBS not fully self-synchronizing Sep 9 21:16:58.750626 kernel: alternatives: applying boot alternatives Sep 9 21:16:58.750634 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=f5bd02e888bbcae51800cf37660dcdbf356eb05540a834019d706c2521a92d30 Sep 9 21:16:58.750642 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 9 21:16:58.750649 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 9 21:16:58.750655 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 9 21:16:58.750662 kernel: Fallback order for Node 0: 0 Sep 9 21:16:58.750668 kernel: Built 1 zonelists, mobility grouping on. Total pages: 643072 Sep 9 21:16:58.750675 kernel: Policy zone: DMA Sep 9 21:16:58.750681 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 9 21:16:58.750688 kernel: software IO TLB: SWIOTLB bounce buffer size adjusted to 2MB Sep 9 21:16:58.750694 kernel: software IO TLB: area num 4. Sep 9 21:16:58.750700 kernel: software IO TLB: SWIOTLB bounce buffer size roundup to 4MB Sep 9 21:16:58.750707 kernel: software IO TLB: mapped [mem 0x00000000d7c00000-0x00000000d8000000] (4MB) Sep 9 21:16:58.750714 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 9 21:16:58.750721 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 9 21:16:58.750728 kernel: rcu: RCU event tracing is enabled. Sep 9 21:16:58.750735 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 9 21:16:58.750741 kernel: Trampoline variant of Tasks RCU enabled. Sep 9 21:16:58.750748 kernel: Tracing variant of Tasks RCU enabled. Sep 9 21:16:58.750754 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 9 21:16:58.750761 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 9 21:16:58.750767 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 9 21:16:58.750774 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 9 21:16:58.750780 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Sep 9 21:16:58.750813 kernel: GICv3: 256 SPIs implemented Sep 9 21:16:58.750820 kernel: GICv3: 0 Extended SPIs implemented Sep 9 21:16:58.750826 kernel: Root IRQ handler: gic_handle_irq Sep 9 21:16:58.750833 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Sep 9 21:16:58.750839 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Sep 9 21:16:58.750846 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Sep 9 21:16:58.750852 kernel: ITS [mem 0x08080000-0x0809ffff] Sep 9 21:16:58.750859 kernel: ITS@0x0000000008080000: allocated 8192 Devices @40110000 (indirect, esz 8, psz 64K, shr 1) Sep 9 21:16:58.750865 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @40120000 (flat, esz 8, psz 64K, shr 1) Sep 9 21:16:58.750872 kernel: GICv3: using LPI property table @0x0000000040130000 Sep 9 21:16:58.750878 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040140000 Sep 9 21:16:58.750885 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 9 21:16:58.750893 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 9 21:16:58.750900 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Sep 9 21:16:58.750907 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Sep 9 21:16:58.750913 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Sep 9 21:16:58.750920 kernel: arm-pv: using stolen time PV Sep 9 21:16:58.750927 kernel: Console: colour dummy device 80x25 Sep 9 21:16:58.750933 kernel: ACPI: Core revision 20240827 Sep 9 21:16:58.750940 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Sep 9 21:16:58.750947 kernel: pid_max: default: 32768 minimum: 301 Sep 9 21:16:58.750953 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 9 21:16:58.750961 kernel: landlock: Up and running. Sep 9 21:16:58.750968 kernel: SELinux: Initializing. Sep 9 21:16:58.750975 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 9 21:16:58.750981 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 9 21:16:58.750988 kernel: rcu: Hierarchical SRCU implementation. Sep 9 21:16:58.750995 kernel: rcu: Max phase no-delay instances is 400. Sep 9 21:16:58.751002 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 9 21:16:58.751008 kernel: Remapping and enabling EFI services. Sep 9 21:16:58.751015 kernel: smp: Bringing up secondary CPUs ... Sep 9 21:16:58.751027 kernel: Detected PIPT I-cache on CPU1 Sep 9 21:16:58.751034 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Sep 9 21:16:58.751041 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040150000 Sep 9 21:16:58.751050 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 9 21:16:58.751057 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Sep 9 21:16:58.751063 kernel: Detected PIPT I-cache on CPU2 Sep 9 21:16:58.751071 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Sep 9 21:16:58.751078 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040160000 Sep 9 21:16:58.751086 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 9 21:16:58.751093 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Sep 9 21:16:58.751100 kernel: Detected PIPT I-cache on CPU3 Sep 9 21:16:58.751107 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Sep 9 21:16:58.751114 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040170000 Sep 9 21:16:58.751121 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 9 21:16:58.751127 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Sep 9 21:16:58.751134 kernel: smp: Brought up 1 node, 4 CPUs Sep 9 21:16:58.751141 kernel: SMP: Total of 4 processors activated. Sep 9 21:16:58.751150 kernel: CPU: All CPU(s) started at EL1 Sep 9 21:16:58.751157 kernel: CPU features: detected: 32-bit EL0 Support Sep 9 21:16:58.751164 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Sep 9 21:16:58.751171 kernel: CPU features: detected: Common not Private translations Sep 9 21:16:58.751178 kernel: CPU features: detected: CRC32 instructions Sep 9 21:16:58.751185 kernel: CPU features: detected: Enhanced Virtualization Traps Sep 9 21:16:58.751192 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Sep 9 21:16:58.751199 kernel: CPU features: detected: LSE atomic instructions Sep 9 21:16:58.751206 kernel: CPU features: detected: Privileged Access Never Sep 9 21:16:58.751214 kernel: CPU features: detected: RAS Extension Support Sep 9 21:16:58.751221 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Sep 9 21:16:58.751228 kernel: alternatives: applying system-wide alternatives Sep 9 21:16:58.751235 kernel: CPU features: detected: Hardware dirty bit management on CPU0-3 Sep 9 21:16:58.751242 kernel: Memory: 2424480K/2572288K available (11136K kernel code, 2436K rwdata, 9060K rodata, 38976K init, 1038K bss, 125472K reserved, 16384K cma-reserved) Sep 9 21:16:58.751249 kernel: devtmpfs: initialized Sep 9 21:16:58.751256 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 9 21:16:58.751264 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 9 21:16:58.751271 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Sep 9 21:16:58.751279 kernel: 0 pages in range for non-PLT usage Sep 9 21:16:58.751286 kernel: 508560 pages in range for PLT usage Sep 9 21:16:58.751293 kernel: pinctrl core: initialized pinctrl subsystem Sep 9 21:16:58.751300 kernel: SMBIOS 3.0.0 present. Sep 9 21:16:58.751307 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Sep 9 21:16:58.751314 kernel: DMI: Memory slots populated: 1/1 Sep 9 21:16:58.751320 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 9 21:16:58.751328 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Sep 9 21:16:58.751335 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Sep 9 21:16:58.751343 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Sep 9 21:16:58.751350 kernel: audit: initializing netlink subsys (disabled) Sep 9 21:16:58.751357 kernel: audit: type=2000 audit(0.021:1): state=initialized audit_enabled=0 res=1 Sep 9 21:16:58.751364 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 9 21:16:58.751371 kernel: cpuidle: using governor menu Sep 9 21:16:58.751378 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Sep 9 21:16:58.751385 kernel: ASID allocator initialised with 32768 entries Sep 9 21:16:58.751392 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 9 21:16:58.751399 kernel: Serial: AMBA PL011 UART driver Sep 9 21:16:58.751408 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 9 21:16:58.751415 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Sep 9 21:16:58.751422 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Sep 9 21:16:58.751429 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Sep 9 21:16:58.751436 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 9 21:16:58.751443 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Sep 9 21:16:58.751450 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Sep 9 21:16:58.751457 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Sep 9 21:16:58.751463 kernel: ACPI: Added _OSI(Module Device) Sep 9 21:16:58.751472 kernel: ACPI: Added _OSI(Processor Device) Sep 9 21:16:58.751479 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 9 21:16:58.751486 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 9 21:16:58.751493 kernel: ACPI: Interpreter enabled Sep 9 21:16:58.751500 kernel: ACPI: Using GIC for interrupt routing Sep 9 21:16:58.751507 kernel: ACPI: MCFG table detected, 1 entries Sep 9 21:16:58.751513 kernel: ACPI: CPU0 has been hot-added Sep 9 21:16:58.751520 kernel: ACPI: CPU1 has been hot-added Sep 9 21:16:58.751527 kernel: ACPI: CPU2 has been hot-added Sep 9 21:16:58.751534 kernel: ACPI: CPU3 has been hot-added Sep 9 21:16:58.751542 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Sep 9 21:16:58.751549 kernel: printk: legacy console [ttyAMA0] enabled Sep 9 21:16:58.751556 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 9 21:16:58.751699 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 9 21:16:58.751766 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 9 21:16:58.751842 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 9 21:16:58.751902 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Sep 9 21:16:58.751965 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Sep 9 21:16:58.751975 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Sep 9 21:16:58.751982 kernel: PCI host bridge to bus 0000:00 Sep 9 21:16:58.752056 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Sep 9 21:16:58.752113 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Sep 9 21:16:58.752167 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Sep 9 21:16:58.752221 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 9 21:16:58.752303 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Sep 9 21:16:58.752375 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 9 21:16:58.752438 kernel: pci 0000:00:01.0: BAR 0 [io 0x0000-0x001f] Sep 9 21:16:58.752499 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff] Sep 9 21:16:58.752559 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Sep 9 21:16:58.752634 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Sep 9 21:16:58.752695 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff]: assigned Sep 9 21:16:58.752757 kernel: pci 0000:00:01.0: BAR 0 [io 0x1000-0x101f]: assigned Sep 9 21:16:58.752822 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Sep 9 21:16:58.752875 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Sep 9 21:16:58.752928 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Sep 9 21:16:58.752937 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Sep 9 21:16:58.752944 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Sep 9 21:16:58.752951 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Sep 9 21:16:58.752961 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Sep 9 21:16:58.752968 kernel: iommu: Default domain type: Translated Sep 9 21:16:58.752975 kernel: iommu: DMA domain TLB invalidation policy: strict mode Sep 9 21:16:58.752982 kernel: efivars: Registered efivars operations Sep 9 21:16:58.752989 kernel: vgaarb: loaded Sep 9 21:16:58.752996 kernel: clocksource: Switched to clocksource arch_sys_counter Sep 9 21:16:58.753003 kernel: VFS: Disk quotas dquot_6.6.0 Sep 9 21:16:58.753010 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 9 21:16:58.753017 kernel: pnp: PnP ACPI init Sep 9 21:16:58.753087 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Sep 9 21:16:58.753097 kernel: pnp: PnP ACPI: found 1 devices Sep 9 21:16:58.753104 kernel: NET: Registered PF_INET protocol family Sep 9 21:16:58.753111 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 9 21:16:58.753119 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 9 21:16:58.753126 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 9 21:16:58.753133 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 9 21:16:58.753140 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 9 21:16:58.753149 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 9 21:16:58.753156 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 9 21:16:58.753163 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 9 21:16:58.753170 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 9 21:16:58.753177 kernel: PCI: CLS 0 bytes, default 64 Sep 9 21:16:58.753184 kernel: kvm [1]: HYP mode not available Sep 9 21:16:58.753191 kernel: Initialise system trusted keyrings Sep 9 21:16:58.753198 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 9 21:16:58.753205 kernel: Key type asymmetric registered Sep 9 21:16:58.753213 kernel: Asymmetric key parser 'x509' registered Sep 9 21:16:58.753220 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 9 21:16:58.753228 kernel: io scheduler mq-deadline registered Sep 9 21:16:58.753235 kernel: io scheduler kyber registered Sep 9 21:16:58.753242 kernel: io scheduler bfq registered Sep 9 21:16:58.753249 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Sep 9 21:16:58.753256 kernel: ACPI: button: Power Button [PWRB] Sep 9 21:16:58.753263 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Sep 9 21:16:58.753323 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Sep 9 21:16:58.753335 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 9 21:16:58.753342 kernel: thunder_xcv, ver 1.0 Sep 9 21:16:58.753349 kernel: thunder_bgx, ver 1.0 Sep 9 21:16:58.753356 kernel: nicpf, ver 1.0 Sep 9 21:16:58.753363 kernel: nicvf, ver 1.0 Sep 9 21:16:58.753430 kernel: rtc-efi rtc-efi.0: registered as rtc0 Sep 9 21:16:58.753487 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-09-09T21:16:58 UTC (1757452618) Sep 9 21:16:58.753497 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 9 21:16:58.753504 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Sep 9 21:16:58.753512 kernel: watchdog: NMI not fully supported Sep 9 21:16:58.753519 kernel: watchdog: Hard watchdog permanently disabled Sep 9 21:16:58.753526 kernel: NET: Registered PF_INET6 protocol family Sep 9 21:16:58.753533 kernel: Segment Routing with IPv6 Sep 9 21:16:58.753540 kernel: In-situ OAM (IOAM) with IPv6 Sep 9 21:16:58.753547 kernel: NET: Registered PF_PACKET protocol family Sep 9 21:16:58.753554 kernel: Key type dns_resolver registered Sep 9 21:16:58.753561 kernel: registered taskstats version 1 Sep 9 21:16:58.753568 kernel: Loading compiled-in X.509 certificates Sep 9 21:16:58.753576 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.45-flatcar: f5007e8dd2a6cc57a1fe19052a0aaf9985861c4d' Sep 9 21:16:58.753583 kernel: Demotion targets for Node 0: null Sep 9 21:16:58.753598 kernel: Key type .fscrypt registered Sep 9 21:16:58.753606 kernel: Key type fscrypt-provisioning registered Sep 9 21:16:58.753613 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 9 21:16:58.753620 kernel: ima: Allocated hash algorithm: sha1 Sep 9 21:16:58.753627 kernel: ima: No architecture policies found Sep 9 21:16:58.753634 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Sep 9 21:16:58.753642 kernel: clk: Disabling unused clocks Sep 9 21:16:58.753649 kernel: PM: genpd: Disabling unused power domains Sep 9 21:16:58.753656 kernel: Warning: unable to open an initial console. Sep 9 21:16:58.753663 kernel: Freeing unused kernel memory: 38976K Sep 9 21:16:58.753670 kernel: Run /init as init process Sep 9 21:16:58.753677 kernel: with arguments: Sep 9 21:16:58.753684 kernel: /init Sep 9 21:16:58.753691 kernel: with environment: Sep 9 21:16:58.753697 kernel: HOME=/ Sep 9 21:16:58.753704 kernel: TERM=linux Sep 9 21:16:58.753712 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 9 21:16:58.753720 systemd[1]: Successfully made /usr/ read-only. Sep 9 21:16:58.753730 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 9 21:16:58.753738 systemd[1]: Detected virtualization kvm. Sep 9 21:16:58.753746 systemd[1]: Detected architecture arm64. Sep 9 21:16:58.753753 systemd[1]: Running in initrd. Sep 9 21:16:58.753760 systemd[1]: No hostname configured, using default hostname. Sep 9 21:16:58.753770 systemd[1]: Hostname set to . Sep 9 21:16:58.753777 systemd[1]: Initializing machine ID from VM UUID. Sep 9 21:16:58.753792 systemd[1]: Queued start job for default target initrd.target. Sep 9 21:16:58.753800 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 9 21:16:58.753808 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 9 21:16:58.753816 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 9 21:16:58.753824 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 9 21:16:58.753831 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 9 21:16:58.753842 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 9 21:16:58.753851 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 9 21:16:58.753858 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 9 21:16:58.753866 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 9 21:16:58.753874 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 9 21:16:58.753881 systemd[1]: Reached target paths.target - Path Units. Sep 9 21:16:58.753889 systemd[1]: Reached target slices.target - Slice Units. Sep 9 21:16:58.753898 systemd[1]: Reached target swap.target - Swaps. Sep 9 21:16:58.753905 systemd[1]: Reached target timers.target - Timer Units. Sep 9 21:16:58.753913 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 9 21:16:58.753921 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 9 21:16:58.753928 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 9 21:16:58.753936 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 9 21:16:58.753943 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 9 21:16:58.753951 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 9 21:16:58.753960 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 9 21:16:58.753967 systemd[1]: Reached target sockets.target - Socket Units. Sep 9 21:16:58.753975 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 9 21:16:58.753983 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 9 21:16:58.753990 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 9 21:16:58.753998 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 9 21:16:58.754006 systemd[1]: Starting systemd-fsck-usr.service... Sep 9 21:16:58.754013 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 9 21:16:58.754021 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 9 21:16:58.754029 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 21:16:58.754037 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 9 21:16:58.754045 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 9 21:16:58.754053 systemd[1]: Finished systemd-fsck-usr.service. Sep 9 21:16:58.754062 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 9 21:16:58.754086 systemd-journald[244]: Collecting audit messages is disabled. Sep 9 21:16:58.754106 systemd-journald[244]: Journal started Sep 9 21:16:58.754125 systemd-journald[244]: Runtime Journal (/run/log/journal/d8cf0b6fb65d43c0830f0e54d33f24e4) is 6M, max 48.5M, 42.4M free. Sep 9 21:16:58.744947 systemd-modules-load[245]: Inserted module 'overlay' Sep 9 21:16:58.758824 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 9 21:16:58.758842 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 21:16:58.760170 systemd-modules-load[245]: Inserted module 'br_netfilter' Sep 9 21:16:58.761545 kernel: Bridge firewalling registered Sep 9 21:16:58.761561 systemd[1]: Started systemd-journald.service - Journal Service. Sep 9 21:16:58.762603 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 9 21:16:58.764009 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 9 21:16:58.768571 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 9 21:16:58.770137 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 9 21:16:58.772003 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 9 21:16:58.786940 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 9 21:16:58.791015 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 9 21:16:58.793166 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 9 21:16:58.795307 systemd-tmpfiles[271]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 9 21:16:58.798353 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 9 21:16:58.800688 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 9 21:16:58.804515 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 9 21:16:58.811408 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 9 21:16:58.826476 dracut-cmdline[290]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=f5bd02e888bbcae51800cf37660dcdbf356eb05540a834019d706c2521a92d30 Sep 9 21:16:58.839753 systemd-resolved[285]: Positive Trust Anchors: Sep 9 21:16:58.839772 systemd-resolved[285]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 9 21:16:58.839816 systemd-resolved[285]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 9 21:16:58.844541 systemd-resolved[285]: Defaulting to hostname 'linux'. Sep 9 21:16:58.845500 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 9 21:16:58.849597 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 9 21:16:58.904814 kernel: SCSI subsystem initialized Sep 9 21:16:58.908810 kernel: Loading iSCSI transport class v2.0-870. Sep 9 21:16:58.916816 kernel: iscsi: registered transport (tcp) Sep 9 21:16:58.928818 kernel: iscsi: registered transport (qla4xxx) Sep 9 21:16:58.928838 kernel: QLogic iSCSI HBA Driver Sep 9 21:16:58.944280 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 9 21:16:58.967865 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 9 21:16:58.969601 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 9 21:16:59.012424 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 9 21:16:59.014499 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 9 21:16:59.072816 kernel: raid6: neonx8 gen() 15760 MB/s Sep 9 21:16:59.089816 kernel: raid6: neonx4 gen() 15771 MB/s Sep 9 21:16:59.106816 kernel: raid6: neonx2 gen() 13227 MB/s Sep 9 21:16:59.123817 kernel: raid6: neonx1 gen() 10416 MB/s Sep 9 21:16:59.140813 kernel: raid6: int64x8 gen() 6874 MB/s Sep 9 21:16:59.157814 kernel: raid6: int64x4 gen() 7278 MB/s Sep 9 21:16:59.174816 kernel: raid6: int64x2 gen() 6082 MB/s Sep 9 21:16:59.191815 kernel: raid6: int64x1 gen() 5031 MB/s Sep 9 21:16:59.191841 kernel: raid6: using algorithm neonx4 gen() 15771 MB/s Sep 9 21:16:59.208813 kernel: raid6: .... xor() 12296 MB/s, rmw enabled Sep 9 21:16:59.208832 kernel: raid6: using neon recovery algorithm Sep 9 21:16:59.213814 kernel: xor: measuring software checksum speed Sep 9 21:16:59.213831 kernel: 8regs : 21658 MB/sec Sep 9 21:16:59.214820 kernel: 32regs : 21699 MB/sec Sep 9 21:16:59.214833 kernel: arm64_neon : 28361 MB/sec Sep 9 21:16:59.214845 kernel: xor: using function: arm64_neon (28361 MB/sec) Sep 9 21:16:59.266805 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 9 21:16:59.272739 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 9 21:16:59.274992 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 9 21:16:59.306039 systemd-udevd[498]: Using default interface naming scheme 'v255'. Sep 9 21:16:59.310073 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 9 21:16:59.311659 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 9 21:16:59.336876 dracut-pre-trigger[503]: rd.md=0: removing MD RAID activation Sep 9 21:16:59.358942 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 9 21:16:59.362902 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 9 21:16:59.410538 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 9 21:16:59.413438 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 9 21:16:59.455535 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Sep 9 21:16:59.455846 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 9 21:16:59.462970 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 9 21:16:59.462998 kernel: GPT:9289727 != 19775487 Sep 9 21:16:59.464886 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 9 21:16:59.464913 kernel: GPT:9289727 != 19775487 Sep 9 21:16:59.466334 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 9 21:16:59.465928 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 9 21:16:59.466032 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 21:16:59.470417 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 21:16:59.468135 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 21:16:59.470415 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 21:16:59.499640 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 9 21:16:59.500870 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 21:16:59.507441 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 9 21:16:59.515288 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 9 21:16:59.522487 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 9 21:16:59.528318 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 9 21:16:59.529312 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 9 21:16:59.531157 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 9 21:16:59.533576 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 9 21:16:59.535262 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 9 21:16:59.537646 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 9 21:16:59.539264 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 9 21:16:59.555604 disk-uuid[593]: Primary Header is updated. Sep 9 21:16:59.555604 disk-uuid[593]: Secondary Entries is updated. Sep 9 21:16:59.555604 disk-uuid[593]: Secondary Header is updated. Sep 9 21:16:59.559820 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 21:16:59.560061 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 9 21:17:00.568849 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 21:17:00.569145 disk-uuid[598]: The operation has completed successfully. Sep 9 21:17:00.594037 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 9 21:17:00.594137 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 9 21:17:00.617906 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 9 21:17:00.639620 sh[613]: Success Sep 9 21:17:00.651085 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 9 21:17:00.651139 kernel: device-mapper: uevent: version 1.0.3 Sep 9 21:17:00.652061 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 9 21:17:00.658819 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Sep 9 21:17:00.682648 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 9 21:17:00.685389 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 9 21:17:00.699155 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 9 21:17:00.705025 kernel: BTRFS: device fsid 0420e954-c3c6-4e24-9a07-863b2151b564 devid 1 transid 36 /dev/mapper/usr (253:0) scanned by mount (625) Sep 9 21:17:00.705059 kernel: BTRFS info (device dm-0): first mount of filesystem 0420e954-c3c6-4e24-9a07-863b2151b564 Sep 9 21:17:00.705070 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Sep 9 21:17:00.709802 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 9 21:17:00.709840 kernel: BTRFS info (device dm-0): enabling free space tree Sep 9 21:17:00.710418 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 9 21:17:00.711510 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 9 21:17:00.712603 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 9 21:17:00.713448 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 9 21:17:00.716121 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 9 21:17:00.732812 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (655) Sep 9 21:17:00.735019 kernel: BTRFS info (device vda6): first mount of filesystem 65698167-02fe-46cf-95a3-7944ec314f1c Sep 9 21:17:00.735054 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 9 21:17:00.737816 kernel: BTRFS info (device vda6): turning on async discard Sep 9 21:17:00.737857 kernel: BTRFS info (device vda6): enabling free space tree Sep 9 21:17:00.741810 kernel: BTRFS info (device vda6): last unmount of filesystem 65698167-02fe-46cf-95a3-7944ec314f1c Sep 9 21:17:00.742658 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 9 21:17:00.744464 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 9 21:17:00.821436 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 9 21:17:00.824710 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 9 21:17:00.842149 ignition[699]: Ignition 2.22.0 Sep 9 21:17:00.842167 ignition[699]: Stage: fetch-offline Sep 9 21:17:00.842199 ignition[699]: no configs at "/usr/lib/ignition/base.d" Sep 9 21:17:00.842207 ignition[699]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 21:17:00.842286 ignition[699]: parsed url from cmdline: "" Sep 9 21:17:00.842290 ignition[699]: no config URL provided Sep 9 21:17:00.842294 ignition[699]: reading system config file "/usr/lib/ignition/user.ign" Sep 9 21:17:00.842300 ignition[699]: no config at "/usr/lib/ignition/user.ign" Sep 9 21:17:00.842319 ignition[699]: op(1): [started] loading QEMU firmware config module Sep 9 21:17:00.842323 ignition[699]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 9 21:17:00.847206 ignition[699]: op(1): [finished] loading QEMU firmware config module Sep 9 21:17:00.859390 systemd-networkd[803]: lo: Link UP Sep 9 21:17:00.859403 systemd-networkd[803]: lo: Gained carrier Sep 9 21:17:00.860121 systemd-networkd[803]: Enumeration completed Sep 9 21:17:00.860197 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 9 21:17:00.860505 systemd-networkd[803]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 21:17:00.860508 systemd-networkd[803]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 9 21:17:00.861849 systemd[1]: Reached target network.target - Network. Sep 9 21:17:00.862208 systemd-networkd[803]: eth0: Link UP Sep 9 21:17:00.862571 systemd-networkd[803]: eth0: Gained carrier Sep 9 21:17:00.862596 systemd-networkd[803]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 21:17:00.884835 systemd-networkd[803]: eth0: DHCPv4 address 10.0.0.66/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 9 21:17:00.899741 ignition[699]: parsing config with SHA512: 854073898bd58556f17750e7d3c518c99bbf14804074a39f610267edb9f265085cf0bbe8f5b27c75f757fae425e44d1619d1250d69f5aaea3c37335796ed84c1 Sep 9 21:17:00.905681 unknown[699]: fetched base config from "system" Sep 9 21:17:00.905697 unknown[699]: fetched user config from "qemu" Sep 9 21:17:00.906092 ignition[699]: fetch-offline: fetch-offline passed Sep 9 21:17:00.906149 ignition[699]: Ignition finished successfully Sep 9 21:17:00.908037 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 9 21:17:00.909415 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 9 21:17:00.910168 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 9 21:17:00.937346 ignition[811]: Ignition 2.22.0 Sep 9 21:17:00.937367 ignition[811]: Stage: kargs Sep 9 21:17:00.937496 ignition[811]: no configs at "/usr/lib/ignition/base.d" Sep 9 21:17:00.937504 ignition[811]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 21:17:00.938224 ignition[811]: kargs: kargs passed Sep 9 21:17:00.938269 ignition[811]: Ignition finished successfully Sep 9 21:17:00.941142 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 9 21:17:00.942882 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 9 21:17:00.971898 ignition[819]: Ignition 2.22.0 Sep 9 21:17:00.971916 ignition[819]: Stage: disks Sep 9 21:17:00.972047 ignition[819]: no configs at "/usr/lib/ignition/base.d" Sep 9 21:17:00.972055 ignition[819]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 21:17:00.974727 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 9 21:17:00.972809 ignition[819]: disks: disks passed Sep 9 21:17:00.976334 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 9 21:17:00.972853 ignition[819]: Ignition finished successfully Sep 9 21:17:00.977577 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 9 21:17:00.978955 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 9 21:17:00.980359 systemd[1]: Reached target sysinit.target - System Initialization. Sep 9 21:17:00.981553 systemd[1]: Reached target basic.target - Basic System. Sep 9 21:17:00.983898 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 9 21:17:01.004781 systemd-fsck[830]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 9 21:17:01.008634 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 9 21:17:01.011069 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 9 21:17:01.071825 kernel: EXT4-fs (vda9): mounted filesystem 09d5f77d-9531-4ec2-9062-5fa777d03891 r/w with ordered data mode. Quota mode: none. Sep 9 21:17:01.071980 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 9 21:17:01.072976 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 9 21:17:01.075589 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 9 21:17:01.077544 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 9 21:17:01.078483 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 9 21:17:01.078535 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 9 21:17:01.078559 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 9 21:17:01.091295 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 9 21:17:01.093120 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 9 21:17:01.097211 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (838) Sep 9 21:17:01.097243 kernel: BTRFS info (device vda6): first mount of filesystem 65698167-02fe-46cf-95a3-7944ec314f1c Sep 9 21:17:01.097259 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 9 21:17:01.100150 kernel: BTRFS info (device vda6): turning on async discard Sep 9 21:17:01.100208 kernel: BTRFS info (device vda6): enabling free space tree Sep 9 21:17:01.101321 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 9 21:17:01.128490 initrd-setup-root[862]: cut: /sysroot/etc/passwd: No such file or directory Sep 9 21:17:01.132664 initrd-setup-root[869]: cut: /sysroot/etc/group: No such file or directory Sep 9 21:17:01.136412 initrd-setup-root[876]: cut: /sysroot/etc/shadow: No such file or directory Sep 9 21:17:01.140056 initrd-setup-root[883]: cut: /sysroot/etc/gshadow: No such file or directory Sep 9 21:17:01.204681 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 9 21:17:01.206727 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 9 21:17:01.208213 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 9 21:17:01.228862 kernel: BTRFS info (device vda6): last unmount of filesystem 65698167-02fe-46cf-95a3-7944ec314f1c Sep 9 21:17:01.239921 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 9 21:17:01.253935 ignition[951]: INFO : Ignition 2.22.0 Sep 9 21:17:01.253935 ignition[951]: INFO : Stage: mount Sep 9 21:17:01.255210 ignition[951]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 9 21:17:01.255210 ignition[951]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 21:17:01.255210 ignition[951]: INFO : mount: mount passed Sep 9 21:17:01.258501 ignition[951]: INFO : Ignition finished successfully Sep 9 21:17:01.258313 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 9 21:17:01.260185 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 9 21:17:01.837878 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 9 21:17:01.839429 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 9 21:17:01.855815 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (964) Sep 9 21:17:01.857803 kernel: BTRFS info (device vda6): first mount of filesystem 65698167-02fe-46cf-95a3-7944ec314f1c Sep 9 21:17:01.857844 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 9 21:17:01.860055 kernel: BTRFS info (device vda6): turning on async discard Sep 9 21:17:01.860089 kernel: BTRFS info (device vda6): enabling free space tree Sep 9 21:17:01.861419 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 9 21:17:01.901776 ignition[981]: INFO : Ignition 2.22.0 Sep 9 21:17:01.901776 ignition[981]: INFO : Stage: files Sep 9 21:17:01.903089 ignition[981]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 9 21:17:01.903089 ignition[981]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 21:17:01.903089 ignition[981]: DEBUG : files: compiled without relabeling support, skipping Sep 9 21:17:01.905662 ignition[981]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 9 21:17:01.905662 ignition[981]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 9 21:17:01.905662 ignition[981]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 9 21:17:01.905662 ignition[981]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 9 21:17:01.909765 ignition[981]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 9 21:17:01.909765 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Sep 9 21:17:01.909765 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Sep 9 21:17:01.905707 unknown[981]: wrote ssh authorized keys file for user: core Sep 9 21:17:02.233878 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 9 21:17:02.236053 systemd-networkd[803]: eth0: Gained IPv6LL Sep 9 21:17:02.712870 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Sep 9 21:17:02.712870 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 9 21:17:02.716272 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Sep 9 21:17:02.730219 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Sep 9 21:17:02.730219 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Sep 9 21:17:02.730219 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-arm64.raw: attempt #1 Sep 9 21:17:03.245455 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 9 21:17:03.580711 ignition[981]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-arm64.raw" Sep 9 21:17:03.580711 ignition[981]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 9 21:17:03.583880 ignition[981]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 9 21:17:03.597669 ignition[981]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 9 21:17:03.600935 ignition[981]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 9 21:17:03.603485 ignition[981]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 9 21:17:03.603485 ignition[981]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 9 21:17:03.603485 ignition[981]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 9 21:17:03.603485 ignition[981]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 9 21:17:03.603485 ignition[981]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 9 21:17:03.603485 ignition[981]: INFO : files: files passed Sep 9 21:17:03.603485 ignition[981]: INFO : Ignition finished successfully Sep 9 21:17:03.603996 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 9 21:17:03.606475 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 9 21:17:03.608155 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 9 21:17:03.621098 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 9 21:17:03.622152 initrd-setup-root-after-ignition[1010]: grep: /sysroot/oem/oem-release: No such file or directory Sep 9 21:17:03.623255 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 9 21:17:03.625866 initrd-setup-root-after-ignition[1012]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 9 21:17:03.625866 initrd-setup-root-after-ignition[1012]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 9 21:17:03.628704 initrd-setup-root-after-ignition[1016]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 9 21:17:03.627811 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 9 21:17:03.629853 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 9 21:17:03.633727 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 9 21:17:03.664686 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 9 21:17:03.664832 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 9 21:17:03.666675 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 9 21:17:03.668061 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 9 21:17:03.669551 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 9 21:17:03.670340 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 9 21:17:03.688852 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 9 21:17:03.690939 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 9 21:17:03.713390 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 9 21:17:03.714396 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 9 21:17:03.715947 systemd[1]: Stopped target timers.target - Timer Units. Sep 9 21:17:03.717327 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 9 21:17:03.717443 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 9 21:17:03.719429 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 9 21:17:03.721038 systemd[1]: Stopped target basic.target - Basic System. Sep 9 21:17:03.722485 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 9 21:17:03.723758 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 9 21:17:03.725404 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 9 21:17:03.726893 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 9 21:17:03.728345 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 9 21:17:03.729699 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 9 21:17:03.731381 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 9 21:17:03.732821 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 9 21:17:03.734298 systemd[1]: Stopped target swap.target - Swaps. Sep 9 21:17:03.735433 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 9 21:17:03.735546 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 9 21:17:03.737551 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 9 21:17:03.738840 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 9 21:17:03.740353 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 9 21:17:03.743860 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 9 21:17:03.744859 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 9 21:17:03.744981 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 9 21:17:03.747267 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 9 21:17:03.747388 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 9 21:17:03.748822 systemd[1]: Stopped target paths.target - Path Units. Sep 9 21:17:03.750076 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 9 21:17:03.754865 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 9 21:17:03.755867 systemd[1]: Stopped target slices.target - Slice Units. Sep 9 21:17:03.757510 systemd[1]: Stopped target sockets.target - Socket Units. Sep 9 21:17:03.758718 systemd[1]: iscsid.socket: Deactivated successfully. Sep 9 21:17:03.758816 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 9 21:17:03.760097 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 9 21:17:03.760173 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 9 21:17:03.761351 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 9 21:17:03.761464 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 9 21:17:03.762868 systemd[1]: ignition-files.service: Deactivated successfully. Sep 9 21:17:03.762968 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 9 21:17:03.764934 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 9 21:17:03.767989 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 9 21:17:03.769094 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 9 21:17:03.769225 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 9 21:17:03.770699 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 9 21:17:03.770814 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 9 21:17:03.775328 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 9 21:17:03.776910 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 9 21:17:03.784947 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 9 21:17:03.791924 ignition[1037]: INFO : Ignition 2.22.0 Sep 9 21:17:03.793897 ignition[1037]: INFO : Stage: umount Sep 9 21:17:03.793897 ignition[1037]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 9 21:17:03.793897 ignition[1037]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 21:17:03.793897 ignition[1037]: INFO : umount: umount passed Sep 9 21:17:03.797099 ignition[1037]: INFO : Ignition finished successfully Sep 9 21:17:03.798596 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 9 21:17:03.799531 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 9 21:17:03.801765 systemd[1]: Stopped target network.target - Network. Sep 9 21:17:03.803315 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 9 21:17:03.804228 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 9 21:17:03.805831 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 9 21:17:03.805894 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 9 21:17:03.807257 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 9 21:17:03.807302 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 9 21:17:03.808654 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 9 21:17:03.808690 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 9 21:17:03.810279 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 9 21:17:03.811623 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 9 21:17:03.819033 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 9 21:17:03.819884 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 9 21:17:03.822555 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 9 21:17:03.822766 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 9 21:17:03.822872 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 9 21:17:03.825757 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 9 21:17:03.826319 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 9 21:17:03.827855 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 9 21:17:03.827894 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 9 21:17:03.830255 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 9 21:17:03.831626 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 9 21:17:03.831684 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 9 21:17:03.833449 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 9 21:17:03.833495 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 9 21:17:03.836510 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 9 21:17:03.836562 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 9 21:17:03.838053 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 9 21:17:03.838097 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 9 21:17:03.840600 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 9 21:17:03.845898 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 9 21:17:03.845963 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 9 21:17:03.847937 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 9 21:17:03.848062 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 9 21:17:03.849981 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 9 21:17:03.850073 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 9 21:17:03.861514 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 9 21:17:03.861647 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 9 21:17:03.877393 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 9 21:17:03.877561 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 9 21:17:03.879309 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 9 21:17:03.879350 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 9 21:17:03.880614 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 9 21:17:03.880641 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 9 21:17:03.882034 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 9 21:17:03.882079 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 9 21:17:03.884159 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 9 21:17:03.884203 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 9 21:17:03.886317 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 9 21:17:03.886370 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 9 21:17:03.889293 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 9 21:17:03.890144 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 9 21:17:03.890216 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 9 21:17:03.893670 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 9 21:17:03.893739 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 9 21:17:03.896465 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 9 21:17:03.896507 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 9 21:17:03.899024 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 9 21:17:03.899066 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 9 21:17:03.900813 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 9 21:17:03.900860 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 21:17:03.904454 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Sep 9 21:17:03.904514 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Sep 9 21:17:03.904543 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 9 21:17:03.904580 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 9 21:17:03.911651 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 9 21:17:03.911776 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 9 21:17:03.913758 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 9 21:17:03.915834 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 9 21:17:03.936332 systemd[1]: Switching root. Sep 9 21:17:03.969945 systemd-journald[244]: Journal stopped Sep 9 21:17:04.672714 systemd-journald[244]: Received SIGTERM from PID 1 (systemd). Sep 9 21:17:04.672769 kernel: SELinux: policy capability network_peer_controls=1 Sep 9 21:17:04.672803 kernel: SELinux: policy capability open_perms=1 Sep 9 21:17:04.672814 kernel: SELinux: policy capability extended_socket_class=1 Sep 9 21:17:04.672823 kernel: SELinux: policy capability always_check_network=0 Sep 9 21:17:04.672833 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 9 21:17:04.672843 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 9 21:17:04.672852 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 9 21:17:04.672861 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 9 21:17:04.672872 kernel: SELinux: policy capability userspace_initial_context=0 Sep 9 21:17:04.672881 kernel: audit: type=1403 audit(1757452624.130:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 9 21:17:04.672894 systemd[1]: Successfully loaded SELinux policy in 51.765ms. Sep 9 21:17:04.672913 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 5.261ms. Sep 9 21:17:04.672924 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 9 21:17:04.672934 systemd[1]: Detected virtualization kvm. Sep 9 21:17:04.672944 systemd[1]: Detected architecture arm64. Sep 9 21:17:04.672954 systemd[1]: Detected first boot. Sep 9 21:17:04.672963 systemd[1]: Initializing machine ID from VM UUID. Sep 9 21:17:04.672975 zram_generator::config[1086]: No configuration found. Sep 9 21:17:04.672985 kernel: NET: Registered PF_VSOCK protocol family Sep 9 21:17:04.672995 systemd[1]: Populated /etc with preset unit settings. Sep 9 21:17:04.673005 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 9 21:17:04.673016 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 9 21:17:04.673025 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 9 21:17:04.673035 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 9 21:17:04.673045 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 9 21:17:04.673056 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 9 21:17:04.673066 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 9 21:17:04.673076 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 9 21:17:04.673087 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 9 21:17:04.673097 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 9 21:17:04.673110 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 9 21:17:04.673119 systemd[1]: Created slice user.slice - User and Session Slice. Sep 9 21:17:04.673129 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 9 21:17:04.673139 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 9 21:17:04.673150 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 9 21:17:04.673160 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 9 21:17:04.673170 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 9 21:17:04.673180 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 9 21:17:04.673190 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Sep 9 21:17:04.673200 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 9 21:17:04.673210 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 9 21:17:04.673221 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 9 21:17:04.673231 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 9 21:17:04.673245 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 9 21:17:04.673255 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 9 21:17:04.673265 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 9 21:17:04.673274 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 9 21:17:04.673284 systemd[1]: Reached target slices.target - Slice Units. Sep 9 21:17:04.673294 systemd[1]: Reached target swap.target - Swaps. Sep 9 21:17:04.673303 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 9 21:17:04.673313 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 9 21:17:04.673325 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 9 21:17:04.673335 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 9 21:17:04.673345 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 9 21:17:04.673355 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 9 21:17:04.673365 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 9 21:17:04.673374 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 9 21:17:04.673384 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 9 21:17:04.673394 systemd[1]: Mounting media.mount - External Media Directory... Sep 9 21:17:04.673404 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 9 21:17:04.673415 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 9 21:17:04.673425 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 9 21:17:04.673435 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 9 21:17:04.673446 systemd[1]: Reached target machines.target - Containers. Sep 9 21:17:04.673458 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 9 21:17:04.673468 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 9 21:17:04.673478 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 9 21:17:04.673488 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 9 21:17:04.673499 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 9 21:17:04.673509 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 9 21:17:04.673519 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 9 21:17:04.673529 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 9 21:17:04.673538 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 9 21:17:04.673549 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 9 21:17:04.673559 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 9 21:17:04.673576 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 9 21:17:04.673590 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 9 21:17:04.673602 systemd[1]: Stopped systemd-fsck-usr.service. Sep 9 21:17:04.673613 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 9 21:17:04.673623 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 9 21:17:04.673633 kernel: loop: module loaded Sep 9 21:17:04.673643 kernel: fuse: init (API version 7.41) Sep 9 21:17:04.673653 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 9 21:17:04.673667 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 9 21:17:04.673677 kernel: ACPI: bus type drm_connector registered Sep 9 21:17:04.673688 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 9 21:17:04.673698 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 9 21:17:04.673708 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 9 21:17:04.673718 systemd[1]: verity-setup.service: Deactivated successfully. Sep 9 21:17:04.673728 systemd[1]: Stopped verity-setup.service. Sep 9 21:17:04.673739 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 9 21:17:04.673749 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 9 21:17:04.673782 systemd-journald[1154]: Collecting audit messages is disabled. Sep 9 21:17:04.673848 systemd[1]: Mounted media.mount - External Media Directory. Sep 9 21:17:04.673862 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 9 21:17:04.673872 systemd-journald[1154]: Journal started Sep 9 21:17:04.673893 systemd-journald[1154]: Runtime Journal (/run/log/journal/d8cf0b6fb65d43c0830f0e54d33f24e4) is 6M, max 48.5M, 42.4M free. Sep 9 21:17:04.496555 systemd[1]: Queued start job for default target multi-user.target. Sep 9 21:17:04.506681 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 9 21:17:04.507059 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 9 21:17:04.675815 systemd[1]: Started systemd-journald.service - Journal Service. Sep 9 21:17:04.677433 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 9 21:17:04.678502 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 9 21:17:04.679596 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 9 21:17:04.681985 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 9 21:17:04.682153 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 9 21:17:04.685231 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 9 21:17:04.686418 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 9 21:17:04.686598 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 9 21:17:04.687868 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 9 21:17:04.688034 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 9 21:17:04.689087 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 9 21:17:04.689274 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 9 21:17:04.690450 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 9 21:17:04.690628 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 9 21:17:04.691805 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 9 21:17:04.693825 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 9 21:17:04.694934 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 9 21:17:04.696079 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 9 21:17:04.697384 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 9 21:17:04.698713 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 9 21:17:04.710873 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 9 21:17:04.712881 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 9 21:17:04.714696 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 9 21:17:04.715667 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 9 21:17:04.715694 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 9 21:17:04.717526 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 9 21:17:04.723580 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 9 21:17:04.724835 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 9 21:17:04.725945 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 9 21:17:04.727643 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 9 21:17:04.728727 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 9 21:17:04.731933 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 9 21:17:04.733309 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 9 21:17:04.734218 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 9 21:17:04.736527 systemd-journald[1154]: Time spent on flushing to /var/log/journal/d8cf0b6fb65d43c0830f0e54d33f24e4 is 11.785ms for 889 entries. Sep 9 21:17:04.736527 systemd-journald[1154]: System Journal (/var/log/journal/d8cf0b6fb65d43c0830f0e54d33f24e4) is 8M, max 195.6M, 187.6M free. Sep 9 21:17:04.753818 systemd-journald[1154]: Received client request to flush runtime journal. Sep 9 21:17:04.737644 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 9 21:17:04.741004 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 9 21:17:04.748941 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 9 21:17:04.750275 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 9 21:17:04.752535 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 9 21:17:04.757812 kernel: loop0: detected capacity change from 0 to 207008 Sep 9 21:17:04.759825 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 9 21:17:04.770805 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 9 21:17:04.771046 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 9 21:17:04.773129 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 9 21:17:04.775773 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 9 21:17:04.781000 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 9 21:17:04.785732 systemd-tmpfiles[1203]: ACLs are not supported, ignoring. Sep 9 21:17:04.785753 systemd-tmpfiles[1203]: ACLs are not supported, ignoring. Sep 9 21:17:04.790281 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 9 21:17:04.794922 kernel: loop1: detected capacity change from 0 to 100632 Sep 9 21:17:04.795323 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 9 21:17:04.808979 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 9 21:17:04.826827 kernel: loop2: detected capacity change from 0 to 119368 Sep 9 21:17:04.840632 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 9 21:17:04.843100 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 9 21:17:04.853820 kernel: loop3: detected capacity change from 0 to 207008 Sep 9 21:17:04.859834 kernel: loop4: detected capacity change from 0 to 100632 Sep 9 21:17:04.863425 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Sep 9 21:17:04.863445 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Sep 9 21:17:04.867893 kernel: loop5: detected capacity change from 0 to 119368 Sep 9 21:17:04.866101 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 9 21:17:04.872100 (sd-merge)[1226]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 9 21:17:04.872455 (sd-merge)[1226]: Merged extensions into '/usr'. Sep 9 21:17:04.876988 systemd[1]: Reload requested from client PID 1202 ('systemd-sysext') (unit systemd-sysext.service)... Sep 9 21:17:04.877001 systemd[1]: Reloading... Sep 9 21:17:04.930820 zram_generator::config[1254]: No configuration found. Sep 9 21:17:04.992616 ldconfig[1197]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 9 21:17:05.095215 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 9 21:17:05.095809 systemd[1]: Reloading finished in 218 ms. Sep 9 21:17:05.112651 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 9 21:17:05.113966 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 9 21:17:05.125019 systemd[1]: Starting ensure-sysext.service... Sep 9 21:17:05.126618 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 9 21:17:05.135839 systemd[1]: Reload requested from client PID 1290 ('systemctl') (unit ensure-sysext.service)... Sep 9 21:17:05.135854 systemd[1]: Reloading... Sep 9 21:17:05.140283 systemd-tmpfiles[1291]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 9 21:17:05.140621 systemd-tmpfiles[1291]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 9 21:17:05.140959 systemd-tmpfiles[1291]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 9 21:17:05.141245 systemd-tmpfiles[1291]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 9 21:17:05.142022 systemd-tmpfiles[1291]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 9 21:17:05.142329 systemd-tmpfiles[1291]: ACLs are not supported, ignoring. Sep 9 21:17:05.142439 systemd-tmpfiles[1291]: ACLs are not supported, ignoring. Sep 9 21:17:05.145148 systemd-tmpfiles[1291]: Detected autofs mount point /boot during canonicalization of boot. Sep 9 21:17:05.145211 systemd-tmpfiles[1291]: Skipping /boot Sep 9 21:17:05.151161 systemd-tmpfiles[1291]: Detected autofs mount point /boot during canonicalization of boot. Sep 9 21:17:05.151248 systemd-tmpfiles[1291]: Skipping /boot Sep 9 21:17:05.181823 zram_generator::config[1321]: No configuration found. Sep 9 21:17:05.322435 systemd[1]: Reloading finished in 186 ms. Sep 9 21:17:05.344314 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 9 21:17:05.350954 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 9 21:17:05.360711 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 9 21:17:05.362986 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 9 21:17:05.365204 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 9 21:17:05.367869 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 9 21:17:05.370048 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 9 21:17:05.375017 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 9 21:17:05.381940 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 9 21:17:05.387866 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 9 21:17:05.390064 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 9 21:17:05.392936 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 9 21:17:05.393783 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 9 21:17:05.393918 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 9 21:17:05.395912 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 9 21:17:05.400556 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 9 21:17:05.402530 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 9 21:17:05.406865 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 9 21:17:05.408530 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 9 21:17:05.408704 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 9 21:17:05.413341 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 9 21:17:05.413529 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 9 21:17:05.417592 systemd-udevd[1362]: Using default interface naming scheme 'v255'. Sep 9 21:17:05.419148 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 9 21:17:05.424014 augenrules[1384]: No rules Sep 9 21:17:05.425016 systemd[1]: audit-rules.service: Deactivated successfully. Sep 9 21:17:05.425272 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 9 21:17:05.431867 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 9 21:17:05.433476 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 9 21:17:05.438984 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 9 21:17:05.440971 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 9 21:17:05.443331 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 9 21:17:05.446058 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 9 21:17:05.449219 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 9 21:17:05.462299 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 9 21:17:05.463172 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 9 21:17:05.463216 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 9 21:17:05.465268 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 9 21:17:05.469031 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 9 21:17:05.470121 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 9 21:17:05.470544 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 9 21:17:05.472448 systemd[1]: Finished ensure-sysext.service. Sep 9 21:17:05.473638 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 9 21:17:05.473798 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 9 21:17:05.476161 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 9 21:17:05.476316 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 9 21:17:05.477418 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 9 21:17:05.477573 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 9 21:17:05.479007 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 9 21:17:05.479153 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 9 21:17:05.481824 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 9 21:17:05.487910 augenrules[1407]: /sbin/augenrules: No change Sep 9 21:17:05.492164 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 9 21:17:05.492222 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 9 21:17:05.493721 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 9 21:17:05.497377 augenrules[1458]: No rules Sep 9 21:17:05.498750 systemd[1]: audit-rules.service: Deactivated successfully. Sep 9 21:17:05.499095 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 9 21:17:05.517240 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Sep 9 21:17:05.562143 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 9 21:17:05.564415 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 9 21:17:05.590859 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 9 21:17:05.617865 systemd-networkd[1433]: lo: Link UP Sep 9 21:17:05.617873 systemd-networkd[1433]: lo: Gained carrier Sep 9 21:17:05.618650 systemd-networkd[1433]: Enumeration completed Sep 9 21:17:05.618757 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 9 21:17:05.619070 systemd-networkd[1433]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 21:17:05.619081 systemd-networkd[1433]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 9 21:17:05.619663 systemd-networkd[1433]: eth0: Link UP Sep 9 21:17:05.619814 systemd-networkd[1433]: eth0: Gained carrier Sep 9 21:17:05.619829 systemd-networkd[1433]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 21:17:05.621131 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 9 21:17:05.623317 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 9 21:17:05.629014 systemd-resolved[1358]: Positive Trust Anchors: Sep 9 21:17:05.629047 systemd-resolved[1358]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 9 21:17:05.629079 systemd-resolved[1358]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 9 21:17:05.636844 systemd-networkd[1433]: eth0: DHCPv4 address 10.0.0.66/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 9 21:17:05.637860 systemd-resolved[1358]: Defaulting to hostname 'linux'. Sep 9 21:17:05.640526 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 9 21:17:05.642100 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 9 21:17:05.643647 systemd[1]: Reached target network.target - Network. Sep 9 21:17:05.644753 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 9 21:17:05.646357 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 9 21:17:05.647481 systemd-timesyncd[1457]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 9 21:17:05.647544 systemd-timesyncd[1457]: Initial clock synchronization to Tue 2025-09-09 21:17:05.589858 UTC. Sep 9 21:17:05.647963 systemd[1]: Reached target sysinit.target - System Initialization. Sep 9 21:17:05.648879 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 9 21:17:05.649830 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 9 21:17:05.650710 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 9 21:17:05.652512 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 9 21:17:05.652548 systemd[1]: Reached target paths.target - Path Units. Sep 9 21:17:05.653272 systemd[1]: Reached target time-set.target - System Time Set. Sep 9 21:17:05.654152 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 9 21:17:05.655104 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 9 21:17:05.656026 systemd[1]: Reached target timers.target - Timer Units. Sep 9 21:17:05.657455 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 9 21:17:05.660440 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 9 21:17:05.663128 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 9 21:17:05.664233 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 9 21:17:05.665334 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 9 21:17:05.675492 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 9 21:17:05.678233 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 9 21:17:05.679815 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 9 21:17:05.680691 systemd[1]: Reached target sockets.target - Socket Units. Sep 9 21:17:05.681947 systemd[1]: Reached target basic.target - Basic System. Sep 9 21:17:05.683433 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 9 21:17:05.683463 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 9 21:17:05.685369 systemd[1]: Starting containerd.service - containerd container runtime... Sep 9 21:17:05.687889 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 9 21:17:05.706626 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 9 21:17:05.708491 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 9 21:17:05.710255 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 9 21:17:05.711066 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 9 21:17:05.722918 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 9 21:17:05.724242 jq[1499]: false Sep 9 21:17:05.725650 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 9 21:17:05.727842 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 9 21:17:05.729575 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 9 21:17:05.733185 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 9 21:17:05.735642 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 9 21:17:05.735831 extend-filesystems[1500]: Found /dev/vda6 Sep 9 21:17:05.736057 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 9 21:17:05.736575 systemd[1]: Starting update-engine.service - Update Engine... Sep 9 21:17:05.738661 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 9 21:17:05.741523 extend-filesystems[1500]: Found /dev/vda9 Sep 9 21:17:05.743840 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 9 21:17:05.744900 extend-filesystems[1500]: Checking size of /dev/vda9 Sep 9 21:17:05.745146 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 9 21:17:05.745302 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 9 21:17:05.745929 jq[1517]: true Sep 9 21:17:05.748080 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 9 21:17:05.748267 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 9 21:17:05.757623 jq[1524]: true Sep 9 21:17:05.767087 update_engine[1511]: I20250909 21:17:05.766856 1511 main.cc:92] Flatcar Update Engine starting Sep 9 21:17:05.768183 (ntainerd)[1525]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 9 21:17:05.772137 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 21:17:05.773915 systemd[1]: motdgen.service: Deactivated successfully. Sep 9 21:17:05.774129 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 9 21:17:05.776651 tar[1521]: linux-arm64/LICENSE Sep 9 21:17:05.776651 tar[1521]: linux-arm64/helm Sep 9 21:17:05.778931 extend-filesystems[1500]: Resized partition /dev/vda9 Sep 9 21:17:05.781044 extend-filesystems[1544]: resize2fs 1.47.3 (8-Jul-2025) Sep 9 21:17:05.785809 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 9 21:17:05.801429 dbus-daemon[1496]: [system] SELinux support is enabled Sep 9 21:17:05.801649 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 9 21:17:05.804320 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 9 21:17:05.804352 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 9 21:17:05.805386 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 9 21:17:05.805409 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 9 21:17:05.805608 update_engine[1511]: I20250909 21:17:05.805490 1511 update_check_scheduler.cc:74] Next update check in 8m27s Sep 9 21:17:05.809079 systemd[1]: Started update-engine.service - Update Engine. Sep 9 21:17:05.822255 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 9 21:17:05.841808 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 9 21:17:05.855713 extend-filesystems[1544]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 9 21:17:05.855713 extend-filesystems[1544]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 9 21:17:05.855713 extend-filesystems[1544]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 9 21:17:05.861297 extend-filesystems[1500]: Resized filesystem in /dev/vda9 Sep 9 21:17:05.862095 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 9 21:17:05.862293 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 9 21:17:05.864699 systemd-logind[1509]: Watching system buttons on /dev/input/event0 (Power Button) Sep 9 21:17:05.865851 bash[1560]: Updated "/home/core/.ssh/authorized_keys" Sep 9 21:17:05.866595 systemd-logind[1509]: New seat seat0. Sep 9 21:17:05.870262 systemd[1]: Started systemd-logind.service - User Login Management. Sep 9 21:17:05.871446 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 9 21:17:05.873831 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 21:17:05.877528 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 9 21:17:05.918334 locksmithd[1559]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 9 21:17:05.970696 containerd[1525]: time="2025-09-09T21:17:05Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 9 21:17:05.971540 containerd[1525]: time="2025-09-09T21:17:05.971505040Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 9 21:17:05.983124 containerd[1525]: time="2025-09-09T21:17:05.983089640Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="8.72µs" Sep 9 21:17:05.983155 containerd[1525]: time="2025-09-09T21:17:05.983123960Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 9 21:17:05.983155 containerd[1525]: time="2025-09-09T21:17:05.983141960Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 9 21:17:05.983413 containerd[1525]: time="2025-09-09T21:17:05.983387680Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 9 21:17:05.983438 containerd[1525]: time="2025-09-09T21:17:05.983417720Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 9 21:17:05.983459 containerd[1525]: time="2025-09-09T21:17:05.983446680Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 9 21:17:05.983581 containerd[1525]: time="2025-09-09T21:17:05.983548600Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 9 21:17:05.983609 containerd[1525]: time="2025-09-09T21:17:05.983580560Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 9 21:17:05.983931 containerd[1525]: time="2025-09-09T21:17:05.983901720Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 9 21:17:05.983954 containerd[1525]: time="2025-09-09T21:17:05.983931320Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 9 21:17:05.983954 containerd[1525]: time="2025-09-09T21:17:05.983945200Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 9 21:17:05.984108 containerd[1525]: time="2025-09-09T21:17:05.984087760Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 9 21:17:05.984201 containerd[1525]: time="2025-09-09T21:17:05.984182040Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 9 21:17:05.984509 containerd[1525]: time="2025-09-09T21:17:05.984483560Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 9 21:17:05.984534 containerd[1525]: time="2025-09-09T21:17:05.984524440Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 9 21:17:05.984556 containerd[1525]: time="2025-09-09T21:17:05.984536720Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 9 21:17:05.984648 containerd[1525]: time="2025-09-09T21:17:05.984628320Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 9 21:17:05.984916 containerd[1525]: time="2025-09-09T21:17:05.984893720Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 9 21:17:05.985053 containerd[1525]: time="2025-09-09T21:17:05.985030160Z" level=info msg="metadata content store policy set" policy=shared Sep 9 21:17:05.988472 containerd[1525]: time="2025-09-09T21:17:05.988435840Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 9 21:17:05.988513 containerd[1525]: time="2025-09-09T21:17:05.988500000Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 9 21:17:05.988533 containerd[1525]: time="2025-09-09T21:17:05.988520560Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 9 21:17:05.988550 containerd[1525]: time="2025-09-09T21:17:05.988533840Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 9 21:17:05.988550 containerd[1525]: time="2025-09-09T21:17:05.988544880Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 9 21:17:05.988644 containerd[1525]: time="2025-09-09T21:17:05.988555960Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 9 21:17:05.988711 containerd[1525]: time="2025-09-09T21:17:05.988649600Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 9 21:17:05.988743 containerd[1525]: time="2025-09-09T21:17:05.988714920Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 9 21:17:05.988743 containerd[1525]: time="2025-09-09T21:17:05.988730720Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 9 21:17:05.988743 containerd[1525]: time="2025-09-09T21:17:05.988741680Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 9 21:17:05.988804 containerd[1525]: time="2025-09-09T21:17:05.988750840Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 9 21:17:05.988804 containerd[1525]: time="2025-09-09T21:17:05.988762840Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 9 21:17:05.988965 containerd[1525]: time="2025-09-09T21:17:05.988941080Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 9 21:17:05.988996 containerd[1525]: time="2025-09-09T21:17:05.988972680Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 9 21:17:05.989018 containerd[1525]: time="2025-09-09T21:17:05.988995920Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 9 21:17:05.989018 containerd[1525]: time="2025-09-09T21:17:05.989007280Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 9 21:17:05.989089 containerd[1525]: time="2025-09-09T21:17:05.989068560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 9 21:17:05.989115 containerd[1525]: time="2025-09-09T21:17:05.989091600Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 9 21:17:05.989115 containerd[1525]: time="2025-09-09T21:17:05.989104840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 9 21:17:05.989155 containerd[1525]: time="2025-09-09T21:17:05.989115160Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 9 21:17:05.989155 containerd[1525]: time="2025-09-09T21:17:05.989126280Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 9 21:17:05.989155 containerd[1525]: time="2025-09-09T21:17:05.989136160Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 9 21:17:05.989155 containerd[1525]: time="2025-09-09T21:17:05.989146800Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 9 21:17:05.989399 containerd[1525]: time="2025-09-09T21:17:05.989377080Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 9 21:17:05.989428 containerd[1525]: time="2025-09-09T21:17:05.989402400Z" level=info msg="Start snapshots syncer" Sep 9 21:17:05.989507 containerd[1525]: time="2025-09-09T21:17:05.989486400Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 9 21:17:05.989975 containerd[1525]: time="2025-09-09T21:17:05.989933640Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 9 21:17:05.990074 containerd[1525]: time="2025-09-09T21:17:05.990050160Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 9 21:17:05.990161 containerd[1525]: time="2025-09-09T21:17:05.990136320Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 9 21:17:05.990489 containerd[1525]: time="2025-09-09T21:17:05.990410000Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 9 21:17:05.990515 containerd[1525]: time="2025-09-09T21:17:05.990497880Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 9 21:17:05.990515 containerd[1525]: time="2025-09-09T21:17:05.990510160Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 9 21:17:05.990548 containerd[1525]: time="2025-09-09T21:17:05.990523160Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 9 21:17:05.990576 containerd[1525]: time="2025-09-09T21:17:05.990534840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 9 21:17:05.990598 containerd[1525]: time="2025-09-09T21:17:05.990574320Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 9 21:17:05.990598 containerd[1525]: time="2025-09-09T21:17:05.990587200Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 9 21:17:05.990681 containerd[1525]: time="2025-09-09T21:17:05.990659840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 9 21:17:05.990705 containerd[1525]: time="2025-09-09T21:17:05.990684120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 9 21:17:05.990705 containerd[1525]: time="2025-09-09T21:17:05.990696960Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 9 21:17:05.990765 containerd[1525]: time="2025-09-09T21:17:05.990749720Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 9 21:17:05.990799 containerd[1525]: time="2025-09-09T21:17:05.990767640Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 9 21:17:05.990799 containerd[1525]: time="2025-09-09T21:17:05.990776960Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 9 21:17:05.990925 containerd[1525]: time="2025-09-09T21:17:05.990905160Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 9 21:17:05.990948 containerd[1525]: time="2025-09-09T21:17:05.990924320Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 9 21:17:05.990948 containerd[1525]: time="2025-09-09T21:17:05.990936800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 9 21:17:05.990980 containerd[1525]: time="2025-09-09T21:17:05.990946920Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 9 21:17:05.991081 containerd[1525]: time="2025-09-09T21:17:05.991064360Z" level=info msg="runtime interface created" Sep 9 21:17:05.991081 containerd[1525]: time="2025-09-09T21:17:05.991078920Z" level=info msg="created NRI interface" Sep 9 21:17:05.991119 containerd[1525]: time="2025-09-09T21:17:05.991089560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 9 21:17:05.991119 containerd[1525]: time="2025-09-09T21:17:05.991101400Z" level=info msg="Connect containerd service" Sep 9 21:17:05.991155 containerd[1525]: time="2025-09-09T21:17:05.991129000Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 9 21:17:05.994795 containerd[1525]: time="2025-09-09T21:17:05.993848680Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 9 21:17:06.059383 containerd[1525]: time="2025-09-09T21:17:06.059340234Z" level=info msg="Start subscribing containerd event" Sep 9 21:17:06.059462 containerd[1525]: time="2025-09-09T21:17:06.059402903Z" level=info msg="Start recovering state" Sep 9 21:17:06.059520 containerd[1525]: time="2025-09-09T21:17:06.059502943Z" level=info msg="Start event monitor" Sep 9 21:17:06.059564 containerd[1525]: time="2025-09-09T21:17:06.059523634Z" level=info msg="Start cni network conf syncer for default" Sep 9 21:17:06.059564 containerd[1525]: time="2025-09-09T21:17:06.059531934Z" level=info msg="Start streaming server" Sep 9 21:17:06.059564 containerd[1525]: time="2025-09-09T21:17:06.059539956Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 9 21:17:06.059564 containerd[1525]: time="2025-09-09T21:17:06.059546628Z" level=info msg="runtime interface starting up..." Sep 9 21:17:06.059564 containerd[1525]: time="2025-09-09T21:17:06.059552665Z" level=info msg="starting plugins..." Sep 9 21:17:06.059564 containerd[1525]: time="2025-09-09T21:17:06.059565890Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 9 21:17:06.059974 containerd[1525]: time="2025-09-09T21:17:06.059950481Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 9 21:17:06.060008 containerd[1525]: time="2025-09-09T21:17:06.060001077Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 9 21:17:06.060166 systemd[1]: Started containerd.service - containerd container runtime. Sep 9 21:17:06.061322 containerd[1525]: time="2025-09-09T21:17:06.061296671Z" level=info msg="containerd successfully booted in 0.090932s" Sep 9 21:17:06.122829 tar[1521]: linux-arm64/README.md Sep 9 21:17:06.146774 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 9 21:17:07.291912 systemd-networkd[1433]: eth0: Gained IPv6LL Sep 9 21:17:07.294370 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 9 21:17:07.297863 systemd[1]: Reached target network-online.target - Network is Online. Sep 9 21:17:07.299922 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 9 21:17:07.302109 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:07.303846 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 9 21:17:07.324575 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 9 21:17:07.324848 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 9 21:17:07.327811 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 9 21:17:07.329339 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 9 21:17:07.446762 sshd_keygen[1523]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 9 21:17:07.466015 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 9 21:17:07.468428 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 9 21:17:07.486082 systemd[1]: issuegen.service: Deactivated successfully. Sep 9 21:17:07.486288 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 9 21:17:07.488635 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 9 21:17:07.507417 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 9 21:17:07.509968 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 9 21:17:07.512053 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Sep 9 21:17:07.513266 systemd[1]: Reached target getty.target - Login Prompts. Sep 9 21:17:07.844114 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:07.845377 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 9 21:17:07.846470 systemd[1]: Startup finished in 1.974s (kernel) + 5.531s (initrd) + 3.767s (userspace) = 11.273s. Sep 9 21:17:07.849136 (kubelet)[1635]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 9 21:17:08.182333 kubelet[1635]: E0909 21:17:08.182206 1635 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 9 21:17:08.184634 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 9 21:17:08.184774 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 9 21:17:08.186896 systemd[1]: kubelet.service: Consumed 749ms CPU time, 254.9M memory peak. Sep 9 21:17:11.251240 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 9 21:17:11.252539 systemd[1]: Started sshd@0-10.0.0.66:22-10.0.0.1:47716.service - OpenSSH per-connection server daemon (10.0.0.1:47716). Sep 9 21:17:11.319023 sshd[1649]: Accepted publickey for core from 10.0.0.1 port 47716 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:11.320718 sshd-session[1649]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:11.326657 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 9 21:17:11.327581 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 9 21:17:11.332733 systemd-logind[1509]: New session 1 of user core. Sep 9 21:17:11.345818 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 9 21:17:11.348208 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 9 21:17:11.363844 (systemd)[1654]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 9 21:17:11.366057 systemd-logind[1509]: New session c1 of user core. Sep 9 21:17:11.466888 systemd[1654]: Queued start job for default target default.target. Sep 9 21:17:11.476716 systemd[1654]: Created slice app.slice - User Application Slice. Sep 9 21:17:11.476741 systemd[1654]: Reached target paths.target - Paths. Sep 9 21:17:11.476965 systemd[1654]: Reached target timers.target - Timers. Sep 9 21:17:11.478206 systemd[1654]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 9 21:17:11.486805 systemd[1654]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 9 21:17:11.486865 systemd[1654]: Reached target sockets.target - Sockets. Sep 9 21:17:11.486902 systemd[1654]: Reached target basic.target - Basic System. Sep 9 21:17:11.486929 systemd[1654]: Reached target default.target - Main User Target. Sep 9 21:17:11.486952 systemd[1654]: Startup finished in 115ms. Sep 9 21:17:11.487056 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 9 21:17:11.488316 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 9 21:17:11.554006 systemd[1]: Started sshd@1-10.0.0.66:22-10.0.0.1:47720.service - OpenSSH per-connection server daemon (10.0.0.1:47720). Sep 9 21:17:11.595734 sshd[1665]: Accepted publickey for core from 10.0.0.1 port 47720 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:11.596917 sshd-session[1665]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:11.601078 systemd-logind[1509]: New session 2 of user core. Sep 9 21:17:11.611924 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 9 21:17:11.662731 sshd[1668]: Connection closed by 10.0.0.1 port 47720 Sep 9 21:17:11.663045 sshd-session[1665]: pam_unix(sshd:session): session closed for user core Sep 9 21:17:11.677901 systemd[1]: sshd@1-10.0.0.66:22-10.0.0.1:47720.service: Deactivated successfully. Sep 9 21:17:11.680165 systemd[1]: session-2.scope: Deactivated successfully. Sep 9 21:17:11.680792 systemd-logind[1509]: Session 2 logged out. Waiting for processes to exit. Sep 9 21:17:11.682958 systemd[1]: Started sshd@2-10.0.0.66:22-10.0.0.1:47728.service - OpenSSH per-connection server daemon (10.0.0.1:47728). Sep 9 21:17:11.683515 systemd-logind[1509]: Removed session 2. Sep 9 21:17:11.731376 sshd[1674]: Accepted publickey for core from 10.0.0.1 port 47728 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:11.732517 sshd-session[1674]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:11.736206 systemd-logind[1509]: New session 3 of user core. Sep 9 21:17:11.751947 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 9 21:17:11.799614 sshd[1677]: Connection closed by 10.0.0.1 port 47728 Sep 9 21:17:11.799485 sshd-session[1674]: pam_unix(sshd:session): session closed for user core Sep 9 21:17:11.808909 systemd[1]: sshd@2-10.0.0.66:22-10.0.0.1:47728.service: Deactivated successfully. Sep 9 21:17:11.811058 systemd[1]: session-3.scope: Deactivated successfully. Sep 9 21:17:11.811693 systemd-logind[1509]: Session 3 logged out. Waiting for processes to exit. Sep 9 21:17:11.813731 systemd[1]: Started sshd@3-10.0.0.66:22-10.0.0.1:47744.service - OpenSSH per-connection server daemon (10.0.0.1:47744). Sep 9 21:17:11.814477 systemd-logind[1509]: Removed session 3. Sep 9 21:17:11.873391 sshd[1683]: Accepted publickey for core from 10.0.0.1 port 47744 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:11.874628 sshd-session[1683]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:11.879002 systemd-logind[1509]: New session 4 of user core. Sep 9 21:17:11.892936 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 9 21:17:11.944956 sshd[1686]: Connection closed by 10.0.0.1 port 47744 Sep 9 21:17:11.945423 sshd-session[1683]: pam_unix(sshd:session): session closed for user core Sep 9 21:17:11.954914 systemd[1]: sshd@3-10.0.0.66:22-10.0.0.1:47744.service: Deactivated successfully. Sep 9 21:17:11.956299 systemd[1]: session-4.scope: Deactivated successfully. Sep 9 21:17:11.958955 systemd-logind[1509]: Session 4 logged out. Waiting for processes to exit. Sep 9 21:17:11.960942 systemd[1]: Started sshd@4-10.0.0.66:22-10.0.0.1:47752.service - OpenSSH per-connection server daemon (10.0.0.1:47752). Sep 9 21:17:11.961569 systemd-logind[1509]: Removed session 4. Sep 9 21:17:12.007455 sshd[1692]: Accepted publickey for core from 10.0.0.1 port 47752 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:12.008488 sshd-session[1692]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:12.012845 systemd-logind[1509]: New session 5 of user core. Sep 9 21:17:12.027965 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 9 21:17:12.085071 sudo[1696]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 9 21:17:12.085331 sudo[1696]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 21:17:12.098686 sudo[1696]: pam_unix(sudo:session): session closed for user root Sep 9 21:17:12.100050 sshd[1695]: Connection closed by 10.0.0.1 port 47752 Sep 9 21:17:12.100432 sshd-session[1692]: pam_unix(sshd:session): session closed for user core Sep 9 21:17:12.111050 systemd[1]: sshd@4-10.0.0.66:22-10.0.0.1:47752.service: Deactivated successfully. Sep 9 21:17:12.113189 systemd[1]: session-5.scope: Deactivated successfully. Sep 9 21:17:12.113948 systemd-logind[1509]: Session 5 logged out. Waiting for processes to exit. Sep 9 21:17:12.116220 systemd[1]: Started sshd@5-10.0.0.66:22-10.0.0.1:47756.service - OpenSSH per-connection server daemon (10.0.0.1:47756). Sep 9 21:17:12.117143 systemd-logind[1509]: Removed session 5. Sep 9 21:17:12.182365 sshd[1702]: Accepted publickey for core from 10.0.0.1 port 47756 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:12.183487 sshd-session[1702]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:12.187008 systemd-logind[1509]: New session 6 of user core. Sep 9 21:17:12.202965 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 9 21:17:12.252480 sudo[1707]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 9 21:17:12.252748 sudo[1707]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 21:17:12.257179 sudo[1707]: pam_unix(sudo:session): session closed for user root Sep 9 21:17:12.261537 sudo[1706]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 9 21:17:12.261805 sudo[1706]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 21:17:12.270564 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 9 21:17:12.306646 augenrules[1729]: No rules Sep 9 21:17:12.307901 systemd[1]: audit-rules.service: Deactivated successfully. Sep 9 21:17:12.309845 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 9 21:17:12.311854 sudo[1706]: pam_unix(sudo:session): session closed for user root Sep 9 21:17:12.313357 sshd[1705]: Connection closed by 10.0.0.1 port 47756 Sep 9 21:17:12.313234 sshd-session[1702]: pam_unix(sshd:session): session closed for user core Sep 9 21:17:12.320829 systemd[1]: sshd@5-10.0.0.66:22-10.0.0.1:47756.service: Deactivated successfully. Sep 9 21:17:12.322185 systemd[1]: session-6.scope: Deactivated successfully. Sep 9 21:17:12.322916 systemd-logind[1509]: Session 6 logged out. Waiting for processes to exit. Sep 9 21:17:12.325059 systemd[1]: Started sshd@6-10.0.0.66:22-10.0.0.1:47772.service - OpenSSH per-connection server daemon (10.0.0.1:47772). Sep 9 21:17:12.325682 systemd-logind[1509]: Removed session 6. Sep 9 21:17:12.379137 sshd[1738]: Accepted publickey for core from 10.0.0.1 port 47772 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:17:12.380230 sshd-session[1738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:17:12.384080 systemd-logind[1509]: New session 7 of user core. Sep 9 21:17:12.394948 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 9 21:17:12.445758 sudo[1742]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 9 21:17:12.446055 sudo[1742]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 21:17:12.710176 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 9 21:17:12.727079 (dockerd)[1762]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 9 21:17:12.916586 dockerd[1762]: time="2025-09-09T21:17:12.916520104Z" level=info msg="Starting up" Sep 9 21:17:12.917334 dockerd[1762]: time="2025-09-09T21:17:12.917299078Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 9 21:17:12.927696 dockerd[1762]: time="2025-09-09T21:17:12.927637332Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 9 21:17:13.029703 dockerd[1762]: time="2025-09-09T21:17:13.029438965Z" level=info msg="Loading containers: start." Sep 9 21:17:13.036831 kernel: Initializing XFRM netlink socket Sep 9 21:17:13.225329 systemd-networkd[1433]: docker0: Link UP Sep 9 21:17:13.228158 dockerd[1762]: time="2025-09-09T21:17:13.228112340Z" level=info msg="Loading containers: done." Sep 9 21:17:13.240052 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1122679757-merged.mount: Deactivated successfully. Sep 9 21:17:13.241232 dockerd[1762]: time="2025-09-09T21:17:13.241178526Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 9 21:17:13.241303 dockerd[1762]: time="2025-09-09T21:17:13.241260775Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 9 21:17:13.241353 dockerd[1762]: time="2025-09-09T21:17:13.241337040Z" level=info msg="Initializing buildkit" Sep 9 21:17:13.262524 dockerd[1762]: time="2025-09-09T21:17:13.262490586Z" level=info msg="Completed buildkit initialization" Sep 9 21:17:13.267329 dockerd[1762]: time="2025-09-09T21:17:13.267297098Z" level=info msg="Daemon has completed initialization" Sep 9 21:17:13.267493 dockerd[1762]: time="2025-09-09T21:17:13.267357289Z" level=info msg="API listen on /run/docker.sock" Sep 9 21:17:13.267538 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 9 21:17:13.828141 containerd[1525]: time="2025-09-09T21:17:13.828103407Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.8\"" Sep 9 21:17:14.399003 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3136029842.mount: Deactivated successfully. Sep 9 21:17:15.429429 containerd[1525]: time="2025-09-09T21:17:15.429376750Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:15.430302 containerd[1525]: time="2025-09-09T21:17:15.430264399Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.8: active requests=0, bytes read=26328359" Sep 9 21:17:15.431831 containerd[1525]: time="2025-09-09T21:17:15.430949523Z" level=info msg="ImageCreate event name:\"sha256:61d628eec7e2101b908b4476f1e8e620490a9e8754184860c8eed25183acaa8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:15.433808 containerd[1525]: time="2025-09-09T21:17:15.433767415Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6e1a2f9b24f69ee77d0c0edaf32b31fdbb5e1a613f4476272197e6e1e239050b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:15.435194 containerd[1525]: time="2025-09-09T21:17:15.435155945Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.8\" with image id \"sha256:61d628eec7e2101b908b4476f1e8e620490a9e8754184860c8eed25183acaa8a\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6e1a2f9b24f69ee77d0c0edaf32b31fdbb5e1a613f4476272197e6e1e239050b\", size \"26325157\" in 1.607009493s" Sep 9 21:17:15.435235 containerd[1525]: time="2025-09-09T21:17:15.435197576Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.8\" returns image reference \"sha256:61d628eec7e2101b908b4476f1e8e620490a9e8754184860c8eed25183acaa8a\"" Sep 9 21:17:15.436069 containerd[1525]: time="2025-09-09T21:17:15.436041518Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.8\"" Sep 9 21:17:16.832129 containerd[1525]: time="2025-09-09T21:17:16.832081208Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:16.832667 containerd[1525]: time="2025-09-09T21:17:16.832631132Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.8: active requests=0, bytes read=22528554" Sep 9 21:17:16.833387 containerd[1525]: time="2025-09-09T21:17:16.833359321Z" level=info msg="ImageCreate event name:\"sha256:f17de36e40fc7cc372be0021b2c58ad61f05d3ebe4d430551bc5e4cd9ed2a061\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:16.836511 containerd[1525]: time="2025-09-09T21:17:16.836477926Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:8788ccd28ceed9e2e5f8fc31375ef5771df8ea6e518b362c9a06f3cc709cd6c7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:16.838129 containerd[1525]: time="2025-09-09T21:17:16.838021020Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.8\" with image id \"sha256:f17de36e40fc7cc372be0021b2c58ad61f05d3ebe4d430551bc5e4cd9ed2a061\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:8788ccd28ceed9e2e5f8fc31375ef5771df8ea6e518b362c9a06f3cc709cd6c7\", size \"24065666\" in 1.40194697s" Sep 9 21:17:16.838129 containerd[1525]: time="2025-09-09T21:17:16.838054756Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.8\" returns image reference \"sha256:f17de36e40fc7cc372be0021b2c58ad61f05d3ebe4d430551bc5e4cd9ed2a061\"" Sep 9 21:17:16.838561 containerd[1525]: time="2025-09-09T21:17:16.838531498Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.8\"" Sep 9 21:17:18.164181 containerd[1525]: time="2025-09-09T21:17:18.164121383Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:18.164973 containerd[1525]: time="2025-09-09T21:17:18.164908409Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.8: active requests=0, bytes read=17483529" Sep 9 21:17:18.165580 containerd[1525]: time="2025-09-09T21:17:18.165552201Z" level=info msg="ImageCreate event name:\"sha256:fe86d26bce3ccd5f0c4057c205b63fde1c8c752778025aea4605ffc3b0f80211\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:18.168297 containerd[1525]: time="2025-09-09T21:17:18.168263892Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:43c58bcbd1c7812dd19f8bfa5ae11093ebefd28699453ce86fc710869e155cd4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:18.170012 containerd[1525]: time="2025-09-09T21:17:18.169981935Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.8\" with image id \"sha256:fe86d26bce3ccd5f0c4057c205b63fde1c8c752778025aea4605ffc3b0f80211\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:43c58bcbd1c7812dd19f8bfa5ae11093ebefd28699453ce86fc710869e155cd4\", size \"19020659\" in 1.331412225s" Sep 9 21:17:18.170055 containerd[1525]: time="2025-09-09T21:17:18.170016485Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.8\" returns image reference \"sha256:fe86d26bce3ccd5f0c4057c205b63fde1c8c752778025aea4605ffc3b0f80211\"" Sep 9 21:17:18.170627 containerd[1525]: time="2025-09-09T21:17:18.170484211Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.8\"" Sep 9 21:17:18.435120 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 9 21:17:18.436534 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:18.570315 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:18.573819 (kubelet)[2054]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 9 21:17:18.617361 kubelet[2054]: E0909 21:17:18.617294 2054 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 9 21:17:18.620523 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 9 21:17:18.620780 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 9 21:17:18.621179 systemd[1]: kubelet.service: Consumed 148ms CPU time, 107.5M memory peak. Sep 9 21:17:19.236399 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1989761617.mount: Deactivated successfully. Sep 9 21:17:19.585650 containerd[1525]: time="2025-09-09T21:17:19.584847508Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:19.586014 containerd[1525]: time="2025-09-09T21:17:19.585968295Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.8: active requests=0, bytes read=27376726" Sep 9 21:17:19.586687 containerd[1525]: time="2025-09-09T21:17:19.586664137Z" level=info msg="ImageCreate event name:\"sha256:2cf30e39f99f8f4ee1a736a4f3175cc2d8d3f58936d8fa83ec5523658fdc7b8b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:19.588782 containerd[1525]: time="2025-09-09T21:17:19.588744154Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:adc1335b480ddd833aac3b0bd20f68ff0f3c3cf7a0bd337933b006d9f5cec40a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:19.589609 containerd[1525]: time="2025-09-09T21:17:19.589579141Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.8\" with image id \"sha256:2cf30e39f99f8f4ee1a736a4f3175cc2d8d3f58936d8fa83ec5523658fdc7b8b\", repo tag \"registry.k8s.io/kube-proxy:v1.32.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:adc1335b480ddd833aac3b0bd20f68ff0f3c3cf7a0bd337933b006d9f5cec40a\", size \"27375743\" in 1.419063654s" Sep 9 21:17:19.589904 containerd[1525]: time="2025-09-09T21:17:19.589885395Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.8\" returns image reference \"sha256:2cf30e39f99f8f4ee1a736a4f3175cc2d8d3f58936d8fa83ec5523658fdc7b8b\"" Sep 9 21:17:19.590445 containerd[1525]: time="2025-09-09T21:17:19.590332990Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 9 21:17:20.115339 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1447933368.mount: Deactivated successfully. Sep 9 21:17:20.753135 containerd[1525]: time="2025-09-09T21:17:20.753093441Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:20.753918 containerd[1525]: time="2025-09-09T21:17:20.753877896Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951624" Sep 9 21:17:20.754710 containerd[1525]: time="2025-09-09T21:17:20.754686084Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:20.758162 containerd[1525]: time="2025-09-09T21:17:20.758111345Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:20.759823 containerd[1525]: time="2025-09-09T21:17:20.759777107Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.169398453s" Sep 9 21:17:20.759871 containerd[1525]: time="2025-09-09T21:17:20.759827492Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Sep 9 21:17:20.760444 containerd[1525]: time="2025-09-09T21:17:20.760418120Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 9 21:17:21.230877 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1487159945.mount: Deactivated successfully. Sep 9 21:17:21.234370 containerd[1525]: time="2025-09-09T21:17:21.234335762Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 21:17:21.235068 containerd[1525]: time="2025-09-09T21:17:21.235044838Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Sep 9 21:17:21.235806 containerd[1525]: time="2025-09-09T21:17:21.235664519Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 21:17:21.238601 containerd[1525]: time="2025-09-09T21:17:21.237435570Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 21:17:21.238654 containerd[1525]: time="2025-09-09T21:17:21.238608917Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 478.15979ms" Sep 9 21:17:21.238654 containerd[1525]: time="2025-09-09T21:17:21.238636730Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Sep 9 21:17:21.239061 containerd[1525]: time="2025-09-09T21:17:21.239008611Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Sep 9 21:17:21.694243 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount443133696.mount: Deactivated successfully. Sep 9 21:17:23.430580 containerd[1525]: time="2025-09-09T21:17:23.430512271Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:23.431158 containerd[1525]: time="2025-09-09T21:17:23.431126258Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67943167" Sep 9 21:17:23.432043 containerd[1525]: time="2025-09-09T21:17:23.432017719Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:23.567159 containerd[1525]: time="2025-09-09T21:17:23.566996148Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:23.568281 containerd[1525]: time="2025-09-09T21:17:23.568158889Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 2.329123821s" Sep 9 21:17:23.568281 containerd[1525]: time="2025-09-09T21:17:23.568191345Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Sep 9 21:17:28.833696 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 9 21:17:28.835115 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:28.847382 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 9 21:17:28.847445 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 9 21:17:28.847848 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:28.849923 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:28.869806 systemd[1]: Reload requested from client PID 2211 ('systemctl') (unit session-7.scope)... Sep 9 21:17:28.869825 systemd[1]: Reloading... Sep 9 21:17:28.950868 zram_generator::config[2260]: No configuration found. Sep 9 21:17:29.197882 systemd[1]: Reloading finished in 327 ms. Sep 9 21:17:29.259195 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 9 21:17:29.259265 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 9 21:17:29.259475 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:29.259519 systemd[1]: kubelet.service: Consumed 86ms CPU time, 94.9M memory peak. Sep 9 21:17:29.260853 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:29.368226 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:29.371674 (kubelet)[2299]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 9 21:17:29.405223 kubelet[2299]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 21:17:29.405223 kubelet[2299]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 9 21:17:29.405223 kubelet[2299]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 21:17:29.405486 kubelet[2299]: I0909 21:17:29.405290 2299 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 9 21:17:30.225634 kubelet[2299]: I0909 21:17:30.225580 2299 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 9 21:17:30.225634 kubelet[2299]: I0909 21:17:30.225615 2299 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 9 21:17:30.225920 kubelet[2299]: I0909 21:17:30.225888 2299 server.go:954] "Client rotation is on, will bootstrap in background" Sep 9 21:17:30.246602 kubelet[2299]: E0909 21:17:30.246560 2299 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.66:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.66:6443: connect: connection refused" logger="UnhandledError" Sep 9 21:17:30.252821 kubelet[2299]: I0909 21:17:30.252704 2299 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 9 21:17:30.257968 kubelet[2299]: I0909 21:17:30.257947 2299 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 9 21:17:30.261124 kubelet[2299]: I0909 21:17:30.261096 2299 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 9 21:17:30.261747 kubelet[2299]: I0909 21:17:30.261693 2299 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 9 21:17:30.261928 kubelet[2299]: I0909 21:17:30.261735 2299 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 9 21:17:30.262064 kubelet[2299]: I0909 21:17:30.261996 2299 topology_manager.go:138] "Creating topology manager with none policy" Sep 9 21:17:30.262064 kubelet[2299]: I0909 21:17:30.262004 2299 container_manager_linux.go:304] "Creating device plugin manager" Sep 9 21:17:30.262205 kubelet[2299]: I0909 21:17:30.262189 2299 state_mem.go:36] "Initialized new in-memory state store" Sep 9 21:17:30.264497 kubelet[2299]: I0909 21:17:30.264462 2299 kubelet.go:446] "Attempting to sync node with API server" Sep 9 21:17:30.264497 kubelet[2299]: I0909 21:17:30.264493 2299 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 9 21:17:30.264559 kubelet[2299]: I0909 21:17:30.264516 2299 kubelet.go:352] "Adding apiserver pod source" Sep 9 21:17:30.264559 kubelet[2299]: I0909 21:17:30.264526 2299 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 9 21:17:30.267655 kubelet[2299]: I0909 21:17:30.267020 2299 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 9 21:17:30.267655 kubelet[2299]: I0909 21:17:30.267608 2299 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 9 21:17:30.267655 kubelet[2299]: W0909 21:17:30.267621 2299 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.66:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.66:6443: connect: connection refused Sep 9 21:17:30.267781 kubelet[2299]: E0909 21:17:30.267674 2299 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.66:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.66:6443: connect: connection refused" logger="UnhandledError" Sep 9 21:17:30.267781 kubelet[2299]: W0909 21:17:30.267723 2299 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 9 21:17:30.267781 kubelet[2299]: W0909 21:17:30.267684 2299 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.66:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.66:6443: connect: connection refused Sep 9 21:17:30.267862 kubelet[2299]: E0909 21:17:30.267774 2299 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.66:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.66:6443: connect: connection refused" logger="UnhandledError" Sep 9 21:17:30.268614 kubelet[2299]: I0909 21:17:30.268594 2299 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 9 21:17:30.268662 kubelet[2299]: I0909 21:17:30.268632 2299 server.go:1287] "Started kubelet" Sep 9 21:17:30.269148 kubelet[2299]: I0909 21:17:30.269115 2299 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 9 21:17:30.270064 kubelet[2299]: I0909 21:17:30.269991 2299 server.go:479] "Adding debug handlers to kubelet server" Sep 9 21:17:30.272710 kubelet[2299]: I0909 21:17:30.272644 2299 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 9 21:17:30.273094 kubelet[2299]: I0909 21:17:30.273065 2299 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 9 21:17:30.273999 kubelet[2299]: E0909 21:17:30.273727 2299 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.66:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.66:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1863b9dfc0d45048 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-09 21:17:30.268610632 +0000 UTC m=+0.894379604,LastTimestamp:2025-09-09 21:17:30.268610632 +0000 UTC m=+0.894379604,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 9 21:17:30.274474 kubelet[2299]: I0909 21:17:30.274448 2299 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 9 21:17:30.275106 kubelet[2299]: I0909 21:17:30.275086 2299 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 9 21:17:30.278560 kubelet[2299]: I0909 21:17:30.278140 2299 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 9 21:17:30.278560 kubelet[2299]: I0909 21:17:30.278247 2299 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 9 21:17:30.278560 kubelet[2299]: I0909 21:17:30.278320 2299 reconciler.go:26] "Reconciler: start to sync state" Sep 9 21:17:30.278560 kubelet[2299]: E0909 21:17:30.278496 2299 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 21:17:30.278704 kubelet[2299]: E0909 21:17:30.278582 2299 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.66:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.66:6443: connect: connection refused" interval="200ms" Sep 9 21:17:30.278704 kubelet[2299]: W0909 21:17:30.278597 2299 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.66:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.66:6443: connect: connection refused Sep 9 21:17:30.278704 kubelet[2299]: E0909 21:17:30.278642 2299 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.66:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.66:6443: connect: connection refused" logger="UnhandledError" Sep 9 21:17:30.279261 kubelet[2299]: I0909 21:17:30.279237 2299 factory.go:221] Registration of the systemd container factory successfully Sep 9 21:17:30.279359 kubelet[2299]: I0909 21:17:30.279328 2299 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 9 21:17:30.279880 kubelet[2299]: E0909 21:17:30.279863 2299 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 9 21:17:30.280062 kubelet[2299]: I0909 21:17:30.280045 2299 factory.go:221] Registration of the containerd container factory successfully Sep 9 21:17:30.290513 kubelet[2299]: I0909 21:17:30.290497 2299 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 9 21:17:30.290609 kubelet[2299]: I0909 21:17:30.290598 2299 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 9 21:17:30.290660 kubelet[2299]: I0909 21:17:30.290651 2299 state_mem.go:36] "Initialized new in-memory state store" Sep 9 21:17:30.293135 kubelet[2299]: I0909 21:17:30.293099 2299 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 9 21:17:30.294154 kubelet[2299]: I0909 21:17:30.294122 2299 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 9 21:17:30.294154 kubelet[2299]: I0909 21:17:30.294154 2299 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 9 21:17:30.294245 kubelet[2299]: I0909 21:17:30.294173 2299 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 9 21:17:30.294245 kubelet[2299]: I0909 21:17:30.294180 2299 kubelet.go:2382] "Starting kubelet main sync loop" Sep 9 21:17:30.294245 kubelet[2299]: E0909 21:17:30.294221 2299 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 9 21:17:30.370740 kubelet[2299]: I0909 21:17:30.370588 2299 policy_none.go:49] "None policy: Start" Sep 9 21:17:30.370740 kubelet[2299]: I0909 21:17:30.370617 2299 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 9 21:17:30.370740 kubelet[2299]: I0909 21:17:30.370631 2299 state_mem.go:35] "Initializing new in-memory state store" Sep 9 21:17:30.370740 kubelet[2299]: W0909 21:17:30.370619 2299 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.66:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.66:6443: connect: connection refused Sep 9 21:17:30.370740 kubelet[2299]: E0909 21:17:30.370685 2299 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.66:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.66:6443: connect: connection refused" logger="UnhandledError" Sep 9 21:17:30.377559 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 9 21:17:30.379590 kubelet[2299]: E0909 21:17:30.379544 2299 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 21:17:30.394782 kubelet[2299]: E0909 21:17:30.394724 2299 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 9 21:17:30.395192 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 9 21:17:30.398597 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 9 21:17:30.409544 kubelet[2299]: I0909 21:17:30.409479 2299 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 9 21:17:30.409755 kubelet[2299]: I0909 21:17:30.409662 2299 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 9 21:17:30.409755 kubelet[2299]: I0909 21:17:30.409673 2299 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 9 21:17:30.410212 kubelet[2299]: I0909 21:17:30.410136 2299 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 9 21:17:30.410991 kubelet[2299]: E0909 21:17:30.410965 2299 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 9 21:17:30.411041 kubelet[2299]: E0909 21:17:30.411012 2299 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 9 21:17:30.479536 kubelet[2299]: E0909 21:17:30.479425 2299 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.66:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.66:6443: connect: connection refused" interval="400ms" Sep 9 21:17:30.511968 kubelet[2299]: I0909 21:17:30.511906 2299 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 21:17:30.512414 kubelet[2299]: E0909 21:17:30.512368 2299 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.66:6443/api/v1/nodes\": dial tcp 10.0.0.66:6443: connect: connection refused" node="localhost" Sep 9 21:17:30.604300 systemd[1]: Created slice kubepods-burstable-pode7f756fcb231b8ddff623f611bf53702.slice - libcontainer container kubepods-burstable-pode7f756fcb231b8ddff623f611bf53702.slice. Sep 9 21:17:30.623486 kubelet[2299]: E0909 21:17:30.623462 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:30.626403 systemd[1]: Created slice kubepods-burstable-poda9176403b596d0b29ae8ad12d635226d.slice - libcontainer container kubepods-burstable-poda9176403b596d0b29ae8ad12d635226d.slice. Sep 9 21:17:30.642845 kubelet[2299]: E0909 21:17:30.642824 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:30.645705 systemd[1]: Created slice kubepods-burstable-poda88c9297c136b0f15880bf567e89a977.slice - libcontainer container kubepods-burstable-poda88c9297c136b0f15880bf567e89a977.slice. Sep 9 21:17:30.647243 kubelet[2299]: E0909 21:17:30.647221 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:30.680523 kubelet[2299]: I0909 21:17:30.680484 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:30.680583 kubelet[2299]: I0909 21:17:30.680527 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e7f756fcb231b8ddff623f611bf53702-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7f756fcb231b8ddff623f611bf53702\") " pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:30.680583 kubelet[2299]: I0909 21:17:30.680547 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:30.680640 kubelet[2299]: I0909 21:17:30.680606 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:30.680640 kubelet[2299]: I0909 21:17:30.680624 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:30.680682 kubelet[2299]: I0909 21:17:30.680638 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e7f756fcb231b8ddff623f611bf53702-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"e7f756fcb231b8ddff623f611bf53702\") " pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:30.680702 kubelet[2299]: I0909 21:17:30.680689 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:30.680720 kubelet[2299]: I0909 21:17:30.680706 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a9176403b596d0b29ae8ad12d635226d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a9176403b596d0b29ae8ad12d635226d\") " pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:30.680745 kubelet[2299]: I0909 21:17:30.680720 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e7f756fcb231b8ddff623f611bf53702-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7f756fcb231b8ddff623f611bf53702\") " pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:30.713579 kubelet[2299]: I0909 21:17:30.713522 2299 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 21:17:30.714060 kubelet[2299]: E0909 21:17:30.714028 2299 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.66:6443/api/v1/nodes\": dial tcp 10.0.0.66:6443: connect: connection refused" node="localhost" Sep 9 21:17:30.880480 kubelet[2299]: E0909 21:17:30.880380 2299 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.66:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.66:6443: connect: connection refused" interval="800ms" Sep 9 21:17:30.924866 kubelet[2299]: E0909 21:17:30.924808 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:30.925863 containerd[1525]: time="2025-09-09T21:17:30.925599852Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:e7f756fcb231b8ddff623f611bf53702,Namespace:kube-system,Attempt:0,}" Sep 9 21:17:30.942422 containerd[1525]: time="2025-09-09T21:17:30.942387783Z" level=info msg="connecting to shim f238ac418ddea969ed2f60f79307c3f4d9f12224cdb6fdae844c41d9dd1f41f1" address="unix:///run/containerd/s/29d55ae077e19f521e59b640bf9792d308a4899709364f92c16ad14d159fefd0" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:30.943874 kubelet[2299]: E0909 21:17:30.943829 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:30.944419 containerd[1525]: time="2025-09-09T21:17:30.944394721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a9176403b596d0b29ae8ad12d635226d,Namespace:kube-system,Attempt:0,}" Sep 9 21:17:30.949048 kubelet[2299]: E0909 21:17:30.948469 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:30.949333 containerd[1525]: time="2025-09-09T21:17:30.949302538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:a88c9297c136b0f15880bf567e89a977,Namespace:kube-system,Attempt:0,}" Sep 9 21:17:30.965941 systemd[1]: Started cri-containerd-f238ac418ddea969ed2f60f79307c3f4d9f12224cdb6fdae844c41d9dd1f41f1.scope - libcontainer container f238ac418ddea969ed2f60f79307c3f4d9f12224cdb6fdae844c41d9dd1f41f1. Sep 9 21:17:30.973573 containerd[1525]: time="2025-09-09T21:17:30.973523074Z" level=info msg="connecting to shim 0f7ff173e1028a634bfd932b8906f357fbcfeab18c9cce5079da0ab359c93486" address="unix:///run/containerd/s/5c542a900fa4476fafb769d119685d67a5627563bec5383903e80d67643ad6ea" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:30.983013 containerd[1525]: time="2025-09-09T21:17:30.982973973Z" level=info msg="connecting to shim f87a48e950ee11f880da9d3f9b799daf28d580cb1431b6fda7edf2d68e3e3704" address="unix:///run/containerd/s/ca39fbf8eb99580ab3f6fa9c1e239fd09ed4e08ee846a1e0fab0fb6c8915a4aa" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:31.002950 systemd[1]: Started cri-containerd-0f7ff173e1028a634bfd932b8906f357fbcfeab18c9cce5079da0ab359c93486.scope - libcontainer container 0f7ff173e1028a634bfd932b8906f357fbcfeab18c9cce5079da0ab359c93486. Sep 9 21:17:31.005674 systemd[1]: Started cri-containerd-f87a48e950ee11f880da9d3f9b799daf28d580cb1431b6fda7edf2d68e3e3704.scope - libcontainer container f87a48e950ee11f880da9d3f9b799daf28d580cb1431b6fda7edf2d68e3e3704. Sep 9 21:17:31.009803 containerd[1525]: time="2025-09-09T21:17:31.009635881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:e7f756fcb231b8ddff623f611bf53702,Namespace:kube-system,Attempt:0,} returns sandbox id \"f238ac418ddea969ed2f60f79307c3f4d9f12224cdb6fdae844c41d9dd1f41f1\"" Sep 9 21:17:31.012821 kubelet[2299]: E0909 21:17:31.011318 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:31.014873 containerd[1525]: time="2025-09-09T21:17:31.014819486Z" level=info msg="CreateContainer within sandbox \"f238ac418ddea969ed2f60f79307c3f4d9f12224cdb6fdae844c41d9dd1f41f1\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 9 21:17:31.022915 containerd[1525]: time="2025-09-09T21:17:31.022885759Z" level=info msg="Container 2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:31.029641 containerd[1525]: time="2025-09-09T21:17:31.029601055Z" level=info msg="CreateContainer within sandbox \"f238ac418ddea969ed2f60f79307c3f4d9f12224cdb6fdae844c41d9dd1f41f1\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906\"" Sep 9 21:17:31.031237 containerd[1525]: time="2025-09-09T21:17:31.031202408Z" level=info msg="StartContainer for \"2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906\"" Sep 9 21:17:31.032413 containerd[1525]: time="2025-09-09T21:17:31.032382269Z" level=info msg="connecting to shim 2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906" address="unix:///run/containerd/s/29d55ae077e19f521e59b640bf9792d308a4899709364f92c16ad14d159fefd0" protocol=ttrpc version=3 Sep 9 21:17:31.041891 containerd[1525]: time="2025-09-09T21:17:31.041861144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:a88c9297c136b0f15880bf567e89a977,Namespace:kube-system,Attempt:0,} returns sandbox id \"f87a48e950ee11f880da9d3f9b799daf28d580cb1431b6fda7edf2d68e3e3704\"" Sep 9 21:17:31.042505 kubelet[2299]: E0909 21:17:31.042464 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:31.047131 containerd[1525]: time="2025-09-09T21:17:31.046566829Z" level=info msg="CreateContainer within sandbox \"f87a48e950ee11f880da9d3f9b799daf28d580cb1431b6fda7edf2d68e3e3704\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 9 21:17:31.053102 containerd[1525]: time="2025-09-09T21:17:31.053073338Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a9176403b596d0b29ae8ad12d635226d,Namespace:kube-system,Attempt:0,} returns sandbox id \"0f7ff173e1028a634bfd932b8906f357fbcfeab18c9cce5079da0ab359c93486\"" Sep 9 21:17:31.053917 kubelet[2299]: E0909 21:17:31.053759 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:31.055885 containerd[1525]: time="2025-09-09T21:17:31.055855832Z" level=info msg="CreateContainer within sandbox \"0f7ff173e1028a634bfd932b8906f357fbcfeab18c9cce5079da0ab359c93486\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 9 21:17:31.057208 containerd[1525]: time="2025-09-09T21:17:31.057183415Z" level=info msg="Container 54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:31.057946 systemd[1]: Started cri-containerd-2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906.scope - libcontainer container 2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906. Sep 9 21:17:31.061747 containerd[1525]: time="2025-09-09T21:17:31.061719184Z" level=info msg="Container 252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:31.065010 containerd[1525]: time="2025-09-09T21:17:31.064983756Z" level=info msg="CreateContainer within sandbox \"f87a48e950ee11f880da9d3f9b799daf28d580cb1431b6fda7edf2d68e3e3704\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b\"" Sep 9 21:17:31.065446 containerd[1525]: time="2025-09-09T21:17:31.065408568Z" level=info msg="StartContainer for \"54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b\"" Sep 9 21:17:31.067575 containerd[1525]: time="2025-09-09T21:17:31.067547945Z" level=info msg="connecting to shim 54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b" address="unix:///run/containerd/s/ca39fbf8eb99580ab3f6fa9c1e239fd09ed4e08ee846a1e0fab0fb6c8915a4aa" protocol=ttrpc version=3 Sep 9 21:17:31.073306 containerd[1525]: time="2025-09-09T21:17:31.073255217Z" level=info msg="CreateContainer within sandbox \"0f7ff173e1028a634bfd932b8906f357fbcfeab18c9cce5079da0ab359c93486\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348\"" Sep 9 21:17:31.073735 containerd[1525]: time="2025-09-09T21:17:31.073717140Z" level=info msg="StartContainer for \"252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348\"" Sep 9 21:17:31.075225 containerd[1525]: time="2025-09-09T21:17:31.075185247Z" level=info msg="connecting to shim 252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348" address="unix:///run/containerd/s/5c542a900fa4476fafb769d119685d67a5627563bec5383903e80d67643ad6ea" protocol=ttrpc version=3 Sep 9 21:17:31.087957 systemd[1]: Started cri-containerd-54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b.scope - libcontainer container 54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b. Sep 9 21:17:31.093858 systemd[1]: Started cri-containerd-252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348.scope - libcontainer container 252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348. Sep 9 21:17:31.101951 containerd[1525]: time="2025-09-09T21:17:31.101913345Z" level=info msg="StartContainer for \"2c4c702d693a185e5e18627a364080af3d8f4ea8c61be28e722c4d690d849906\" returns successfully" Sep 9 21:17:31.116141 kubelet[2299]: I0909 21:17:31.115875 2299 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 21:17:31.116242 kubelet[2299]: E0909 21:17:31.116214 2299 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.66:6443/api/v1/nodes\": dial tcp 10.0.0.66:6443: connect: connection refused" node="localhost" Sep 9 21:17:31.141875 containerd[1525]: time="2025-09-09T21:17:31.140184633Z" level=info msg="StartContainer for \"252f0d30899de7a09acb2282b2c04ed7a2686e61866c7fa09382c2cd04568348\" returns successfully" Sep 9 21:17:31.143563 containerd[1525]: time="2025-09-09T21:17:31.143525265Z" level=info msg="StartContainer for \"54a0e8004036a93398988d8194e5f141240a6e5dd9b9db550e485813d070fe0b\" returns successfully" Sep 9 21:17:31.300810 kubelet[2299]: E0909 21:17:31.300332 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:31.301073 kubelet[2299]: E0909 21:17:31.301054 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:31.306445 kubelet[2299]: E0909 21:17:31.306419 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:31.306904 kubelet[2299]: E0909 21:17:31.306879 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:31.309636 kubelet[2299]: E0909 21:17:31.309614 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:31.309732 kubelet[2299]: E0909 21:17:31.309717 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:31.918666 kubelet[2299]: I0909 21:17:31.917810 2299 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 21:17:32.312607 kubelet[2299]: E0909 21:17:32.312514 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:32.312937 kubelet[2299]: E0909 21:17:32.312872 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:32.313468 kubelet[2299]: E0909 21:17:32.313445 2299 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 21:17:32.313560 kubelet[2299]: E0909 21:17:32.313544 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:32.681936 kubelet[2299]: E0909 21:17:32.681601 2299 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 9 21:17:32.739773 kubelet[2299]: I0909 21:17:32.739742 2299 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 9 21:17:32.778385 kubelet[2299]: I0909 21:17:32.778347 2299 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:32.833684 kubelet[2299]: E0909 21:17:32.833623 2299 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:32.833684 kubelet[2299]: I0909 21:17:32.833682 2299 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:32.835830 kubelet[2299]: E0909 21:17:32.835763 2299 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:32.835830 kubelet[2299]: I0909 21:17:32.835802 2299 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:32.837614 kubelet[2299]: E0909 21:17:32.837573 2299 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:33.267762 kubelet[2299]: I0909 21:17:33.267516 2299 apiserver.go:52] "Watching apiserver" Sep 9 21:17:33.279002 kubelet[2299]: I0909 21:17:33.278970 2299 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 9 21:17:33.313496 kubelet[2299]: I0909 21:17:33.313475 2299 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:33.315535 kubelet[2299]: E0909 21:17:33.315309 2299 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:33.315535 kubelet[2299]: E0909 21:17:33.315457 2299 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:34.807082 systemd[1]: Reload requested from client PID 2575 ('systemctl') (unit session-7.scope)... Sep 9 21:17:34.807411 systemd[1]: Reloading... Sep 9 21:17:34.869815 zram_generator::config[2618]: No configuration found. Sep 9 21:17:35.055953 systemd[1]: Reloading finished in 248 ms. Sep 9 21:17:35.080006 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:35.097741 systemd[1]: kubelet.service: Deactivated successfully. Sep 9 21:17:35.098029 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:35.098084 systemd[1]: kubelet.service: Consumed 1.245s CPU time, 129.3M memory peak. Sep 9 21:17:35.099645 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 21:17:35.240517 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 21:17:35.244220 (kubelet)[2660]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 9 21:17:35.280679 kubelet[2660]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 21:17:35.280679 kubelet[2660]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 9 21:17:35.280679 kubelet[2660]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 21:17:35.281131 kubelet[2660]: I0909 21:17:35.281096 2660 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 9 21:17:35.286453 kubelet[2660]: I0909 21:17:35.286416 2660 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 9 21:17:35.286453 kubelet[2660]: I0909 21:17:35.286442 2660 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 9 21:17:35.286682 kubelet[2660]: I0909 21:17:35.286654 2660 server.go:954] "Client rotation is on, will bootstrap in background" Sep 9 21:17:35.287775 kubelet[2660]: I0909 21:17:35.287744 2660 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 9 21:17:35.290041 kubelet[2660]: I0909 21:17:35.290013 2660 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 9 21:17:35.295563 kubelet[2660]: I0909 21:17:35.295538 2660 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 9 21:17:35.298127 kubelet[2660]: I0909 21:17:35.298100 2660 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 9 21:17:35.298350 kubelet[2660]: I0909 21:17:35.298316 2660 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 9 21:17:35.298509 kubelet[2660]: I0909 21:17:35.298342 2660 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 9 21:17:35.298579 kubelet[2660]: I0909 21:17:35.298512 2660 topology_manager.go:138] "Creating topology manager with none policy" Sep 9 21:17:35.298579 kubelet[2660]: I0909 21:17:35.298521 2660 container_manager_linux.go:304] "Creating device plugin manager" Sep 9 21:17:35.298579 kubelet[2660]: I0909 21:17:35.298559 2660 state_mem.go:36] "Initialized new in-memory state store" Sep 9 21:17:35.298689 kubelet[2660]: I0909 21:17:35.298678 2660 kubelet.go:446] "Attempting to sync node with API server" Sep 9 21:17:35.298711 kubelet[2660]: I0909 21:17:35.298693 2660 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 9 21:17:35.298731 kubelet[2660]: I0909 21:17:35.298712 2660 kubelet.go:352] "Adding apiserver pod source" Sep 9 21:17:35.298731 kubelet[2660]: I0909 21:17:35.298721 2660 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.299913 2660 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.300514 2660 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.301161 2660 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.301294 2660 server.go:1287] "Started kubelet" Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.301475 2660 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.301641 2660 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.302350 2660 server.go:479] "Adding debug handlers to kubelet server" Sep 9 21:17:35.302795 kubelet[2660]: I0909 21:17:35.302553 2660 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 9 21:17:35.304554 kubelet[2660]: E0909 21:17:35.304527 2660 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 9 21:17:35.308262 kubelet[2660]: I0909 21:17:35.308242 2660 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 9 21:17:35.308262 kubelet[2660]: I0909 21:17:35.308255 2660 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 9 21:17:35.308344 kubelet[2660]: I0909 21:17:35.308319 2660 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 9 21:17:35.309049 kubelet[2660]: I0909 21:17:35.309025 2660 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 9 21:17:35.309165 kubelet[2660]: I0909 21:17:35.309151 2660 reconciler.go:26] "Reconciler: start to sync state" Sep 9 21:17:35.312799 kubelet[2660]: E0909 21:17:35.312745 2660 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 21:17:35.313918 kubelet[2660]: I0909 21:17:35.313890 2660 factory.go:221] Registration of the systemd container factory successfully Sep 9 21:17:35.314006 kubelet[2660]: I0909 21:17:35.313978 2660 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 9 21:17:35.315049 kubelet[2660]: I0909 21:17:35.315019 2660 factory.go:221] Registration of the containerd container factory successfully Sep 9 21:17:35.323078 kubelet[2660]: I0909 21:17:35.323043 2660 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 9 21:17:35.324659 kubelet[2660]: I0909 21:17:35.324626 2660 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 9 21:17:35.324659 kubelet[2660]: I0909 21:17:35.324655 2660 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 9 21:17:35.324736 kubelet[2660]: I0909 21:17:35.324673 2660 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 9 21:17:35.324736 kubelet[2660]: I0909 21:17:35.324681 2660 kubelet.go:2382] "Starting kubelet main sync loop" Sep 9 21:17:35.324736 kubelet[2660]: E0909 21:17:35.324720 2660 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354621 2660 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354654 2660 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354674 2660 state_mem.go:36] "Initialized new in-memory state store" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354813 2660 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354823 2660 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354839 2660 policy_none.go:49] "None policy: Start" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354848 2660 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354857 2660 state_mem.go:35] "Initializing new in-memory state store" Sep 9 21:17:35.355618 kubelet[2660]: I0909 21:17:35.354943 2660 state_mem.go:75] "Updated machine memory state" Sep 9 21:17:35.358582 kubelet[2660]: I0909 21:17:35.358547 2660 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 9 21:17:35.358719 kubelet[2660]: I0909 21:17:35.358694 2660 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 9 21:17:35.358751 kubelet[2660]: I0909 21:17:35.358713 2660 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 9 21:17:35.359607 kubelet[2660]: I0909 21:17:35.359581 2660 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 9 21:17:35.360826 kubelet[2660]: E0909 21:17:35.360412 2660 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 9 21:17:35.426314 kubelet[2660]: I0909 21:17:35.426286 2660 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:35.426648 kubelet[2660]: I0909 21:17:35.426468 2660 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:35.426812 kubelet[2660]: I0909 21:17:35.426542 2660 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:35.462772 kubelet[2660]: I0909 21:17:35.462753 2660 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 21:17:35.469814 kubelet[2660]: I0909 21:17:35.469765 2660 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 9 21:17:35.469884 kubelet[2660]: I0909 21:17:35.469847 2660 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 9 21:17:35.510397 kubelet[2660]: I0909 21:17:35.510247 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:35.510397 kubelet[2660]: I0909 21:17:35.510282 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:35.510397 kubelet[2660]: I0909 21:17:35.510301 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:35.510397 kubelet[2660]: I0909 21:17:35.510318 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:35.510397 kubelet[2660]: I0909 21:17:35.510335 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a9176403b596d0b29ae8ad12d635226d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a9176403b596d0b29ae8ad12d635226d\") " pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:35.510566 kubelet[2660]: I0909 21:17:35.510354 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e7f756fcb231b8ddff623f611bf53702-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7f756fcb231b8ddff623f611bf53702\") " pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:35.510566 kubelet[2660]: I0909 21:17:35.510370 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e7f756fcb231b8ddff623f611bf53702-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"e7f756fcb231b8ddff623f611bf53702\") " pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:35.510566 kubelet[2660]: I0909 21:17:35.510385 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 21:17:35.510566 kubelet[2660]: I0909 21:17:35.510408 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e7f756fcb231b8ddff623f611bf53702-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7f756fcb231b8ddff623f611bf53702\") " pod="kube-system/kube-apiserver-localhost" Sep 9 21:17:35.731441 kubelet[2660]: E0909 21:17:35.731330 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:35.731672 kubelet[2660]: E0909 21:17:35.731647 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:35.732730 kubelet[2660]: E0909 21:17:35.732682 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:36.300065 kubelet[2660]: I0909 21:17:36.300013 2660 apiserver.go:52] "Watching apiserver" Sep 9 21:17:36.309537 kubelet[2660]: I0909 21:17:36.309511 2660 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 9 21:17:36.342609 kubelet[2660]: E0909 21:17:36.342387 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:36.342609 kubelet[2660]: I0909 21:17:36.342395 2660 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:36.344161 kubelet[2660]: E0909 21:17:36.342874 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:36.346605 kubelet[2660]: E0909 21:17:36.346584 2660 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 9 21:17:36.346784 kubelet[2660]: E0909 21:17:36.346769 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:36.367715 kubelet[2660]: I0909 21:17:36.367441 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.367425162 podStartE2EDuration="1.367425162s" podCreationTimestamp="2025-09-09 21:17:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:17:36.367423722 +0000 UTC m=+1.119979641" watchObservedRunningTime="2025-09-09 21:17:36.367425162 +0000 UTC m=+1.119981081" Sep 9 21:17:36.367715 kubelet[2660]: I0909 21:17:36.367619 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.367602738 podStartE2EDuration="1.367602738s" podCreationTimestamp="2025-09-09 21:17:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:17:36.360711555 +0000 UTC m=+1.113267474" watchObservedRunningTime="2025-09-09 21:17:36.367602738 +0000 UTC m=+1.120158657" Sep 9 21:17:36.374804 kubelet[2660]: I0909 21:17:36.374739 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.374713693 podStartE2EDuration="1.374713693s" podCreationTimestamp="2025-09-09 21:17:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:17:36.373962271 +0000 UTC m=+1.126518190" watchObservedRunningTime="2025-09-09 21:17:36.374713693 +0000 UTC m=+1.127269612" Sep 9 21:17:37.343215 kubelet[2660]: E0909 21:17:37.343164 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:37.343524 kubelet[2660]: E0909 21:17:37.343301 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:37.344272 kubelet[2660]: E0909 21:17:37.343546 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:38.580419 kubelet[2660]: E0909 21:17:38.580390 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:41.384587 kubelet[2660]: I0909 21:17:41.384462 2660 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 9 21:17:41.384935 containerd[1525]: time="2025-09-09T21:17:41.384872620Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 9 21:17:41.385116 kubelet[2660]: I0909 21:17:41.385092 2660 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 9 21:17:41.600252 kubelet[2660]: E0909 21:17:41.600181 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:41.827846 systemd[1]: Created slice kubepods-besteffort-pod72c007e7_0f0a_4da8_8a2b_fcef6855ddae.slice - libcontainer container kubepods-besteffort-pod72c007e7_0f0a_4da8_8a2b_fcef6855ddae.slice. Sep 9 21:17:41.852400 kubelet[2660]: I0909 21:17:41.852225 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/72c007e7-0f0a-4da8-8a2b-fcef6855ddae-lib-modules\") pod \"kube-proxy-j2rp5\" (UID: \"72c007e7-0f0a-4da8-8a2b-fcef6855ddae\") " pod="kube-system/kube-proxy-j2rp5" Sep 9 21:17:41.852400 kubelet[2660]: I0909 21:17:41.852306 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/72c007e7-0f0a-4da8-8a2b-fcef6855ddae-kube-proxy\") pod \"kube-proxy-j2rp5\" (UID: \"72c007e7-0f0a-4da8-8a2b-fcef6855ddae\") " pod="kube-system/kube-proxy-j2rp5" Sep 9 21:17:41.852400 kubelet[2660]: I0909 21:17:41.852334 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/72c007e7-0f0a-4da8-8a2b-fcef6855ddae-xtables-lock\") pod \"kube-proxy-j2rp5\" (UID: \"72c007e7-0f0a-4da8-8a2b-fcef6855ddae\") " pod="kube-system/kube-proxy-j2rp5" Sep 9 21:17:41.852400 kubelet[2660]: I0909 21:17:41.852352 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb9sc\" (UniqueName: \"kubernetes.io/projected/72c007e7-0f0a-4da8-8a2b-fcef6855ddae-kube-api-access-vb9sc\") pod \"kube-proxy-j2rp5\" (UID: \"72c007e7-0f0a-4da8-8a2b-fcef6855ddae\") " pod="kube-system/kube-proxy-j2rp5" Sep 9 21:17:42.138393 kubelet[2660]: E0909 21:17:42.138275 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:42.138812 containerd[1525]: time="2025-09-09T21:17:42.138761126Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-j2rp5,Uid:72c007e7-0f0a-4da8-8a2b-fcef6855ddae,Namespace:kube-system,Attempt:0,}" Sep 9 21:17:42.154456 containerd[1525]: time="2025-09-09T21:17:42.154415927Z" level=info msg="connecting to shim abbd56ff87b912485c175c93bdc195fc6e166b0e298c660c9619772ba53470b4" address="unix:///run/containerd/s/4886170931d1b3c9a33491f8e8d64f3ccb288fe1f1d602f1b831c1e3e20352fa" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:42.180011 systemd[1]: Started cri-containerd-abbd56ff87b912485c175c93bdc195fc6e166b0e298c660c9619772ba53470b4.scope - libcontainer container abbd56ff87b912485c175c93bdc195fc6e166b0e298c660c9619772ba53470b4. Sep 9 21:17:42.200028 containerd[1525]: time="2025-09-09T21:17:42.199691257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-j2rp5,Uid:72c007e7-0f0a-4da8-8a2b-fcef6855ddae,Namespace:kube-system,Attempt:0,} returns sandbox id \"abbd56ff87b912485c175c93bdc195fc6e166b0e298c660c9619772ba53470b4\"" Sep 9 21:17:42.200543 kubelet[2660]: E0909 21:17:42.200519 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:42.203700 containerd[1525]: time="2025-09-09T21:17:42.203668505Z" level=info msg="CreateContainer within sandbox \"abbd56ff87b912485c175c93bdc195fc6e166b0e298c660c9619772ba53470b4\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 9 21:17:42.216813 containerd[1525]: time="2025-09-09T21:17:42.215867723Z" level=info msg="Container bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:42.222702 containerd[1525]: time="2025-09-09T21:17:42.222642053Z" level=info msg="CreateContainer within sandbox \"abbd56ff87b912485c175c93bdc195fc6e166b0e298c660c9619772ba53470b4\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0\"" Sep 9 21:17:42.223956 containerd[1525]: time="2025-09-09T21:17:42.223930497Z" level=info msg="StartContainer for \"bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0\"" Sep 9 21:17:42.227130 containerd[1525]: time="2025-09-09T21:17:42.227084889Z" level=info msg="connecting to shim bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0" address="unix:///run/containerd/s/4886170931d1b3c9a33491f8e8d64f3ccb288fe1f1d602f1b831c1e3e20352fa" protocol=ttrpc version=3 Sep 9 21:17:42.248956 systemd[1]: Started cri-containerd-bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0.scope - libcontainer container bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0. Sep 9 21:17:42.281115 containerd[1525]: time="2025-09-09T21:17:42.281050655Z" level=info msg="StartContainer for \"bf32482fbf0269261c99ef1220b9cdf81ac83dd6247b4af0b3c6fd15d4435de0\" returns successfully" Sep 9 21:17:42.355091 kubelet[2660]: E0909 21:17:42.355062 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:42.355334 kubelet[2660]: E0909 21:17:42.355149 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:42.488947 kubelet[2660]: I0909 21:17:42.488815 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-j2rp5" podStartSLOduration=1.488794669 podStartE2EDuration="1.488794669s" podCreationTimestamp="2025-09-09 21:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:17:42.364459836 +0000 UTC m=+7.117015755" watchObservedRunningTime="2025-09-09 21:17:42.488794669 +0000 UTC m=+7.241350588" Sep 9 21:17:42.496702 systemd[1]: Created slice kubepods-besteffort-pod1e1110e4_d4b0_43fc_9778_01828fd61b8a.slice - libcontainer container kubepods-besteffort-pod1e1110e4_d4b0_43fc_9778_01828fd61b8a.slice. Sep 9 21:17:42.556780 kubelet[2660]: I0909 21:17:42.556711 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9wkn\" (UniqueName: \"kubernetes.io/projected/1e1110e4-d4b0-43fc-9778-01828fd61b8a-kube-api-access-r9wkn\") pod \"tigera-operator-755d956888-hh92k\" (UID: \"1e1110e4-d4b0-43fc-9778-01828fd61b8a\") " pod="tigera-operator/tigera-operator-755d956888-hh92k" Sep 9 21:17:42.556780 kubelet[2660]: I0909 21:17:42.556750 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/1e1110e4-d4b0-43fc-9778-01828fd61b8a-var-lib-calico\") pod \"tigera-operator-755d956888-hh92k\" (UID: \"1e1110e4-d4b0-43fc-9778-01828fd61b8a\") " pod="tigera-operator/tigera-operator-755d956888-hh92k" Sep 9 21:17:42.799150 containerd[1525]: time="2025-09-09T21:17:42.799034209Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-hh92k,Uid:1e1110e4-d4b0-43fc-9778-01828fd61b8a,Namespace:tigera-operator,Attempt:0,}" Sep 9 21:17:42.815695 containerd[1525]: time="2025-09-09T21:17:42.815658302Z" level=info msg="connecting to shim 40d379945af019a8d839498fe6a82c221df58ff86d346979daadf78b11c29021" address="unix:///run/containerd/s/68a230915bd9c1eaf61410b20c13ca60cc8c0e9d12ff692fb37465e7e95c2e69" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:42.835991 systemd[1]: Started cri-containerd-40d379945af019a8d839498fe6a82c221df58ff86d346979daadf78b11c29021.scope - libcontainer container 40d379945af019a8d839498fe6a82c221df58ff86d346979daadf78b11c29021. Sep 9 21:17:42.862904 containerd[1525]: time="2025-09-09T21:17:42.862855139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-hh92k,Uid:1e1110e4-d4b0-43fc-9778-01828fd61b8a,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"40d379945af019a8d839498fe6a82c221df58ff86d346979daadf78b11c29021\"" Sep 9 21:17:42.864592 containerd[1525]: time="2025-09-09T21:17:42.864550331Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 9 21:17:43.355960 kubelet[2660]: E0909 21:17:43.355937 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:44.383007 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3395673670.mount: Deactivated successfully. Sep 9 21:17:45.033341 containerd[1525]: time="2025-09-09T21:17:45.033300038Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:45.034484 containerd[1525]: time="2025-09-09T21:17:45.034453211Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=22152365" Sep 9 21:17:45.035461 containerd[1525]: time="2025-09-09T21:17:45.035407468Z" level=info msg="ImageCreate event name:\"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:45.039946 containerd[1525]: time="2025-09-09T21:17:45.039837603Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"22148360\" in 2.175255712s" Sep 9 21:17:45.039946 containerd[1525]: time="2025-09-09T21:17:45.039872242Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\"" Sep 9 21:17:45.040259 containerd[1525]: time="2025-09-09T21:17:45.040235353Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:45.042177 containerd[1525]: time="2025-09-09T21:17:45.042147348Z" level=info msg="CreateContainer within sandbox \"40d379945af019a8d839498fe6a82c221df58ff86d346979daadf78b11c29021\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 9 21:17:45.049676 containerd[1525]: time="2025-09-09T21:17:45.049180420Z" level=info msg="Container ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:45.055731 containerd[1525]: time="2025-09-09T21:17:45.055698385Z" level=info msg="CreateContainer within sandbox \"40d379945af019a8d839498fe6a82c221df58ff86d346979daadf78b11c29021\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c\"" Sep 9 21:17:45.056160 containerd[1525]: time="2025-09-09T21:17:45.056089376Z" level=info msg="StartContainer for \"ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c\"" Sep 9 21:17:45.056823 containerd[1525]: time="2025-09-09T21:17:45.056780799Z" level=info msg="connecting to shim ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c" address="unix:///run/containerd/s/68a230915bd9c1eaf61410b20c13ca60cc8c0e9d12ff692fb37465e7e95c2e69" protocol=ttrpc version=3 Sep 9 21:17:45.082924 systemd[1]: Started cri-containerd-ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c.scope - libcontainer container ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c. Sep 9 21:17:45.108816 containerd[1525]: time="2025-09-09T21:17:45.107285638Z" level=info msg="StartContainer for \"ec79aae350422de7345a6581336da7ad662b55e0b3a53ab055dcbba855b0457c\" returns successfully" Sep 9 21:17:45.374182 kubelet[2660]: I0909 21:17:45.374047 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-hh92k" podStartSLOduration=1.196882345 podStartE2EDuration="3.37403077s" podCreationTimestamp="2025-09-09 21:17:42 +0000 UTC" firstStartedPulling="2025-09-09 21:17:42.863979787 +0000 UTC m=+7.616535666" lastFinishedPulling="2025-09-09 21:17:45.041128172 +0000 UTC m=+9.793684091" observedRunningTime="2025-09-09 21:17:45.373383226 +0000 UTC m=+10.125939145" watchObservedRunningTime="2025-09-09 21:17:45.37403077 +0000 UTC m=+10.126586729" Sep 9 21:17:46.820954 kubelet[2660]: E0909 21:17:46.820906 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:48.589363 kubelet[2660]: E0909 21:17:48.589319 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:49.370702 kubelet[2660]: E0909 21:17:49.370667 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:50.152940 sudo[1742]: pam_unix(sudo:session): session closed for user root Sep 9 21:17:50.154138 sshd[1741]: Connection closed by 10.0.0.1 port 47772 Sep 9 21:17:50.156005 sshd-session[1738]: pam_unix(sshd:session): session closed for user core Sep 9 21:17:50.160614 systemd[1]: sshd@6-10.0.0.66:22-10.0.0.1:47772.service: Deactivated successfully. Sep 9 21:17:50.163490 systemd[1]: session-7.scope: Deactivated successfully. Sep 9 21:17:50.163852 systemd[1]: session-7.scope: Consumed 7.079s CPU time, 223M memory peak. Sep 9 21:17:50.164953 systemd-logind[1509]: Session 7 logged out. Waiting for processes to exit. Sep 9 21:17:50.167213 systemd-logind[1509]: Removed session 7. Sep 9 21:17:50.573747 update_engine[1511]: I20250909 21:17:50.573236 1511 update_attempter.cc:509] Updating boot flags... Sep 9 21:17:53.829534 systemd[1]: Created slice kubepods-besteffort-pod0df556b7_58b5_4582_8712_ea3f54da080d.slice - libcontainer container kubepods-besteffort-pod0df556b7_58b5_4582_8712_ea3f54da080d.slice. Sep 9 21:17:53.850518 kubelet[2660]: I0909 21:17:53.850466 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/0df556b7-58b5-4582-8712-ea3f54da080d-typha-certs\") pod \"calico-typha-54cd69d7d-hqn2k\" (UID: \"0df556b7-58b5-4582-8712-ea3f54da080d\") " pod="calico-system/calico-typha-54cd69d7d-hqn2k" Sep 9 21:17:53.850518 kubelet[2660]: I0909 21:17:53.850515 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0df556b7-58b5-4582-8712-ea3f54da080d-tigera-ca-bundle\") pod \"calico-typha-54cd69d7d-hqn2k\" (UID: \"0df556b7-58b5-4582-8712-ea3f54da080d\") " pod="calico-system/calico-typha-54cd69d7d-hqn2k" Sep 9 21:17:53.851003 kubelet[2660]: I0909 21:17:53.850534 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhkxj\" (UniqueName: \"kubernetes.io/projected/0df556b7-58b5-4582-8712-ea3f54da080d-kube-api-access-fhkxj\") pod \"calico-typha-54cd69d7d-hqn2k\" (UID: \"0df556b7-58b5-4582-8712-ea3f54da080d\") " pod="calico-system/calico-typha-54cd69d7d-hqn2k" Sep 9 21:17:54.046839 systemd[1]: Created slice kubepods-besteffort-pod0c906d6e_720a_4e85_b1cc_483453715bcd.slice - libcontainer container kubepods-besteffort-pod0c906d6e_720a_4e85_b1cc_483453715bcd.slice. Sep 9 21:17:54.138410 kubelet[2660]: E0909 21:17:54.138259 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:54.138728 containerd[1525]: time="2025-09-09T21:17:54.138694219Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-54cd69d7d-hqn2k,Uid:0df556b7-58b5-4582-8712-ea3f54da080d,Namespace:calico-system,Attempt:0,}" Sep 9 21:17:54.152976 kubelet[2660]: I0909 21:17:54.152919 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-policysync\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.152976 kubelet[2660]: I0909 21:17:54.152963 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-var-lib-calico\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153107 kubelet[2660]: I0909 21:17:54.152991 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-cni-log-dir\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153107 kubelet[2660]: I0909 21:17:54.153009 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-xtables-lock\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153107 kubelet[2660]: I0909 21:17:54.153061 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-cni-bin-dir\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153178 kubelet[2660]: I0909 21:17:54.153112 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-flexvol-driver-host\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153178 kubelet[2660]: I0909 21:17:54.153141 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c906d6e-720a-4e85-b1cc-483453715bcd-tigera-ca-bundle\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153219 kubelet[2660]: I0909 21:17:54.153184 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-cni-net-dir\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153219 kubelet[2660]: I0909 21:17:54.153207 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/0c906d6e-720a-4e85-b1cc-483453715bcd-node-certs\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153261 kubelet[2660]: I0909 21:17:54.153231 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-var-run-calico\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153283 kubelet[2660]: I0909 21:17:54.153261 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c906d6e-720a-4e85-b1cc-483453715bcd-lib-modules\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.153751 kubelet[2660]: I0909 21:17:54.153303 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w64r\" (UniqueName: \"kubernetes.io/projected/0c906d6e-720a-4e85-b1cc-483453715bcd-kube-api-access-5w64r\") pod \"calico-node-tp66b\" (UID: \"0c906d6e-720a-4e85-b1cc-483453715bcd\") " pod="calico-system/calico-node-tp66b" Sep 9 21:17:54.175720 containerd[1525]: time="2025-09-09T21:17:54.175675663Z" level=info msg="connecting to shim 089832aa22019b0e678c363b66bc6d767ba784e2539aadad5de96a4a96b61445" address="unix:///run/containerd/s/e7f7cdb5b5ed0a80300ed79992a1642180a3772a4ad297a0511d4de318171374" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:54.224005 systemd[1]: Started cri-containerd-089832aa22019b0e678c363b66bc6d767ba784e2539aadad5de96a4a96b61445.scope - libcontainer container 089832aa22019b0e678c363b66bc6d767ba784e2539aadad5de96a4a96b61445. Sep 9 21:17:54.256882 containerd[1525]: time="2025-09-09T21:17:54.256832363Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-54cd69d7d-hqn2k,Uid:0df556b7-58b5-4582-8712-ea3f54da080d,Namespace:calico-system,Attempt:0,} returns sandbox id \"089832aa22019b0e678c363b66bc6d767ba784e2539aadad5de96a4a96b61445\"" Sep 9 21:17:54.258548 kubelet[2660]: E0909 21:17:54.258518 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.258548 kubelet[2660]: W0909 21:17:54.258543 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.265083 kubelet[2660]: E0909 21:17:54.265012 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:54.269052 containerd[1525]: time="2025-09-09T21:17:54.269014060Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 9 21:17:54.271683 kubelet[2660]: E0909 21:17:54.271651 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.271683 kubelet[2660]: W0909 21:17:54.271677 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.274942 kubelet[2660]: E0909 21:17:54.274895 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.279863 kubelet[2660]: E0909 21:17:54.279814 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.280253 kubelet[2660]: E0909 21:17:54.280219 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.280253 kubelet[2660]: W0909 21:17:54.280239 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.280305 kubelet[2660]: E0909 21:17:54.280256 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.341745 kubelet[2660]: E0909 21:17:54.341697 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-q45ms" podUID="ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03" Sep 9 21:17:54.352829 containerd[1525]: time="2025-09-09T21:17:54.352690682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-tp66b,Uid:0c906d6e-720a-4e85-b1cc-483453715bcd,Namespace:calico-system,Attempt:0,}" Sep 9 21:17:54.384239 containerd[1525]: time="2025-09-09T21:17:54.384191808Z" level=info msg="connecting to shim ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad" address="unix:///run/containerd/s/bad8c6e66a1a5fae73144afbdd5ada42469da364b42e32847d99b13abfb0cbb4" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:17:54.408027 systemd[1]: Started cri-containerd-ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad.scope - libcontainer container ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad. Sep 9 21:17:54.438970 kubelet[2660]: E0909 21:17:54.438935 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.439185 kubelet[2660]: W0909 21:17:54.439058 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.439185 kubelet[2660]: E0909 21:17:54.439082 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.439437 kubelet[2660]: E0909 21:17:54.439411 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.439596 kubelet[2660]: W0909 21:17:54.439487 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.439596 kubelet[2660]: E0909 21:17:54.439552 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.439902 kubelet[2660]: E0909 21:17:54.439890 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.440120 kubelet[2660]: W0909 21:17:54.439970 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.440120 kubelet[2660]: E0909 21:17:54.439985 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.440544 kubelet[2660]: E0909 21:17:54.440487 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.440544 kubelet[2660]: W0909 21:17:54.440500 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.440544 kubelet[2660]: E0909 21:17:54.440511 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.440883 kubelet[2660]: E0909 21:17:54.440860 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.441031 kubelet[2660]: W0909 21:17:54.440964 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.441031 kubelet[2660]: E0909 21:17:54.440981 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.441310 kubelet[2660]: E0909 21:17:54.441239 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.441310 kubelet[2660]: W0909 21:17:54.441250 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.441310 kubelet[2660]: E0909 21:17:54.441259 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.441639 kubelet[2660]: E0909 21:17:54.441574 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.441639 kubelet[2660]: W0909 21:17:54.441586 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.441639 kubelet[2660]: E0909 21:17:54.441596 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.442169 kubelet[2660]: E0909 21:17:54.442032 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.442169 kubelet[2660]: W0909 21:17:54.442071 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.442169 kubelet[2660]: E0909 21:17:54.442083 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.442667 kubelet[2660]: E0909 21:17:54.442640 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.442732 kubelet[2660]: W0909 21:17:54.442719 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.442783 kubelet[2660]: E0909 21:17:54.442772 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.443108 kubelet[2660]: E0909 21:17:54.443062 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.443108 kubelet[2660]: W0909 21:17:54.443077 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.443108 kubelet[2660]: E0909 21:17:54.443087 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.443463 kubelet[2660]: E0909 21:17:54.443450 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.443547 kubelet[2660]: W0909 21:17:54.443536 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.443613 kubelet[2660]: E0909 21:17:54.443602 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.443964 kubelet[2660]: E0909 21:17:54.443850 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.443964 kubelet[2660]: W0909 21:17:54.443866 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.443964 kubelet[2660]: E0909 21:17:54.443876 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.444127 kubelet[2660]: E0909 21:17:54.444102 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.444187 kubelet[2660]: W0909 21:17:54.444176 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.444257 kubelet[2660]: E0909 21:17:54.444246 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.444469 kubelet[2660]: E0909 21:17:54.444442 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.444590 kubelet[2660]: W0909 21:17:54.444577 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.444641 kubelet[2660]: E0909 21:17:54.444630 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.444988 kubelet[2660]: E0909 21:17:54.444975 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.445088 kubelet[2660]: W0909 21:17:54.445075 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.445252 kubelet[2660]: E0909 21:17:54.445188 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.445538 kubelet[2660]: E0909 21:17:54.445525 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.445691 kubelet[2660]: W0909 21:17:54.445608 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.445691 kubelet[2660]: E0909 21:17:54.445624 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.445913 kubelet[2660]: E0909 21:17:54.445899 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.445972 kubelet[2660]: W0909 21:17:54.445960 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.446038 kubelet[2660]: E0909 21:17:54.446026 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.446382 kubelet[2660]: E0909 21:17:54.446331 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.446382 kubelet[2660]: W0909 21:17:54.446345 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.446563 kubelet[2660]: E0909 21:17:54.446467 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.446721 kubelet[2660]: E0909 21:17:54.446704 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.446865 kubelet[2660]: W0909 21:17:54.446850 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.447129 kubelet[2660]: E0909 21:17:54.447039 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.447321 kubelet[2660]: E0909 21:17:54.447298 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.447404 kubelet[2660]: W0909 21:17:54.447393 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.447476 kubelet[2660]: E0909 21:17:54.447453 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.452597 containerd[1525]: time="2025-09-09T21:17:54.452562220Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-tp66b,Uid:0c906d6e-720a-4e85-b1cc-483453715bcd,Namespace:calico-system,Attempt:0,} returns sandbox id \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\"" Sep 9 21:17:54.456705 kubelet[2660]: E0909 21:17:54.456569 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.456705 kubelet[2660]: W0909 21:17:54.456584 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.456705 kubelet[2660]: E0909 21:17:54.456595 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.456705 kubelet[2660]: I0909 21:17:54.456620 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03-kubelet-dir\") pod \"csi-node-driver-q45ms\" (UID: \"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03\") " pod="calico-system/csi-node-driver-q45ms" Sep 9 21:17:54.456919 kubelet[2660]: E0909 21:17:54.456903 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.456983 kubelet[2660]: W0909 21:17:54.456971 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.457055 kubelet[2660]: E0909 21:17:54.457046 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.457152 kubelet[2660]: I0909 21:17:54.457107 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03-socket-dir\") pod \"csi-node-driver-q45ms\" (UID: \"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03\") " pod="calico-system/csi-node-driver-q45ms" Sep 9 21:17:54.457290 kubelet[2660]: E0909 21:17:54.457274 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.457322 kubelet[2660]: W0909 21:17:54.457291 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.457322 kubelet[2660]: E0909 21:17:54.457310 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.457448 kubelet[2660]: E0909 21:17:54.457436 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.457448 kubelet[2660]: W0909 21:17:54.457446 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.457448 kubelet[2660]: E0909 21:17:54.457461 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.457603 kubelet[2660]: E0909 21:17:54.457591 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.457603 kubelet[2660]: W0909 21:17:54.457601 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.457648 kubelet[2660]: E0909 21:17:54.457614 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.457648 kubelet[2660]: I0909 21:17:54.457632 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03-registration-dir\") pod \"csi-node-driver-q45ms\" (UID: \"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03\") " pod="calico-system/csi-node-driver-q45ms" Sep 9 21:17:54.457899 kubelet[2660]: E0909 21:17:54.457850 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.457899 kubelet[2660]: W0909 21:17:54.457864 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.457899 kubelet[2660]: E0909 21:17:54.457877 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.457899 kubelet[2660]: I0909 21:17:54.457891 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dkdk\" (UniqueName: \"kubernetes.io/projected/ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03-kube-api-access-5dkdk\") pod \"csi-node-driver-q45ms\" (UID: \"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03\") " pod="calico-system/csi-node-driver-q45ms" Sep 9 21:17:54.458062 kubelet[2660]: E0909 21:17:54.458024 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458062 kubelet[2660]: W0909 21:17:54.458033 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.458062 kubelet[2660]: E0909 21:17:54.458049 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.458149 kubelet[2660]: I0909 21:17:54.458063 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03-varrun\") pod \"csi-node-driver-q45ms\" (UID: \"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03\") " pod="calico-system/csi-node-driver-q45ms" Sep 9 21:17:54.458224 kubelet[2660]: E0909 21:17:54.458216 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458250 kubelet[2660]: W0909 21:17:54.458224 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.458250 kubelet[2660]: E0909 21:17:54.458238 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.458391 kubelet[2660]: E0909 21:17:54.458381 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458391 kubelet[2660]: W0909 21:17:54.458391 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.458472 kubelet[2660]: E0909 21:17:54.458417 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.458531 kubelet[2660]: E0909 21:17:54.458521 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458531 kubelet[2660]: W0909 21:17:54.458531 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.458601 kubelet[2660]: E0909 21:17:54.458552 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.458662 kubelet[2660]: E0909 21:17:54.458651 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458662 kubelet[2660]: W0909 21:17:54.458660 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.458731 kubelet[2660]: E0909 21:17:54.458672 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.458820 kubelet[2660]: E0909 21:17:54.458810 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458847 kubelet[2660]: W0909 21:17:54.458820 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.458847 kubelet[2660]: E0909 21:17:54.458832 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.458956 kubelet[2660]: E0909 21:17:54.458946 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.458956 kubelet[2660]: W0909 21:17:54.458956 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.459023 kubelet[2660]: E0909 21:17:54.458963 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.459102 kubelet[2660]: E0909 21:17:54.459091 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.459102 kubelet[2660]: W0909 21:17:54.459101 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.459153 kubelet[2660]: E0909 21:17:54.459108 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.459239 kubelet[2660]: E0909 21:17:54.459228 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.459239 kubelet[2660]: W0909 21:17:54.459238 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.459290 kubelet[2660]: E0909 21:17:54.459246 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.558478 kubelet[2660]: E0909 21:17:54.558452 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.558478 kubelet[2660]: W0909 21:17:54.558473 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.558604 kubelet[2660]: E0909 21:17:54.558492 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.558712 kubelet[2660]: E0909 21:17:54.558700 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.558742 kubelet[2660]: W0909 21:17:54.558712 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.558742 kubelet[2660]: E0909 21:17:54.558731 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.558936 kubelet[2660]: E0909 21:17:54.558924 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.558936 kubelet[2660]: W0909 21:17:54.558936 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.558988 kubelet[2660]: E0909 21:17:54.558950 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.559157 kubelet[2660]: E0909 21:17:54.559146 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.559157 kubelet[2660]: W0909 21:17:54.559157 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.559210 kubelet[2660]: E0909 21:17:54.559173 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.559367 kubelet[2660]: E0909 21:17:54.559346 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.559452 kubelet[2660]: W0909 21:17:54.559366 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.559452 kubelet[2660]: E0909 21:17:54.559381 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.559568 kubelet[2660]: E0909 21:17:54.559555 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.559568 kubelet[2660]: W0909 21:17:54.559567 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.559615 kubelet[2660]: E0909 21:17:54.559586 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.559733 kubelet[2660]: E0909 21:17:54.559722 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.559733 kubelet[2660]: W0909 21:17:54.559733 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.559811 kubelet[2660]: E0909 21:17:54.559782 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.559901 kubelet[2660]: E0909 21:17:54.559888 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.559901 kubelet[2660]: W0909 21:17:54.559899 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.559957 kubelet[2660]: E0909 21:17:54.559919 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.560120 kubelet[2660]: E0909 21:17:54.560042 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.560120 kubelet[2660]: W0909 21:17:54.560054 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.560120 kubelet[2660]: E0909 21:17:54.560088 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.560220 kubelet[2660]: E0909 21:17:54.560185 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.560220 kubelet[2660]: W0909 21:17:54.560192 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.560220 kubelet[2660]: E0909 21:17:54.560205 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.560410 kubelet[2660]: E0909 21:17:54.560397 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.560410 kubelet[2660]: W0909 21:17:54.560409 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.560753 kubelet[2660]: E0909 21:17:54.560486 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.560753 kubelet[2660]: E0909 21:17:54.560543 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.560753 kubelet[2660]: W0909 21:17:54.560548 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.560753 kubelet[2660]: E0909 21:17:54.560579 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.560753 kubelet[2660]: E0909 21:17:54.560738 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.560753 kubelet[2660]: W0909 21:17:54.560746 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.560753 kubelet[2660]: E0909 21:17:54.560762 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.561237 kubelet[2660]: E0909 21:17:54.561221 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.561318 kubelet[2660]: W0909 21:17:54.561305 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.561461 kubelet[2660]: E0909 21:17:54.561396 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.561680 kubelet[2660]: E0909 21:17:54.561665 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.561797 kubelet[2660]: W0909 21:17:54.561758 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.561877 kubelet[2660]: E0909 21:17:54.561858 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.562264 kubelet[2660]: E0909 21:17:54.562166 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.562264 kubelet[2660]: W0909 21:17:54.562180 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.562264 kubelet[2660]: E0909 21:17:54.562228 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.562679 kubelet[2660]: E0909 21:17:54.562559 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.562679 kubelet[2660]: W0909 21:17:54.562588 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.562679 kubelet[2660]: E0909 21:17:54.562619 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.562900 kubelet[2660]: E0909 21:17:54.562882 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.562955 kubelet[2660]: W0909 21:17:54.562944 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.563028 kubelet[2660]: E0909 21:17:54.563009 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.566782 kubelet[2660]: E0909 21:17:54.566749 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.566946 kubelet[2660]: W0909 21:17:54.566929 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.567046 kubelet[2660]: E0909 21:17:54.567024 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.567479 kubelet[2660]: E0909 21:17:54.567427 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.567897 kubelet[2660]: W0909 21:17:54.567567 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.567897 kubelet[2660]: E0909 21:17:54.567599 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.568333 kubelet[2660]: E0909 21:17:54.568096 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.568333 kubelet[2660]: W0909 21:17:54.568112 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.568333 kubelet[2660]: E0909 21:17:54.568143 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.568592 kubelet[2660]: E0909 21:17:54.568575 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.568649 kubelet[2660]: W0909 21:17:54.568637 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.568737 kubelet[2660]: E0909 21:17:54.568718 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.568913 kubelet[2660]: E0909 21:17:54.568899 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.569094 kubelet[2660]: W0909 21:17:54.568970 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.569137 kubelet[2660]: E0909 21:17:54.569105 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.569361 kubelet[2660]: E0909 21:17:54.569339 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.569437 kubelet[2660]: W0909 21:17:54.569424 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.569573 kubelet[2660]: E0909 21:17:54.569499 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.569882 kubelet[2660]: E0909 21:17:54.569865 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.570041 kubelet[2660]: W0909 21:17:54.569992 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.570041 kubelet[2660]: E0909 21:17:54.570010 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:54.579630 kubelet[2660]: E0909 21:17:54.579607 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:54.579630 kubelet[2660]: W0909 21:17:54.579627 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:54.579731 kubelet[2660]: E0909 21:17:54.579644 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:55.315186 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3768016738.mount: Deactivated successfully. Sep 9 21:17:55.919832 containerd[1525]: time="2025-09-09T21:17:55.919763763Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:55.921137 containerd[1525]: time="2025-09-09T21:17:55.921089344Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=33105775" Sep 9 21:17:55.922487 containerd[1525]: time="2025-09-09T21:17:55.922454365Z" level=info msg="ImageCreate event name:\"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:55.924574 containerd[1525]: time="2025-09-09T21:17:55.924542975Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:55.925180 containerd[1525]: time="2025-09-09T21:17:55.925014528Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"33105629\" in 1.655941629s" Sep 9 21:17:55.925180 containerd[1525]: time="2025-09-09T21:17:55.925044967Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\"" Sep 9 21:17:55.927534 containerd[1525]: time="2025-09-09T21:17:55.927505412Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 9 21:17:55.959397 containerd[1525]: time="2025-09-09T21:17:55.959058520Z" level=info msg="CreateContainer within sandbox \"089832aa22019b0e678c363b66bc6d767ba784e2539aadad5de96a4a96b61445\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 9 21:17:55.966542 containerd[1525]: time="2025-09-09T21:17:55.966505613Z" level=info msg="Container 74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:55.978131 containerd[1525]: time="2025-09-09T21:17:55.978088847Z" level=info msg="CreateContainer within sandbox \"089832aa22019b0e678c363b66bc6d767ba784e2539aadad5de96a4a96b61445\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94\"" Sep 9 21:17:55.980960 containerd[1525]: time="2025-09-09T21:17:55.980927726Z" level=info msg="StartContainer for \"74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94\"" Sep 9 21:17:55.981948 containerd[1525]: time="2025-09-09T21:17:55.981913832Z" level=info msg="connecting to shim 74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94" address="unix:///run/containerd/s/e7f7cdb5b5ed0a80300ed79992a1642180a3772a4ad297a0511d4de318171374" protocol=ttrpc version=3 Sep 9 21:17:56.012937 systemd[1]: Started cri-containerd-74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94.scope - libcontainer container 74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94. Sep 9 21:17:56.048331 containerd[1525]: time="2025-09-09T21:17:56.048293111Z" level=info msg="StartContainer for \"74fb308bc4b71f30ca1edbbb41657e37bfa6233698c27b4d01ce6b1e94619e94\" returns successfully" Sep 9 21:17:56.325581 kubelet[2660]: E0909 21:17:56.325538 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-q45ms" podUID="ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03" Sep 9 21:17:56.386934 kubelet[2660]: E0909 21:17:56.386911 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:56.396835 kubelet[2660]: I0909 21:17:56.396756 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-54cd69d7d-hqn2k" podStartSLOduration=1.7374516020000002 podStartE2EDuration="3.396743982s" podCreationTimestamp="2025-09-09 21:17:53 +0000 UTC" firstStartedPulling="2025-09-09 21:17:54.268057154 +0000 UTC m=+19.020613073" lastFinishedPulling="2025-09-09 21:17:55.927349534 +0000 UTC m=+20.679905453" observedRunningTime="2025-09-09 21:17:56.396634024 +0000 UTC m=+21.149189943" watchObservedRunningTime="2025-09-09 21:17:56.396743982 +0000 UTC m=+21.149299901" Sep 9 21:17:56.457938 kubelet[2660]: E0909 21:17:56.457911 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.457938 kubelet[2660]: W0909 21:17:56.457934 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458070 kubelet[2660]: E0909 21:17:56.457954 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.458132 kubelet[2660]: E0909 21:17:56.458117 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.458132 kubelet[2660]: W0909 21:17:56.458128 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458192 kubelet[2660]: E0909 21:17:56.458138 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.458286 kubelet[2660]: E0909 21:17:56.458274 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.458286 kubelet[2660]: W0909 21:17:56.458284 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458356 kubelet[2660]: E0909 21:17:56.458293 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.458433 kubelet[2660]: E0909 21:17:56.458420 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.458433 kubelet[2660]: W0909 21:17:56.458430 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458497 kubelet[2660]: E0909 21:17:56.458438 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.458573 kubelet[2660]: E0909 21:17:56.458561 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.458573 kubelet[2660]: W0909 21:17:56.458572 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458628 kubelet[2660]: E0909 21:17:56.458580 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.458701 kubelet[2660]: E0909 21:17:56.458689 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.458701 kubelet[2660]: W0909 21:17:56.458699 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458763 kubelet[2660]: E0909 21:17:56.458707 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.458860 kubelet[2660]: E0909 21:17:56.458848 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.458860 kubelet[2660]: W0909 21:17:56.458858 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.458923 kubelet[2660]: E0909 21:17:56.458866 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459028 kubelet[2660]: E0909 21:17:56.459010 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459028 kubelet[2660]: W0909 21:17:56.459020 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459028 kubelet[2660]: E0909 21:17:56.459028 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459182 kubelet[2660]: E0909 21:17:56.459165 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459182 kubelet[2660]: W0909 21:17:56.459175 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459182 kubelet[2660]: E0909 21:17:56.459182 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459305 kubelet[2660]: E0909 21:17:56.459292 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459305 kubelet[2660]: W0909 21:17:56.459303 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459387 kubelet[2660]: E0909 21:17:56.459310 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459475 kubelet[2660]: E0909 21:17:56.459460 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459475 kubelet[2660]: W0909 21:17:56.459473 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459545 kubelet[2660]: E0909 21:17:56.459482 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459622 kubelet[2660]: E0909 21:17:56.459610 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459622 kubelet[2660]: W0909 21:17:56.459620 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459676 kubelet[2660]: E0909 21:17:56.459628 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459797 kubelet[2660]: E0909 21:17:56.459767 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459797 kubelet[2660]: W0909 21:17:56.459776 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459872 kubelet[2660]: E0909 21:17:56.459810 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.459948 kubelet[2660]: E0909 21:17:56.459933 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.459948 kubelet[2660]: W0909 21:17:56.459943 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.459990 kubelet[2660]: E0909 21:17:56.459951 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.460118 kubelet[2660]: E0909 21:17:56.460106 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.460145 kubelet[2660]: W0909 21:17:56.460118 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.460145 kubelet[2660]: E0909 21:17:56.460127 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.473803 kubelet[2660]: E0909 21:17:56.473769 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.473803 kubelet[2660]: W0909 21:17:56.473794 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.473803 kubelet[2660]: E0909 21:17:56.473807 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.474025 kubelet[2660]: E0909 21:17:56.474010 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.474025 kubelet[2660]: W0909 21:17:56.474022 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.474066 kubelet[2660]: E0909 21:17:56.474037 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.474225 kubelet[2660]: E0909 21:17:56.474213 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.474225 kubelet[2660]: W0909 21:17:56.474224 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.474276 kubelet[2660]: E0909 21:17:56.474238 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.474431 kubelet[2660]: E0909 21:17:56.474418 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.474431 kubelet[2660]: W0909 21:17:56.474429 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.474482 kubelet[2660]: E0909 21:17:56.474444 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.474606 kubelet[2660]: E0909 21:17:56.474594 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.474633 kubelet[2660]: W0909 21:17:56.474611 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.474633 kubelet[2660]: E0909 21:17:56.474624 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.474754 kubelet[2660]: E0909 21:17:56.474743 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.474777 kubelet[2660]: W0909 21:17:56.474753 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.474777 kubelet[2660]: E0909 21:17:56.474765 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.474949 kubelet[2660]: E0909 21:17:56.474936 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.474949 kubelet[2660]: W0909 21:17:56.474946 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.475002 kubelet[2660]: E0909 21:17:56.474959 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.475171 kubelet[2660]: E0909 21:17:56.475154 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.475171 kubelet[2660]: W0909 21:17:56.475169 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.475218 kubelet[2660]: E0909 21:17:56.475188 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.475360 kubelet[2660]: E0909 21:17:56.475348 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.475360 kubelet[2660]: W0909 21:17:56.475359 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.475433 kubelet[2660]: E0909 21:17:56.475410 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.475541 kubelet[2660]: E0909 21:17:56.475529 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.475541 kubelet[2660]: W0909 21:17:56.475539 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.475591 kubelet[2660]: E0909 21:17:56.475582 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.475682 kubelet[2660]: E0909 21:17:56.475670 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.475682 kubelet[2660]: W0909 21:17:56.475681 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.475726 kubelet[2660]: E0909 21:17:56.475697 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.475858 kubelet[2660]: E0909 21:17:56.475847 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.475889 kubelet[2660]: W0909 21:17:56.475858 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.475889 kubelet[2660]: E0909 21:17:56.475874 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.476049 kubelet[2660]: E0909 21:17:56.476037 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.476049 kubelet[2660]: W0909 21:17:56.476048 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.476099 kubelet[2660]: E0909 21:17:56.476060 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.476277 kubelet[2660]: E0909 21:17:56.476263 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.476305 kubelet[2660]: W0909 21:17:56.476277 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.476305 kubelet[2660]: E0909 21:17:56.476288 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.476845 kubelet[2660]: E0909 21:17:56.476827 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.481719 kubelet[2660]: W0909 21:17:56.481682 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.481751 kubelet[2660]: E0909 21:17:56.481727 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.481985 kubelet[2660]: E0909 21:17:56.481956 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.481985 kubelet[2660]: W0909 21:17:56.481973 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.482043 kubelet[2660]: E0909 21:17:56.481990 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.482424 kubelet[2660]: E0909 21:17:56.482397 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.482424 kubelet[2660]: W0909 21:17:56.482411 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.482424 kubelet[2660]: E0909 21:17:56.482422 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.482762 kubelet[2660]: E0909 21:17:56.482747 2660 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 9 21:17:56.482762 kubelet[2660]: W0909 21:17:56.482760 2660 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 9 21:17:56.482835 kubelet[2660]: E0909 21:17:56.482770 2660 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 9 21:17:56.952589 containerd[1525]: time="2025-09-09T21:17:56.952541736Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:56.953539 containerd[1525]: time="2025-09-09T21:17:56.953279926Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4266814" Sep 9 21:17:56.954230 containerd[1525]: time="2025-09-09T21:17:56.954197553Z" level=info msg="ImageCreate event name:\"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:56.956194 containerd[1525]: time="2025-09-09T21:17:56.956167006Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:17:56.956667 containerd[1525]: time="2025-09-09T21:17:56.956631040Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5636015\" in 1.029086828s" Sep 9 21:17:56.956667 containerd[1525]: time="2025-09-09T21:17:56.956665639Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\"" Sep 9 21:17:56.958907 containerd[1525]: time="2025-09-09T21:17:56.958867449Z" level=info msg="CreateContainer within sandbox \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 9 21:17:56.974731 containerd[1525]: time="2025-09-09T21:17:56.974697473Z" level=info msg="Container 0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:17:56.985069 containerd[1525]: time="2025-09-09T21:17:56.985018411Z" level=info msg="CreateContainer within sandbox \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\"" Sep 9 21:17:56.985923 containerd[1525]: time="2025-09-09T21:17:56.985597963Z" level=info msg="StartContainer for \"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\"" Sep 9 21:17:56.989190 containerd[1525]: time="2025-09-09T21:17:56.989160395Z" level=info msg="connecting to shim 0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9" address="unix:///run/containerd/s/bad8c6e66a1a5fae73144afbdd5ada42469da364b42e32847d99b13abfb0cbb4" protocol=ttrpc version=3 Sep 9 21:17:57.007936 systemd[1]: Started cri-containerd-0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9.scope - libcontainer container 0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9. Sep 9 21:17:57.037401 containerd[1525]: time="2025-09-09T21:17:57.037290078Z" level=info msg="StartContainer for \"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\" returns successfully" Sep 9 21:17:57.050344 systemd[1]: cri-containerd-0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9.scope: Deactivated successfully. Sep 9 21:17:57.059405 containerd[1525]: time="2025-09-09T21:17:57.059292430Z" level=info msg="received exit event container_id:\"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\" id:\"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\" pid:3363 exited_at:{seconds:1757452677 nanos:54113858}" Sep 9 21:17:57.059542 containerd[1525]: time="2025-09-09T21:17:57.059367469Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\" id:\"0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9\" pid:3363 exited_at:{seconds:1757452677 nanos:54113858}" Sep 9 21:17:57.089895 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0be77c5bfd3a39bb75537d0891a7c1e06bf9db6f366890e7b3088888ad3c5fd9-rootfs.mount: Deactivated successfully. Sep 9 21:17:57.390664 kubelet[2660]: I0909 21:17:57.390053 2660 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 9 21:17:57.390664 kubelet[2660]: E0909 21:17:57.390338 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:17:57.392751 containerd[1525]: time="2025-09-09T21:17:57.392715271Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 9 21:17:58.325895 kubelet[2660]: E0909 21:17:58.325846 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-q45ms" podUID="ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03" Sep 9 21:18:00.325305 kubelet[2660]: E0909 21:18:00.325253 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-q45ms" podUID="ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03" Sep 9 21:18:01.037687 containerd[1525]: time="2025-09-09T21:18:01.037647749Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:01.038115 containerd[1525]: time="2025-09-09T21:18:01.038091304Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=65913477" Sep 9 21:18:01.038978 containerd[1525]: time="2025-09-09T21:18:01.038930254Z" level=info msg="ImageCreate event name:\"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:01.044808 containerd[1525]: time="2025-09-09T21:18:01.040835033Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:01.044910 containerd[1525]: time="2025-09-09T21:18:01.041512546Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"67282718\" in 3.648756516s" Sep 9 21:18:01.044949 containerd[1525]: time="2025-09-09T21:18:01.044914269Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\"" Sep 9 21:18:01.047525 containerd[1525]: time="2025-09-09T21:18:01.047495480Z" level=info msg="CreateContainer within sandbox \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 9 21:18:01.057130 containerd[1525]: time="2025-09-09T21:18:01.057020096Z" level=info msg="Container 5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:01.067443 containerd[1525]: time="2025-09-09T21:18:01.067407222Z" level=info msg="CreateContainer within sandbox \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\"" Sep 9 21:18:01.068251 containerd[1525]: time="2025-09-09T21:18:01.068206013Z" level=info msg="StartContainer for \"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\"" Sep 9 21:18:01.084435 containerd[1525]: time="2025-09-09T21:18:01.084404075Z" level=info msg="connecting to shim 5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf" address="unix:///run/containerd/s/bad8c6e66a1a5fae73144afbdd5ada42469da364b42e32847d99b13abfb0cbb4" protocol=ttrpc version=3 Sep 9 21:18:01.113258 systemd[1]: Started cri-containerd-5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf.scope - libcontainer container 5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf. Sep 9 21:18:01.279736 containerd[1525]: time="2025-09-09T21:18:01.279547570Z" level=info msg="StartContainer for \"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\" returns successfully" Sep 9 21:18:01.747407 systemd[1]: cri-containerd-5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf.scope: Deactivated successfully. Sep 9 21:18:01.747736 systemd[1]: cri-containerd-5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf.scope: Consumed 479ms CPU time, 176.8M memory peak, 2.8M read from disk, 165.8M written to disk. Sep 9 21:18:01.748748 containerd[1525]: time="2025-09-09T21:18:01.748452818Z" level=info msg="received exit event container_id:\"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\" id:\"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\" pid:3425 exited_at:{seconds:1757452681 nanos:748200621}" Sep 9 21:18:01.748748 containerd[1525]: time="2025-09-09T21:18:01.748489178Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\" id:\"5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf\" pid:3425 exited_at:{seconds:1757452681 nanos:748200621}" Sep 9 21:18:01.771136 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5027a4b74b801960e3b4421ccf9adf0b462418d66711f1a4e701f92405da16bf-rootfs.mount: Deactivated successfully. Sep 9 21:18:01.782310 kubelet[2660]: I0909 21:18:01.782278 2660 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 9 21:18:01.856306 systemd[1]: Created slice kubepods-burstable-podac77a038_6bc7_49b6_acb6_74bcc3a30baa.slice - libcontainer container kubepods-burstable-podac77a038_6bc7_49b6_acb6_74bcc3a30baa.slice. Sep 9 21:18:01.864268 systemd[1]: Created slice kubepods-besteffort-pod7b69b002_3795_4b6e_b64b_23a21c9d9830.slice - libcontainer container kubepods-besteffort-pod7b69b002_3795_4b6e_b64b_23a21c9d9830.slice. Sep 9 21:18:01.871507 systemd[1]: Created slice kubepods-besteffort-pod1d112637_357c_40e9_9e80_6fa0222fdf7a.slice - libcontainer container kubepods-besteffort-pod1d112637_357c_40e9_9e80_6fa0222fdf7a.slice. Sep 9 21:18:01.882047 systemd[1]: Created slice kubepods-besteffort-pod319137ec_563e_4c4c_ba44_c5fde18a3fdf.slice - libcontainer container kubepods-besteffort-pod319137ec_563e_4c4c_ba44_c5fde18a3fdf.slice. Sep 9 21:18:01.887536 systemd[1]: Created slice kubepods-besteffort-poda068b1b8_ddce_47cc_a4be_745982cc66ec.slice - libcontainer container kubepods-besteffort-poda068b1b8_ddce_47cc_a4be_745982cc66ec.slice. Sep 9 21:18:01.893108 systemd[1]: Created slice kubepods-burstable-podcaa9dc8e_593a_4c15_8b92_f1c9b4823e5d.slice - libcontainer container kubepods-burstable-podcaa9dc8e_593a_4c15_8b92_f1c9b4823e5d.slice. Sep 9 21:18:01.898686 systemd[1]: Created slice kubepods-besteffort-pod3b79aa0a_edd4_45ee_8637_43c015678f95.slice - libcontainer container kubepods-besteffort-pod3b79aa0a_edd4_45ee_8637_43c015678f95.slice. Sep 9 21:18:01.913583 kubelet[2660]: I0909 21:18:01.913553 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5f4d\" (UniqueName: \"kubernetes.io/projected/7b69b002-3795-4b6e-b64b-23a21c9d9830-kube-api-access-h5f4d\") pod \"calico-apiserver-f65bff7c7-9kk94\" (UID: \"7b69b002-3795-4b6e-b64b-23a21c9d9830\") " pod="calico-apiserver/calico-apiserver-f65bff7c7-9kk94" Sep 9 21:18:01.913675 kubelet[2660]: I0909 21:18:01.913590 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5dvq\" (UniqueName: \"kubernetes.io/projected/3b79aa0a-edd4-45ee-8637-43c015678f95-kube-api-access-f5dvq\") pod \"calico-apiserver-f65bff7c7-vqfpp\" (UID: \"3b79aa0a-edd4-45ee-8637-43c015678f95\") " pod="calico-apiserver/calico-apiserver-f65bff7c7-vqfpp" Sep 9 21:18:01.913675 kubelet[2660]: I0909 21:18:01.913610 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3b79aa0a-edd4-45ee-8637-43c015678f95-calico-apiserver-certs\") pod \"calico-apiserver-f65bff7c7-vqfpp\" (UID: \"3b79aa0a-edd4-45ee-8637-43c015678f95\") " pod="calico-apiserver/calico-apiserver-f65bff7c7-vqfpp" Sep 9 21:18:01.913675 kubelet[2660]: I0909 21:18:01.913638 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac77a038-6bc7-49b6-acb6-74bcc3a30baa-config-volume\") pod \"coredns-668d6bf9bc-7sqj7\" (UID: \"ac77a038-6bc7-49b6-acb6-74bcc3a30baa\") " pod="kube-system/coredns-668d6bf9bc-7sqj7" Sep 9 21:18:01.913675 kubelet[2660]: I0909 21:18:01.913656 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j86r\" (UniqueName: \"kubernetes.io/projected/caa9dc8e-593a-4c15-8b92-f1c9b4823e5d-kube-api-access-7j86r\") pod \"coredns-668d6bf9bc-gr8qj\" (UID: \"caa9dc8e-593a-4c15-8b92-f1c9b4823e5d\") " pod="kube-system/coredns-668d6bf9bc-gr8qj" Sep 9 21:18:01.913778 kubelet[2660]: I0909 21:18:01.913702 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/7b69b002-3795-4b6e-b64b-23a21c9d9830-calico-apiserver-certs\") pod \"calico-apiserver-f65bff7c7-9kk94\" (UID: \"7b69b002-3795-4b6e-b64b-23a21c9d9830\") " pod="calico-apiserver/calico-apiserver-f65bff7c7-9kk94" Sep 9 21:18:01.913778 kubelet[2660]: I0909 21:18:01.913733 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/caa9dc8e-593a-4c15-8b92-f1c9b4823e5d-config-volume\") pod \"coredns-668d6bf9bc-gr8qj\" (UID: \"caa9dc8e-593a-4c15-8b92-f1c9b4823e5d\") " pod="kube-system/coredns-668d6bf9bc-gr8qj" Sep 9 21:18:01.913778 kubelet[2660]: I0909 21:18:01.913750 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-backend-key-pair\") pod \"whisker-7fd69799ff-qdmmt\" (UID: \"a068b1b8-ddce-47cc-a4be-745982cc66ec\") " pod="calico-system/whisker-7fd69799ff-qdmmt" Sep 9 21:18:01.913778 kubelet[2660]: I0909 21:18:01.913767 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/319137ec-563e-4c4c-ba44-c5fde18a3fdf-goldmane-key-pair\") pod \"goldmane-54d579b49d-95rlw\" (UID: \"319137ec-563e-4c4c-ba44-c5fde18a3fdf\") " pod="calico-system/goldmane-54d579b49d-95rlw" Sep 9 21:18:01.913923 kubelet[2660]: I0909 21:18:01.913802 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/319137ec-563e-4c4c-ba44-c5fde18a3fdf-config\") pod \"goldmane-54d579b49d-95rlw\" (UID: \"319137ec-563e-4c4c-ba44-c5fde18a3fdf\") " pod="calico-system/goldmane-54d579b49d-95rlw" Sep 9 21:18:01.913923 kubelet[2660]: I0909 21:18:01.913824 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-ca-bundle\") pod \"whisker-7fd69799ff-qdmmt\" (UID: \"a068b1b8-ddce-47cc-a4be-745982cc66ec\") " pod="calico-system/whisker-7fd69799ff-qdmmt" Sep 9 21:18:01.913923 kubelet[2660]: I0909 21:18:01.913842 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/319137ec-563e-4c4c-ba44-c5fde18a3fdf-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-95rlw\" (UID: \"319137ec-563e-4c4c-ba44-c5fde18a3fdf\") " pod="calico-system/goldmane-54d579b49d-95rlw" Sep 9 21:18:01.913923 kubelet[2660]: I0909 21:18:01.913859 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj94q\" (UniqueName: \"kubernetes.io/projected/ac77a038-6bc7-49b6-acb6-74bcc3a30baa-kube-api-access-nj94q\") pod \"coredns-668d6bf9bc-7sqj7\" (UID: \"ac77a038-6bc7-49b6-acb6-74bcc3a30baa\") " pod="kube-system/coredns-668d6bf9bc-7sqj7" Sep 9 21:18:01.913923 kubelet[2660]: I0909 21:18:01.913900 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmpgf\" (UniqueName: \"kubernetes.io/projected/1d112637-357c-40e9-9e80-6fa0222fdf7a-kube-api-access-tmpgf\") pod \"calico-kube-controllers-6f9758d6db-9jkm7\" (UID: \"1d112637-357c-40e9-9e80-6fa0222fdf7a\") " pod="calico-system/calico-kube-controllers-6f9758d6db-9jkm7" Sep 9 21:18:01.914040 kubelet[2660]: I0909 21:18:01.913946 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvw4l\" (UniqueName: \"kubernetes.io/projected/319137ec-563e-4c4c-ba44-c5fde18a3fdf-kube-api-access-kvw4l\") pod \"goldmane-54d579b49d-95rlw\" (UID: \"319137ec-563e-4c4c-ba44-c5fde18a3fdf\") " pod="calico-system/goldmane-54d579b49d-95rlw" Sep 9 21:18:01.914040 kubelet[2660]: I0909 21:18:01.913966 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d112637-357c-40e9-9e80-6fa0222fdf7a-tigera-ca-bundle\") pod \"calico-kube-controllers-6f9758d6db-9jkm7\" (UID: \"1d112637-357c-40e9-9e80-6fa0222fdf7a\") " pod="calico-system/calico-kube-controllers-6f9758d6db-9jkm7" Sep 9 21:18:01.914040 kubelet[2660]: I0909 21:18:01.913981 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52hxc\" (UniqueName: \"kubernetes.io/projected/a068b1b8-ddce-47cc-a4be-745982cc66ec-kube-api-access-52hxc\") pod \"whisker-7fd69799ff-qdmmt\" (UID: \"a068b1b8-ddce-47cc-a4be-745982cc66ec\") " pod="calico-system/whisker-7fd69799ff-qdmmt" Sep 9 21:18:02.160176 kubelet[2660]: E0909 21:18:02.160143 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:02.161104 containerd[1525]: time="2025-09-09T21:18:02.160865917Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sqj7,Uid:ac77a038-6bc7-49b6-acb6-74bcc3a30baa,Namespace:kube-system,Attempt:0,}" Sep 9 21:18:02.168809 containerd[1525]: time="2025-09-09T21:18:02.168750114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-9kk94,Uid:7b69b002-3795-4b6e-b64b-23a21c9d9830,Namespace:calico-apiserver,Attempt:0,}" Sep 9 21:18:02.175192 containerd[1525]: time="2025-09-09T21:18:02.175092607Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6f9758d6db-9jkm7,Uid:1d112637-357c-40e9-9e80-6fa0222fdf7a,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:02.185958 containerd[1525]: time="2025-09-09T21:18:02.185877293Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-95rlw,Uid:319137ec-563e-4c4c-ba44-c5fde18a3fdf,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:02.193094 containerd[1525]: time="2025-09-09T21:18:02.192940539Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7fd69799ff-qdmmt,Uid:a068b1b8-ddce-47cc-a4be-745982cc66ec,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:02.195267 kubelet[2660]: E0909 21:18:02.195236 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:02.196078 containerd[1525]: time="2025-09-09T21:18:02.196051466Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gr8qj,Uid:caa9dc8e-593a-4c15-8b92-f1c9b4823e5d,Namespace:kube-system,Attempt:0,}" Sep 9 21:18:02.204188 containerd[1525]: time="2025-09-09T21:18:02.204141341Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-vqfpp,Uid:3b79aa0a-edd4-45ee-8637-43c015678f95,Namespace:calico-apiserver,Attempt:0,}" Sep 9 21:18:02.280633 containerd[1525]: time="2025-09-09T21:18:02.280548495Z" level=error msg="Failed to destroy network for sandbox \"44e4b843856736943580e1af754f3c9e2aa10bdf569eace723f6740736d1be62\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.281802 containerd[1525]: time="2025-09-09T21:18:02.281742882Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sqj7,Uid:ac77a038-6bc7-49b6-acb6-74bcc3a30baa,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"44e4b843856736943580e1af754f3c9e2aa10bdf569eace723f6740736d1be62\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.285858 kubelet[2660]: E0909 21:18:02.285781 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"44e4b843856736943580e1af754f3c9e2aa10bdf569eace723f6740736d1be62\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.288765 kubelet[2660]: E0909 21:18:02.288728 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"44e4b843856736943580e1af754f3c9e2aa10bdf569eace723f6740736d1be62\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-7sqj7" Sep 9 21:18:02.288905 kubelet[2660]: E0909 21:18:02.288887 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"44e4b843856736943580e1af754f3c9e2aa10bdf569eace723f6740736d1be62\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-7sqj7" Sep 9 21:18:02.289032 kubelet[2660]: E0909 21:18:02.288996 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-7sqj7_kube-system(ac77a038-6bc7-49b6-acb6-74bcc3a30baa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-7sqj7_kube-system(ac77a038-6bc7-49b6-acb6-74bcc3a30baa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"44e4b843856736943580e1af754f3c9e2aa10bdf569eace723f6740736d1be62\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-7sqj7" podUID="ac77a038-6bc7-49b6-acb6-74bcc3a30baa" Sep 9 21:18:02.293945 containerd[1525]: time="2025-09-09T21:18:02.293912394Z" level=error msg="Failed to destroy network for sandbox \"04c9890d25d4cf53b439522c92c69d22c8fd37562d33954ce050751ddf33225a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.295252 containerd[1525]: time="2025-09-09T21:18:02.295212300Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gr8qj,Uid:caa9dc8e-593a-4c15-8b92-f1c9b4823e5d,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"04c9890d25d4cf53b439522c92c69d22c8fd37562d33954ce050751ddf33225a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.295697 kubelet[2660]: E0909 21:18:02.295661 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04c9890d25d4cf53b439522c92c69d22c8fd37562d33954ce050751ddf33225a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.295758 kubelet[2660]: E0909 21:18:02.295708 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04c9890d25d4cf53b439522c92c69d22c8fd37562d33954ce050751ddf33225a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-gr8qj" Sep 9 21:18:02.295758 kubelet[2660]: E0909 21:18:02.295729 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04c9890d25d4cf53b439522c92c69d22c8fd37562d33954ce050751ddf33225a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-gr8qj" Sep 9 21:18:02.295841 kubelet[2660]: E0909 21:18:02.295765 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-gr8qj_kube-system(caa9dc8e-593a-4c15-8b92-f1c9b4823e5d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-gr8qj_kube-system(caa9dc8e-593a-4c15-8b92-f1c9b4823e5d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"04c9890d25d4cf53b439522c92c69d22c8fd37562d33954ce050751ddf33225a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-gr8qj" podUID="caa9dc8e-593a-4c15-8b92-f1c9b4823e5d" Sep 9 21:18:02.296112 containerd[1525]: time="2025-09-09T21:18:02.295554257Z" level=error msg="Failed to destroy network for sandbox \"442ca1ac4345a73f85759e0df75f96b52e40f5b3eb4a3ebaa62b5b390d8e0250\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.297331 containerd[1525]: time="2025-09-09T21:18:02.297270159Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-vqfpp,Uid:3b79aa0a-edd4-45ee-8637-43c015678f95,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"442ca1ac4345a73f85759e0df75f96b52e40f5b3eb4a3ebaa62b5b390d8e0250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.297480 kubelet[2660]: E0909 21:18:02.297448 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"442ca1ac4345a73f85759e0df75f96b52e40f5b3eb4a3ebaa62b5b390d8e0250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.297525 kubelet[2660]: E0909 21:18:02.297486 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"442ca1ac4345a73f85759e0df75f96b52e40f5b3eb4a3ebaa62b5b390d8e0250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f65bff7c7-vqfpp" Sep 9 21:18:02.297525 kubelet[2660]: E0909 21:18:02.297503 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"442ca1ac4345a73f85759e0df75f96b52e40f5b3eb4a3ebaa62b5b390d8e0250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f65bff7c7-vqfpp" Sep 9 21:18:02.297578 kubelet[2660]: E0909 21:18:02.297532 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f65bff7c7-vqfpp_calico-apiserver(3b79aa0a-edd4-45ee-8637-43c015678f95)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f65bff7c7-vqfpp_calico-apiserver(3b79aa0a-edd4-45ee-8637-43c015678f95)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"442ca1ac4345a73f85759e0df75f96b52e40f5b3eb4a3ebaa62b5b390d8e0250\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f65bff7c7-vqfpp" podUID="3b79aa0a-edd4-45ee-8637-43c015678f95" Sep 9 21:18:02.303632 containerd[1525]: time="2025-09-09T21:18:02.303589292Z" level=error msg="Failed to destroy network for sandbox \"d894921187e6913050438f1f2ed1c8ede8ecf688d9261f8024314af256928f4b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.305104 containerd[1525]: time="2025-09-09T21:18:02.305072516Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7fd69799ff-qdmmt,Uid:a068b1b8-ddce-47cc-a4be-745982cc66ec,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d894921187e6913050438f1f2ed1c8ede8ecf688d9261f8024314af256928f4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.305812 kubelet[2660]: E0909 21:18:02.305721 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d894921187e6913050438f1f2ed1c8ede8ecf688d9261f8024314af256928f4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.305872 kubelet[2660]: E0909 21:18:02.305817 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d894921187e6913050438f1f2ed1c8ede8ecf688d9261f8024314af256928f4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7fd69799ff-qdmmt" Sep 9 21:18:02.305872 kubelet[2660]: E0909 21:18:02.305836 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d894921187e6913050438f1f2ed1c8ede8ecf688d9261f8024314af256928f4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7fd69799ff-qdmmt" Sep 9 21:18:02.305919 kubelet[2660]: E0909 21:18:02.305869 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7fd69799ff-qdmmt_calico-system(a068b1b8-ddce-47cc-a4be-745982cc66ec)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7fd69799ff-qdmmt_calico-system(a068b1b8-ddce-47cc-a4be-745982cc66ec)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d894921187e6913050438f1f2ed1c8ede8ecf688d9261f8024314af256928f4b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7fd69799ff-qdmmt" podUID="a068b1b8-ddce-47cc-a4be-745982cc66ec" Sep 9 21:18:02.306173 containerd[1525]: time="2025-09-09T21:18:02.306149545Z" level=error msg="Failed to destroy network for sandbox \"01e4e22208ba673ac6decafc2bbb70fe4dd96f6547c17c837a8751692d0bba33\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.307071 containerd[1525]: time="2025-09-09T21:18:02.307043216Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-95rlw,Uid:319137ec-563e-4c4c-ba44-c5fde18a3fdf,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"01e4e22208ba673ac6decafc2bbb70fe4dd96f6547c17c837a8751692d0bba33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.307263 kubelet[2660]: E0909 21:18:02.307238 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"01e4e22208ba673ac6decafc2bbb70fe4dd96f6547c17c837a8751692d0bba33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.307308 kubelet[2660]: E0909 21:18:02.307291 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"01e4e22208ba673ac6decafc2bbb70fe4dd96f6547c17c837a8751692d0bba33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-95rlw" Sep 9 21:18:02.307334 kubelet[2660]: E0909 21:18:02.307311 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"01e4e22208ba673ac6decafc2bbb70fe4dd96f6547c17c837a8751692d0bba33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-95rlw" Sep 9 21:18:02.307361 kubelet[2660]: E0909 21:18:02.307344 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-95rlw_calico-system(319137ec-563e-4c4c-ba44-c5fde18a3fdf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-95rlw_calico-system(319137ec-563e-4c4c-ba44-c5fde18a3fdf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"01e4e22208ba673ac6decafc2bbb70fe4dd96f6547c17c837a8751692d0bba33\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-95rlw" podUID="319137ec-563e-4c4c-ba44-c5fde18a3fdf" Sep 9 21:18:02.313043 containerd[1525]: time="2025-09-09T21:18:02.313007113Z" level=error msg="Failed to destroy network for sandbox \"a02e3565fe61eb712dd387dfacdaf9d7577b7db9d757156a71e1438dc9804462\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.314442 containerd[1525]: time="2025-09-09T21:18:02.314151501Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-9kk94,Uid:7b69b002-3795-4b6e-b64b-23a21c9d9830,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a02e3565fe61eb712dd387dfacdaf9d7577b7db9d757156a71e1438dc9804462\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.314529 kubelet[2660]: E0909 21:18:02.314329 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a02e3565fe61eb712dd387dfacdaf9d7577b7db9d757156a71e1438dc9804462\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.314529 kubelet[2660]: E0909 21:18:02.314364 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a02e3565fe61eb712dd387dfacdaf9d7577b7db9d757156a71e1438dc9804462\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f65bff7c7-9kk94" Sep 9 21:18:02.314529 kubelet[2660]: E0909 21:18:02.314388 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a02e3565fe61eb712dd387dfacdaf9d7577b7db9d757156a71e1438dc9804462\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f65bff7c7-9kk94" Sep 9 21:18:02.314617 kubelet[2660]: E0909 21:18:02.314426 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f65bff7c7-9kk94_calico-apiserver(7b69b002-3795-4b6e-b64b-23a21c9d9830)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f65bff7c7-9kk94_calico-apiserver(7b69b002-3795-4b6e-b64b-23a21c9d9830)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a02e3565fe61eb712dd387dfacdaf9d7577b7db9d757156a71e1438dc9804462\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f65bff7c7-9kk94" podUID="7b69b002-3795-4b6e-b64b-23a21c9d9830" Sep 9 21:18:02.325633 containerd[1525]: time="2025-09-09T21:18:02.325594420Z" level=error msg="Failed to destroy network for sandbox \"40491d8154992fd5c9ebe48a5c25d1a9cce70e30f67f6e1b033f1ffc69c8f3b5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.327111 containerd[1525]: time="2025-09-09T21:18:02.327078444Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6f9758d6db-9jkm7,Uid:1d112637-357c-40e9-9e80-6fa0222fdf7a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"40491d8154992fd5c9ebe48a5c25d1a9cce70e30f67f6e1b033f1ffc69c8f3b5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.327313 kubelet[2660]: E0909 21:18:02.327280 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"40491d8154992fd5c9ebe48a5c25d1a9cce70e30f67f6e1b033f1ffc69c8f3b5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.327367 kubelet[2660]: E0909 21:18:02.327325 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"40491d8154992fd5c9ebe48a5c25d1a9cce70e30f67f6e1b033f1ffc69c8f3b5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6f9758d6db-9jkm7" Sep 9 21:18:02.327367 kubelet[2660]: E0909 21:18:02.327342 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"40491d8154992fd5c9ebe48a5c25d1a9cce70e30f67f6e1b033f1ffc69c8f3b5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6f9758d6db-9jkm7" Sep 9 21:18:02.327414 kubelet[2660]: E0909 21:18:02.327371 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6f9758d6db-9jkm7_calico-system(1d112637-357c-40e9-9e80-6fa0222fdf7a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6f9758d6db-9jkm7_calico-system(1d112637-357c-40e9-9e80-6fa0222fdf7a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"40491d8154992fd5c9ebe48a5c25d1a9cce70e30f67f6e1b033f1ffc69c8f3b5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6f9758d6db-9jkm7" podUID="1d112637-357c-40e9-9e80-6fa0222fdf7a" Sep 9 21:18:02.330627 systemd[1]: Created slice kubepods-besteffort-podec7a9498_a0d9_4dbc_a0ab_af5696ebcf03.slice - libcontainer container kubepods-besteffort-podec7a9498_a0d9_4dbc_a0ab_af5696ebcf03.slice. Sep 9 21:18:02.332602 containerd[1525]: time="2025-09-09T21:18:02.332574626Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-q45ms,Uid:ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:02.371891 containerd[1525]: time="2025-09-09T21:18:02.371855292Z" level=error msg="Failed to destroy network for sandbox \"81e43d859acaee4220309059d1b4504164e46f14d89656568af140177ad4cb43\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.372970 containerd[1525]: time="2025-09-09T21:18:02.372939361Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-q45ms,Uid:ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"81e43d859acaee4220309059d1b4504164e46f14d89656568af140177ad4cb43\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.373146 kubelet[2660]: E0909 21:18:02.373112 2660 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81e43d859acaee4220309059d1b4504164e46f14d89656568af140177ad4cb43\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 9 21:18:02.373185 kubelet[2660]: E0909 21:18:02.373166 2660 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81e43d859acaee4220309059d1b4504164e46f14d89656568af140177ad4cb43\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-q45ms" Sep 9 21:18:02.373233 kubelet[2660]: E0909 21:18:02.373191 2660 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81e43d859acaee4220309059d1b4504164e46f14d89656568af140177ad4cb43\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-q45ms" Sep 9 21:18:02.373261 kubelet[2660]: E0909 21:18:02.373232 2660 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-q45ms_calico-system(ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-q45ms_calico-system(ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"81e43d859acaee4220309059d1b4504164e46f14d89656568af140177ad4cb43\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-q45ms" podUID="ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03" Sep 9 21:18:02.409678 containerd[1525]: time="2025-09-09T21:18:02.409639694Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 9 21:18:03.056470 systemd[1]: run-netns-cni\x2d4af9de38\x2d7b35\x2d43f4\x2d6a7d\x2dc5262f1b96ff.mount: Deactivated successfully. Sep 9 21:18:03.056559 systemd[1]: run-netns-cni\x2ddd5d3ef8\x2d87c9\x2dc02c\x2d62aa\x2dd0a5a6a6c645.mount: Deactivated successfully. Sep 9 21:18:03.056610 systemd[1]: run-netns-cni\x2df8e8e8e4\x2dffcc\x2d51d1\x2d84f1\x2d604f83329be8.mount: Deactivated successfully. Sep 9 21:18:06.166297 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1662031304.mount: Deactivated successfully. Sep 9 21:18:06.469979 containerd[1525]: time="2025-09-09T21:18:06.469687613Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=151100457" Sep 9 21:18:06.473317 containerd[1525]: time="2025-09-09T21:18:06.473275141Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:06.474331 containerd[1525]: time="2025-09-09T21:18:06.474295211Z" level=info msg="ImageCreate event name:\"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:06.475113 containerd[1525]: time="2025-09-09T21:18:06.474867486Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"151100319\" in 4.065191993s" Sep 9 21:18:06.475113 containerd[1525]: time="2025-09-09T21:18:06.474899646Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\"" Sep 9 21:18:06.476195 containerd[1525]: time="2025-09-09T21:18:06.476149995Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:06.488026 containerd[1525]: time="2025-09-09T21:18:06.487990608Z" level=info msg="CreateContainer within sandbox \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 9 21:18:06.506821 containerd[1525]: time="2025-09-09T21:18:06.506759398Z" level=info msg="Container 69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:06.515246 containerd[1525]: time="2025-09-09T21:18:06.515123323Z" level=info msg="CreateContainer within sandbox \"ac93293ea103f0e0ea310d87fbcfc6333080ac9c309d65f47ae03fdd8d23ddad\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a\"" Sep 9 21:18:06.516911 containerd[1525]: time="2025-09-09T21:18:06.516803387Z" level=info msg="StartContainer for \"69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a\"" Sep 9 21:18:06.518956 containerd[1525]: time="2025-09-09T21:18:06.518918328Z" level=info msg="connecting to shim 69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a" address="unix:///run/containerd/s/bad8c6e66a1a5fae73144afbdd5ada42469da364b42e32847d99b13abfb0cbb4" protocol=ttrpc version=3 Sep 9 21:18:06.536963 systemd[1]: Started cri-containerd-69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a.scope - libcontainer container 69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a. Sep 9 21:18:06.574924 containerd[1525]: time="2025-09-09T21:18:06.574881583Z" level=info msg="StartContainer for \"69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a\" returns successfully" Sep 9 21:18:06.689814 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 9 21:18:06.689896 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 9 21:18:06.848401 kubelet[2660]: I0909 21:18:06.848352 2660 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-ca-bundle\") pod \"a068b1b8-ddce-47cc-a4be-745982cc66ec\" (UID: \"a068b1b8-ddce-47cc-a4be-745982cc66ec\") " Sep 9 21:18:06.848772 kubelet[2660]: I0909 21:18:06.848416 2660 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52hxc\" (UniqueName: \"kubernetes.io/projected/a068b1b8-ddce-47cc-a4be-745982cc66ec-kube-api-access-52hxc\") pod \"a068b1b8-ddce-47cc-a4be-745982cc66ec\" (UID: \"a068b1b8-ddce-47cc-a4be-745982cc66ec\") " Sep 9 21:18:06.848772 kubelet[2660]: I0909 21:18:06.848444 2660 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-backend-key-pair\") pod \"a068b1b8-ddce-47cc-a4be-745982cc66ec\" (UID: \"a068b1b8-ddce-47cc-a4be-745982cc66ec\") " Sep 9 21:18:06.848772 kubelet[2660]: I0909 21:18:06.848753 2660 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "a068b1b8-ddce-47cc-a4be-745982cc66ec" (UID: "a068b1b8-ddce-47cc-a4be-745982cc66ec"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 9 21:18:06.859280 kubelet[2660]: I0909 21:18:06.859088 2660 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "a068b1b8-ddce-47cc-a4be-745982cc66ec" (UID: "a068b1b8-ddce-47cc-a4be-745982cc66ec"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 9 21:18:06.859280 kubelet[2660]: I0909 21:18:06.859111 2660 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a068b1b8-ddce-47cc-a4be-745982cc66ec-kube-api-access-52hxc" (OuterVolumeSpecName: "kube-api-access-52hxc") pod "a068b1b8-ddce-47cc-a4be-745982cc66ec" (UID: "a068b1b8-ddce-47cc-a4be-745982cc66ec"). InnerVolumeSpecName "kube-api-access-52hxc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 9 21:18:06.949247 kubelet[2660]: I0909 21:18:06.949203 2660 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Sep 9 21:18:06.949247 kubelet[2660]: I0909 21:18:06.949237 2660 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a068b1b8-ddce-47cc-a4be-745982cc66ec-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Sep 9 21:18:06.949247 kubelet[2660]: I0909 21:18:06.949249 2660 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-52hxc\" (UniqueName: \"kubernetes.io/projected/a068b1b8-ddce-47cc-a4be-745982cc66ec-kube-api-access-52hxc\") on node \"localhost\" DevicePath \"\"" Sep 9 21:18:07.167255 systemd[1]: var-lib-kubelet-pods-a068b1b8\x2dddce\x2d47cc\x2da4be\x2d745982cc66ec-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d52hxc.mount: Deactivated successfully. Sep 9 21:18:07.167344 systemd[1]: var-lib-kubelet-pods-a068b1b8\x2dddce\x2d47cc\x2da4be\x2d745982cc66ec-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 9 21:18:07.333922 systemd[1]: Removed slice kubepods-besteffort-poda068b1b8_ddce_47cc_a4be_745982cc66ec.slice - libcontainer container kubepods-besteffort-poda068b1b8_ddce_47cc_a4be_745982cc66ec.slice. Sep 9 21:18:07.451853 kubelet[2660]: I0909 21:18:07.451293 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-tp66b" podStartSLOduration=1.426108082 podStartE2EDuration="13.451276929s" podCreationTimestamp="2025-09-09 21:17:54 +0000 UTC" firstStartedPulling="2025-09-09 21:17:54.453746723 +0000 UTC m=+19.206302642" lastFinishedPulling="2025-09-09 21:18:06.47891557 +0000 UTC m=+31.231471489" observedRunningTime="2025-09-09 21:18:07.442318247 +0000 UTC m=+32.194874206" watchObservedRunningTime="2025-09-09 21:18:07.451276929 +0000 UTC m=+32.203832808" Sep 9 21:18:07.489421 systemd[1]: Created slice kubepods-besteffort-podf0a169da_7843_48ae_bdc2_e15c4bfba01b.slice - libcontainer container kubepods-besteffort-podf0a169da_7843_48ae_bdc2_e15c4bfba01b.slice. Sep 9 21:18:07.553301 kubelet[2660]: I0909 21:18:07.553219 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f0a169da-7843-48ae-bdc2-e15c4bfba01b-whisker-backend-key-pair\") pod \"whisker-68fd97857d-nqls8\" (UID: \"f0a169da-7843-48ae-bdc2-e15c4bfba01b\") " pod="calico-system/whisker-68fd97857d-nqls8" Sep 9 21:18:07.553301 kubelet[2660]: I0909 21:18:07.553306 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0a169da-7843-48ae-bdc2-e15c4bfba01b-whisker-ca-bundle\") pod \"whisker-68fd97857d-nqls8\" (UID: \"f0a169da-7843-48ae-bdc2-e15c4bfba01b\") " pod="calico-system/whisker-68fd97857d-nqls8" Sep 9 21:18:07.553448 kubelet[2660]: I0909 21:18:07.553328 2660 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mplk\" (UniqueName: \"kubernetes.io/projected/f0a169da-7843-48ae-bdc2-e15c4bfba01b-kube-api-access-8mplk\") pod \"whisker-68fd97857d-nqls8\" (UID: \"f0a169da-7843-48ae-bdc2-e15c4bfba01b\") " pod="calico-system/whisker-68fd97857d-nqls8" Sep 9 21:18:07.578547 containerd[1525]: time="2025-09-09T21:18:07.578506140Z" level=info msg="TaskExit event in podsandbox handler container_id:\"69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a\" id:\"2d4529f33a1f116dbb66529dc4ddf9cdd0a8095839c2050227015a9479758e45\" pid:3808 exit_status:1 exited_at:{seconds:1757452687 nanos:578221583}" Sep 9 21:18:07.797347 containerd[1525]: time="2025-09-09T21:18:07.797304153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-68fd97857d-nqls8,Uid:f0a169da-7843-48ae-bdc2-e15c4bfba01b,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:07.939821 systemd-networkd[1433]: cali05e94fd6f8c: Link UP Sep 9 21:18:07.940329 systemd-networkd[1433]: cali05e94fd6f8c: Gained carrier Sep 9 21:18:07.954172 containerd[1525]: 2025-09-09 21:18:07.817 [INFO][3824] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:07.954172 containerd[1525]: 2025-09-09 21:18:07.844 [INFO][3824] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--68fd97857d--nqls8-eth0 whisker-68fd97857d- calico-system f0a169da-7843-48ae-bdc2-e15c4bfba01b 880 0 2025-09-09 21:18:07 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:68fd97857d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-68fd97857d-nqls8 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali05e94fd6f8c [] [] }} ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-" Sep 9 21:18:07.954172 containerd[1525]: 2025-09-09 21:18:07.844 [INFO][3824] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.954172 containerd[1525]: 2025-09-09 21:18:07.900 [INFO][3838] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" HandleID="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Workload="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.900 [INFO][3838] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" HandleID="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Workload="localhost-k8s-whisker--68fd97857d--nqls8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000119b20), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-68fd97857d-nqls8", "timestamp":"2025-09-09 21:18:07.900263096 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.900 [INFO][3838] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.900 [INFO][3838] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.900 [INFO][3838] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.911 [INFO][3838] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" host="localhost" Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.915 [INFO][3838] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.919 [INFO][3838] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.920 [INFO][3838] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.922 [INFO][3838] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:07.954383 containerd[1525]: 2025-09-09 21:18:07.922 [INFO][3838] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" host="localhost" Sep 9 21:18:07.954567 containerd[1525]: 2025-09-09 21:18:07.923 [INFO][3838] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b Sep 9 21:18:07.954567 containerd[1525]: 2025-09-09 21:18:07.927 [INFO][3838] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" host="localhost" Sep 9 21:18:07.954567 containerd[1525]: 2025-09-09 21:18:07.931 [INFO][3838] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" host="localhost" Sep 9 21:18:07.954567 containerd[1525]: 2025-09-09 21:18:07.931 [INFO][3838] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" host="localhost" Sep 9 21:18:07.954567 containerd[1525]: 2025-09-09 21:18:07.931 [INFO][3838] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:07.954567 containerd[1525]: 2025-09-09 21:18:07.931 [INFO][3838] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" HandleID="k8s-pod-network.5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Workload="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.954774 containerd[1525]: 2025-09-09 21:18:07.934 [INFO][3824] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--68fd97857d--nqls8-eth0", GenerateName:"whisker-68fd97857d-", Namespace:"calico-system", SelfLink:"", UID:"f0a169da-7843-48ae-bdc2-e15c4bfba01b", ResourceVersion:"880", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 18, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"68fd97857d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-68fd97857d-nqls8", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali05e94fd6f8c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:07.954774 containerd[1525]: 2025-09-09 21:18:07.934 [INFO][3824] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.954900 containerd[1525]: 2025-09-09 21:18:07.934 [INFO][3824] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali05e94fd6f8c ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.954900 containerd[1525]: 2025-09-09 21:18:07.940 [INFO][3824] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.954943 containerd[1525]: 2025-09-09 21:18:07.940 [INFO][3824] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--68fd97857d--nqls8-eth0", GenerateName:"whisker-68fd97857d-", Namespace:"calico-system", SelfLink:"", UID:"f0a169da-7843-48ae-bdc2-e15c4bfba01b", ResourceVersion:"880", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 18, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"68fd97857d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b", Pod:"whisker-68fd97857d-nqls8", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali05e94fd6f8c", MAC:"8e:1d:6b:cc:0c:95", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:07.955006 containerd[1525]: 2025-09-09 21:18:07.951 [INFO][3824] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" Namespace="calico-system" Pod="whisker-68fd97857d-nqls8" WorkloadEndpoint="localhost-k8s-whisker--68fd97857d--nqls8-eth0" Sep 9 21:18:07.987741 containerd[1525]: time="2025-09-09T21:18:07.987689774Z" level=info msg="connecting to shim 5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b" address="unix:///run/containerd/s/e04d83fb7c3d7eeb13a4a2270ebf8dd161c378e42cadb35d66dbb005cb18b03d" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:08.053975 systemd[1]: Started cri-containerd-5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b.scope - libcontainer container 5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b. Sep 9 21:18:08.083412 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:08.149746 containerd[1525]: time="2025-09-09T21:18:08.149700087Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-68fd97857d-nqls8,Uid:f0a169da-7843-48ae-bdc2-e15c4bfba01b,Namespace:calico-system,Attempt:0,} returns sandbox id \"5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b\"" Sep 9 21:18:08.152167 containerd[1525]: time="2025-09-09T21:18:08.152132706Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 9 21:18:08.489466 containerd[1525]: time="2025-09-09T21:18:08.489332269Z" level=info msg="TaskExit event in podsandbox handler container_id:\"69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a\" id:\"d46a769d2e63c4cd81c8ff2a10c0ec905ca0d16007864da57ceb72b7b66dc6dd\" pid:4012 exit_status:1 exited_at:{seconds:1757452688 nanos:489005272}" Sep 9 21:18:09.115973 systemd-networkd[1433]: cali05e94fd6f8c: Gained IPv6LL Sep 9 21:18:09.328403 kubelet[2660]: I0909 21:18:09.328344 2660 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a068b1b8-ddce-47cc-a4be-745982cc66ec" path="/var/lib/kubelet/pods/a068b1b8-ddce-47cc-a4be-745982cc66ec/volumes" Sep 9 21:18:09.639826 containerd[1525]: time="2025-09-09T21:18:09.639752970Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:09.640471 containerd[1525]: time="2025-09-09T21:18:09.640446044Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4605606" Sep 9 21:18:09.641127 containerd[1525]: time="2025-09-09T21:18:09.641098599Z" level=info msg="ImageCreate event name:\"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:09.643403 containerd[1525]: time="2025-09-09T21:18:09.643379500Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:09.644043 containerd[1525]: time="2025-09-09T21:18:09.643921096Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"5974839\" in 1.49174659s" Sep 9 21:18:09.644043 containerd[1525]: time="2025-09-09T21:18:09.643955576Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\"" Sep 9 21:18:09.645815 containerd[1525]: time="2025-09-09T21:18:09.645695841Z" level=info msg="CreateContainer within sandbox \"5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 9 21:18:09.654918 containerd[1525]: time="2025-09-09T21:18:09.653852455Z" level=info msg="Container 73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:09.656010 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount321162440.mount: Deactivated successfully. Sep 9 21:18:09.660637 containerd[1525]: time="2025-09-09T21:18:09.660594360Z" level=info msg="CreateContainer within sandbox \"5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3\"" Sep 9 21:18:09.660980 containerd[1525]: time="2025-09-09T21:18:09.660961677Z" level=info msg="StartContainer for \"73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3\"" Sep 9 21:18:09.663201 containerd[1525]: time="2025-09-09T21:18:09.663166179Z" level=info msg="connecting to shim 73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3" address="unix:///run/containerd/s/e04d83fb7c3d7eeb13a4a2270ebf8dd161c378e42cadb35d66dbb005cb18b03d" protocol=ttrpc version=3 Sep 9 21:18:09.687955 systemd[1]: Started cri-containerd-73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3.scope - libcontainer container 73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3. Sep 9 21:18:09.717251 containerd[1525]: time="2025-09-09T21:18:09.717220740Z" level=info msg="StartContainer for \"73a5667054956d85e4acedddd0b2ed2fe4a59f53a1f577d00e50a8587e7ae5f3\" returns successfully" Sep 9 21:18:09.719533 containerd[1525]: time="2025-09-09T21:18:09.719502121Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 9 21:18:11.230068 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2919605620.mount: Deactivated successfully. Sep 9 21:18:11.247622 containerd[1525]: time="2025-09-09T21:18:11.247580805Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:11.248460 containerd[1525]: time="2025-09-09T21:18:11.248305400Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=30823700" Sep 9 21:18:11.249282 containerd[1525]: time="2025-09-09T21:18:11.249249552Z" level=info msg="ImageCreate event name:\"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:11.251273 containerd[1525]: time="2025-09-09T21:18:11.251232457Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:11.251992 containerd[1525]: time="2025-09-09T21:18:11.251943012Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"30823530\" in 1.532296012s" Sep 9 21:18:11.252053 containerd[1525]: time="2025-09-09T21:18:11.251997211Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\"" Sep 9 21:18:11.254580 containerd[1525]: time="2025-09-09T21:18:11.254529152Z" level=info msg="CreateContainer within sandbox \"5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 9 21:18:11.261664 containerd[1525]: time="2025-09-09T21:18:11.260945303Z" level=info msg="Container f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:11.269089 containerd[1525]: time="2025-09-09T21:18:11.269056401Z" level=info msg="CreateContainer within sandbox \"5bdc74932af5daf5790ccb4a6bfa389fa47621ad38b49df435618e58be11fa6b\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053\"" Sep 9 21:18:11.271076 containerd[1525]: time="2025-09-09T21:18:11.271040826Z" level=info msg="StartContainer for \"f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053\"" Sep 9 21:18:11.272154 containerd[1525]: time="2025-09-09T21:18:11.272126298Z" level=info msg="connecting to shim f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053" address="unix:///run/containerd/s/e04d83fb7c3d7eeb13a4a2270ebf8dd161c378e42cadb35d66dbb005cb18b03d" protocol=ttrpc version=3 Sep 9 21:18:11.293940 systemd[1]: Started cri-containerd-f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053.scope - libcontainer container f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053. Sep 9 21:18:11.343920 containerd[1525]: time="2025-09-09T21:18:11.343876431Z" level=info msg="StartContainer for \"f3429628c0e0feced03b5f3120a09233bc69cbe0faac3e192f8838a4c76fb053\" returns successfully" Sep 9 21:18:11.506070 kubelet[2660]: I0909 21:18:11.505613 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-68fd97857d-nqls8" podStartSLOduration=1.404393345 podStartE2EDuration="4.505591999s" podCreationTimestamp="2025-09-09 21:18:07 +0000 UTC" firstStartedPulling="2025-09-09 21:18:08.151872749 +0000 UTC m=+32.904428628" lastFinishedPulling="2025-09-09 21:18:11.253071363 +0000 UTC m=+36.005627282" observedRunningTime="2025-09-09 21:18:11.50549208 +0000 UTC m=+36.258047999" watchObservedRunningTime="2025-09-09 21:18:11.505591999 +0000 UTC m=+36.258147918" Sep 9 21:18:13.325955 containerd[1525]: time="2025-09-09T21:18:13.325850506Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-9kk94,Uid:7b69b002-3795-4b6e-b64b-23a21c9d9830,Namespace:calico-apiserver,Attempt:0,}" Sep 9 21:18:13.429234 systemd-networkd[1433]: calieeaef2f2d6a: Link UP Sep 9 21:18:13.430033 systemd-networkd[1433]: calieeaef2f2d6a: Gained carrier Sep 9 21:18:13.446611 containerd[1525]: 2025-09-09 21:18:13.348 [INFO][4211] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:13.446611 containerd[1525]: 2025-09-09 21:18:13.363 [INFO][4211] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0 calico-apiserver-f65bff7c7- calico-apiserver 7b69b002-3795-4b6e-b64b-23a21c9d9830 808 0 2025-09-09 21:17:51 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:f65bff7c7 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-f65bff7c7-9kk94 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calieeaef2f2d6a [] [] }} ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-" Sep 9 21:18:13.446611 containerd[1525]: 2025-09-09 21:18:13.363 [INFO][4211] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.446611 containerd[1525]: 2025-09-09 21:18:13.389 [INFO][4227] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" HandleID="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Workload="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.389 [INFO][4227] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" HandleID="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Workload="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002dd5f0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-f65bff7c7-9kk94", "timestamp":"2025-09-09 21:18:13.389186252 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.389 [INFO][4227] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.389 [INFO][4227] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.389 [INFO][4227] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.398 [INFO][4227] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" host="localhost" Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.402 [INFO][4227] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.406 [INFO][4227] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.408 [INFO][4227] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.412 [INFO][4227] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:13.447040 containerd[1525]: 2025-09-09 21:18:13.412 [INFO][4227] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" host="localhost" Sep 9 21:18:13.447480 containerd[1525]: 2025-09-09 21:18:13.416 [INFO][4227] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a Sep 9 21:18:13.447480 containerd[1525]: 2025-09-09 21:18:13.420 [INFO][4227] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" host="localhost" Sep 9 21:18:13.447480 containerd[1525]: 2025-09-09 21:18:13.424 [INFO][4227] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" host="localhost" Sep 9 21:18:13.447480 containerd[1525]: 2025-09-09 21:18:13.424 [INFO][4227] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" host="localhost" Sep 9 21:18:13.447480 containerd[1525]: 2025-09-09 21:18:13.424 [INFO][4227] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:13.447480 containerd[1525]: 2025-09-09 21:18:13.424 [INFO][4227] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" HandleID="k8s-pod-network.55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Workload="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.447598 containerd[1525]: 2025-09-09 21:18:13.427 [INFO][4211] cni-plugin/k8s.go 418: Populated endpoint ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0", GenerateName:"calico-apiserver-f65bff7c7-", Namespace:"calico-apiserver", SelfLink:"", UID:"7b69b002-3795-4b6e-b64b-23a21c9d9830", ResourceVersion:"808", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f65bff7c7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-f65bff7c7-9kk94", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieeaef2f2d6a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:13.447657 containerd[1525]: 2025-09-09 21:18:13.427 [INFO][4211] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.447657 containerd[1525]: 2025-09-09 21:18:13.427 [INFO][4211] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieeaef2f2d6a ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.447657 containerd[1525]: 2025-09-09 21:18:13.431 [INFO][4211] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.447717 containerd[1525]: 2025-09-09 21:18:13.431 [INFO][4211] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0", GenerateName:"calico-apiserver-f65bff7c7-", Namespace:"calico-apiserver", SelfLink:"", UID:"7b69b002-3795-4b6e-b64b-23a21c9d9830", ResourceVersion:"808", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f65bff7c7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a", Pod:"calico-apiserver-f65bff7c7-9kk94", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieeaef2f2d6a", MAC:"8e:50:97:16:c1:86", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:13.447763 containerd[1525]: 2025-09-09 21:18:13.443 [INFO][4211] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-9kk94" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--9kk94-eth0" Sep 9 21:18:13.470506 containerd[1525]: time="2025-09-09T21:18:13.470459989Z" level=info msg="connecting to shim 55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a" address="unix:///run/containerd/s/691c251f2f0ea98a8bf25d5252008ce9963cb10aece1eb73fda73d7310c9419a" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:13.507346 systemd[1]: Started cri-containerd-55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a.scope - libcontainer container 55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a. Sep 9 21:18:13.521616 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:13.558299 containerd[1525]: time="2025-09-09T21:18:13.558120481Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-9kk94,Uid:7b69b002-3795-4b6e-b64b-23a21c9d9830,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a\"" Sep 9 21:18:13.560860 containerd[1525]: time="2025-09-09T21:18:13.560828501Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 9 21:18:14.748382 systemd-networkd[1433]: calieeaef2f2d6a: Gained IPv6LL Sep 9 21:18:15.331505 containerd[1525]: time="2025-09-09T21:18:15.331256581Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-vqfpp,Uid:3b79aa0a-edd4-45ee-8637-43c015678f95,Namespace:calico-apiserver,Attempt:0,}" Sep 9 21:18:15.331505 containerd[1525]: time="2025-09-09T21:18:15.331266221Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-q45ms,Uid:ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:15.570172 containerd[1525]: time="2025-09-09T21:18:15.570128803Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:15.572324 containerd[1525]: time="2025-09-09T21:18:15.572289428Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=44530807" Sep 9 21:18:15.572383 containerd[1525]: time="2025-09-09T21:18:15.572369067Z" level=info msg="ImageCreate event name:\"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:15.574932 containerd[1525]: time="2025-09-09T21:18:15.574889650Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:15.575567 containerd[1525]: time="2025-09-09T21:18:15.575530406Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"45900064\" in 2.014669505s" Sep 9 21:18:15.575567 containerd[1525]: time="2025-09-09T21:18:15.575560566Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\"" Sep 9 21:18:15.588892 containerd[1525]: time="2025-09-09T21:18:15.588240040Z" level=info msg="CreateContainer within sandbox \"55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 9 21:18:15.607209 containerd[1525]: time="2025-09-09T21:18:15.606911593Z" level=info msg="Container 13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:15.648428 containerd[1525]: time="2025-09-09T21:18:15.647258600Z" level=info msg="CreateContainer within sandbox \"55f69fd4631430e5ead8cd79fe7d1604574dd30e8973a619654cb71546ffc97a\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01\"" Sep 9 21:18:15.649112 containerd[1525]: time="2025-09-09T21:18:15.649080388Z" level=info msg="StartContainer for \"13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01\"" Sep 9 21:18:15.653533 containerd[1525]: time="2025-09-09T21:18:15.653484878Z" level=info msg="connecting to shim 13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01" address="unix:///run/containerd/s/691c251f2f0ea98a8bf25d5252008ce9963cb10aece1eb73fda73d7310c9419a" protocol=ttrpc version=3 Sep 9 21:18:15.654963 systemd-networkd[1433]: cali00e65c2ffce: Link UP Sep 9 21:18:15.657239 systemd-networkd[1433]: cali00e65c2ffce: Gained carrier Sep 9 21:18:15.681280 containerd[1525]: 2025-09-09 21:18:15.540 [INFO][4347] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:15.681280 containerd[1525]: 2025-09-09 21:18:15.559 [INFO][4347] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0 calico-apiserver-f65bff7c7- calico-apiserver 3b79aa0a-edd4-45ee-8637-43c015678f95 807 0 2025-09-09 21:17:51 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:f65bff7c7 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-f65bff7c7-vqfpp eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali00e65c2ffce [] [] }} ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-" Sep 9 21:18:15.681280 containerd[1525]: 2025-09-09 21:18:15.559 [INFO][4347] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.681280 containerd[1525]: 2025-09-09 21:18:15.594 [INFO][4378] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" HandleID="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Workload="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.594 [INFO][4378] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" HandleID="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Workload="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c30b0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-f65bff7c7-vqfpp", "timestamp":"2025-09-09 21:18:15.594544157 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.594 [INFO][4378] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.594 [INFO][4378] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.594 [INFO][4378] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.612 [INFO][4378] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" host="localhost" Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.619 [INFO][4378] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.626 [INFO][4378] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.628 [INFO][4378] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.631 [INFO][4378] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:15.681516 containerd[1525]: 2025-09-09 21:18:15.631 [INFO][4378] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" host="localhost" Sep 9 21:18:15.681725 containerd[1525]: 2025-09-09 21:18:15.633 [INFO][4378] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71 Sep 9 21:18:15.681725 containerd[1525]: 2025-09-09 21:18:15.638 [INFO][4378] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" host="localhost" Sep 9 21:18:15.681725 containerd[1525]: 2025-09-09 21:18:15.643 [INFO][4378] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" host="localhost" Sep 9 21:18:15.681725 containerd[1525]: 2025-09-09 21:18:15.644 [INFO][4378] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" host="localhost" Sep 9 21:18:15.681725 containerd[1525]: 2025-09-09 21:18:15.644 [INFO][4378] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:15.681725 containerd[1525]: 2025-09-09 21:18:15.644 [INFO][4378] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" HandleID="k8s-pod-network.f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Workload="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.681888 containerd[1525]: 2025-09-09 21:18:15.649 [INFO][4347] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0", GenerateName:"calico-apiserver-f65bff7c7-", Namespace:"calico-apiserver", SelfLink:"", UID:"3b79aa0a-edd4-45ee-8637-43c015678f95", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f65bff7c7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-f65bff7c7-vqfpp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali00e65c2ffce", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:15.681940 containerd[1525]: 2025-09-09 21:18:15.650 [INFO][4347] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.681940 containerd[1525]: 2025-09-09 21:18:15.650 [INFO][4347] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali00e65c2ffce ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.681940 containerd[1525]: 2025-09-09 21:18:15.657 [INFO][4347] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.682001 containerd[1525]: 2025-09-09 21:18:15.658 [INFO][4347] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0", GenerateName:"calico-apiserver-f65bff7c7-", Namespace:"calico-apiserver", SelfLink:"", UID:"3b79aa0a-edd4-45ee-8637-43c015678f95", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f65bff7c7", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71", Pod:"calico-apiserver-f65bff7c7-vqfpp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali00e65c2ffce", MAC:"b6:b9:43:a7:e3:21", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:15.682047 containerd[1525]: 2025-09-09 21:18:15.672 [INFO][4347] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" Namespace="calico-apiserver" Pod="calico-apiserver-f65bff7c7-vqfpp" WorkloadEndpoint="localhost-k8s-calico--apiserver--f65bff7c7--vqfpp-eth0" Sep 9 21:18:15.697241 systemd[1]: Started cri-containerd-13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01.scope - libcontainer container 13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01. Sep 9 21:18:15.729453 containerd[1525]: time="2025-09-09T21:18:15.728961647Z" level=info msg="connecting to shim f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71" address="unix:///run/containerd/s/71bc7f8fbe35388a2b019e245e68e0a46428cd1959186fe5ee58b772242c2d2a" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:15.789364 systemd[1]: Started cri-containerd-f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71.scope - libcontainer container f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71. Sep 9 21:18:15.799365 systemd[1]: Started sshd@7-10.0.0.66:22-10.0.0.1:54752.service - OpenSSH per-connection server daemon (10.0.0.1:54752). Sep 9 21:18:15.818295 systemd-networkd[1433]: cali6a3a30f1f30: Link UP Sep 9 21:18:15.819216 systemd-networkd[1433]: cali6a3a30f1f30: Gained carrier Sep 9 21:18:15.824413 containerd[1525]: time="2025-09-09T21:18:15.823254208Z" level=info msg="StartContainer for \"13a011d336168a98916ec3f2e09ab21495f5c0bc4c534791b64c9660f26b7c01\" returns successfully" Sep 9 21:18:15.830826 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:15.845889 containerd[1525]: 2025-09-09 21:18:15.582 [INFO][4361] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:15.845889 containerd[1525]: 2025-09-09 21:18:15.616 [INFO][4361] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--q45ms-eth0 csi-node-driver- calico-system ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03 704 0 2025-09-09 21:17:54 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-q45ms eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali6a3a30f1f30 [] [] }} ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-" Sep 9 21:18:15.845889 containerd[1525]: 2025-09-09 21:18:15.617 [INFO][4361] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.845889 containerd[1525]: 2025-09-09 21:18:15.651 [INFO][4395] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" HandleID="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Workload="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.651 [INFO][4395] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" HandleID="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Workload="localhost-k8s-csi--node--driver--q45ms-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004d7b0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-q45ms", "timestamp":"2025-09-09 21:18:15.651432932 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.651 [INFO][4395] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.651 [INFO][4395] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.651 [INFO][4395] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.712 [INFO][4395] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" host="localhost" Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.735 [INFO][4395] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.754 [INFO][4395] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.760 [INFO][4395] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.771 [INFO][4395] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:15.846216 containerd[1525]: 2025-09-09 21:18:15.772 [INFO][4395] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" host="localhost" Sep 9 21:18:15.846479 containerd[1525]: 2025-09-09 21:18:15.775 [INFO][4395] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0 Sep 9 21:18:15.846479 containerd[1525]: 2025-09-09 21:18:15.781 [INFO][4395] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" host="localhost" Sep 9 21:18:15.846479 containerd[1525]: 2025-09-09 21:18:15.798 [INFO][4395] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" host="localhost" Sep 9 21:18:15.846479 containerd[1525]: 2025-09-09 21:18:15.798 [INFO][4395] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" host="localhost" Sep 9 21:18:15.846479 containerd[1525]: 2025-09-09 21:18:15.798 [INFO][4395] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:15.846479 containerd[1525]: 2025-09-09 21:18:15.798 [INFO][4395] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" HandleID="k8s-pod-network.5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Workload="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.846600 containerd[1525]: 2025-09-09 21:18:15.808 [INFO][4361] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--q45ms-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03", ResourceVersion:"704", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-q45ms", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6a3a30f1f30", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:15.846655 containerd[1525]: 2025-09-09 21:18:15.809 [INFO][4361] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.846655 containerd[1525]: 2025-09-09 21:18:15.810 [INFO][4361] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6a3a30f1f30 ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.846655 containerd[1525]: 2025-09-09 21:18:15.820 [INFO][4361] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.846710 containerd[1525]: 2025-09-09 21:18:15.820 [INFO][4361] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--q45ms-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03", ResourceVersion:"704", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0", Pod:"csi-node-driver-q45ms", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6a3a30f1f30", MAC:"e6:3b:93:73:35:82", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:15.846757 containerd[1525]: 2025-09-09 21:18:15.836 [INFO][4361] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" Namespace="calico-system" Pod="csi-node-driver-q45ms" WorkloadEndpoint="localhost-k8s-csi--node--driver--q45ms-eth0" Sep 9 21:18:15.885021 containerd[1525]: time="2025-09-09T21:18:15.883388440Z" level=info msg="connecting to shim 5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0" address="unix:///run/containerd/s/38873af70b5bf5d47eec87069796b17c7ca083ad07bc514551eb44de3c90496a" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:15.888746 sshd[4487]: Accepted publickey for core from 10.0.0.1 port 54752 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:15.891352 sshd-session[4487]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:15.897086 containerd[1525]: time="2025-09-09T21:18:15.897036068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f65bff7c7-vqfpp,Uid:3b79aa0a-edd4-45ee-8637-43c015678f95,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71\"" Sep 9 21:18:15.899323 systemd-logind[1509]: New session 8 of user core. Sep 9 21:18:15.902508 containerd[1525]: time="2025-09-09T21:18:15.902476711Z" level=info msg="CreateContainer within sandbox \"f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 9 21:18:15.906058 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 9 21:18:15.912973 containerd[1525]: time="2025-09-09T21:18:15.912928560Z" level=info msg="Container f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:15.920996 systemd[1]: Started cri-containerd-5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0.scope - libcontainer container 5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0. Sep 9 21:18:15.923150 containerd[1525]: time="2025-09-09T21:18:15.923096891Z" level=info msg="CreateContainer within sandbox \"f1ee0cfed2cb1139fa33fb3d84eb5504cac310e01cdee4ecf99db9bc18c4ba71\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43\"" Sep 9 21:18:15.923915 containerd[1525]: time="2025-09-09T21:18:15.923880286Z" level=info msg="StartContainer for \"f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43\"" Sep 9 21:18:15.925010 containerd[1525]: time="2025-09-09T21:18:15.924983238Z" level=info msg="connecting to shim f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43" address="unix:///run/containerd/s/71bc7f8fbe35388a2b019e245e68e0a46428cd1959186fe5ee58b772242c2d2a" protocol=ttrpc version=3 Sep 9 21:18:15.942375 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:15.948957 systemd[1]: Started cri-containerd-f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43.scope - libcontainer container f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43. Sep 9 21:18:15.962034 containerd[1525]: time="2025-09-09T21:18:15.961992148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-q45ms,Uid:ec7a9498-a0d9-4dbc-a0ab-af5696ebcf03,Namespace:calico-system,Attempt:0,} returns sandbox id \"5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0\"" Sep 9 21:18:15.966423 containerd[1525]: time="2025-09-09T21:18:15.965366165Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 9 21:18:16.011965 containerd[1525]: time="2025-09-09T21:18:16.011920931Z" level=info msg="StartContainer for \"f955684790845643b533881664ed0c47be1d5333bf8ee7f0dd194f92d9baaf43\" returns successfully" Sep 9 21:18:16.178169 sshd[4547]: Connection closed by 10.0.0.1 port 54752 Sep 9 21:18:16.178611 sshd-session[4487]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:16.182635 systemd[1]: sshd@7-10.0.0.66:22-10.0.0.1:54752.service: Deactivated successfully. Sep 9 21:18:16.184453 systemd[1]: session-8.scope: Deactivated successfully. Sep 9 21:18:16.186744 systemd-logind[1509]: Session 8 logged out. Waiting for processes to exit. Sep 9 21:18:16.188862 systemd-logind[1509]: Removed session 8. Sep 9 21:18:16.326501 kubelet[2660]: E0909 21:18:16.325823 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:16.328336 containerd[1525]: time="2025-09-09T21:18:16.327180052Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-95rlw,Uid:319137ec-563e-4c4c-ba44-c5fde18a3fdf,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:16.329754 kubelet[2660]: E0909 21:18:16.326188 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:16.329833 containerd[1525]: time="2025-09-09T21:18:16.329022640Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sqj7,Uid:ac77a038-6bc7-49b6-acb6-74bcc3a30baa,Namespace:kube-system,Attempt:0,}" Sep 9 21:18:16.330489 containerd[1525]: time="2025-09-09T21:18:16.330333551Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gr8qj,Uid:caa9dc8e-593a-4c15-8b92-f1c9b4823e5d,Namespace:kube-system,Attempt:0,}" Sep 9 21:18:16.477418 systemd-networkd[1433]: calibc0347e3635: Link UP Sep 9 21:18:16.477570 systemd-networkd[1433]: calibc0347e3635: Gained carrier Sep 9 21:18:16.496231 containerd[1525]: 2025-09-09 21:18:16.360 [INFO][4613] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:16.496231 containerd[1525]: 2025-09-09 21:18:16.378 [INFO][4613] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0 coredns-668d6bf9bc- kube-system ac77a038-6bc7-49b6-acb6-74bcc3a30baa 802 0 2025-09-09 21:17:42 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-7sqj7 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calibc0347e3635 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-" Sep 9 21:18:16.496231 containerd[1525]: 2025-09-09 21:18:16.378 [INFO][4613] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.496231 containerd[1525]: 2025-09-09 21:18:16.435 [INFO][4655] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" HandleID="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Workload="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.436 [INFO][4655] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" HandleID="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Workload="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000340240), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-7sqj7", "timestamp":"2025-09-09 21:18:16.435959375 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.436 [INFO][4655] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.436 [INFO][4655] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.436 [INFO][4655] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.448 [INFO][4655] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" host="localhost" Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.452 [INFO][4655] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.456 [INFO][4655] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.458 [INFO][4655] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.461 [INFO][4655] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:16.496734 containerd[1525]: 2025-09-09 21:18:16.461 [INFO][4655] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" host="localhost" Sep 9 21:18:16.497374 containerd[1525]: 2025-09-09 21:18:16.463 [INFO][4655] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8 Sep 9 21:18:16.497374 containerd[1525]: 2025-09-09 21:18:16.466 [INFO][4655] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" host="localhost" Sep 9 21:18:16.497374 containerd[1525]: 2025-09-09 21:18:16.472 [INFO][4655] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" host="localhost" Sep 9 21:18:16.497374 containerd[1525]: 2025-09-09 21:18:16.472 [INFO][4655] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" host="localhost" Sep 9 21:18:16.497374 containerd[1525]: 2025-09-09 21:18:16.472 [INFO][4655] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:16.497374 containerd[1525]: 2025-09-09 21:18:16.472 [INFO][4655] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" HandleID="k8s-pod-network.685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Workload="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.497487 containerd[1525]: 2025-09-09 21:18:16.475 [INFO][4613] cni-plugin/k8s.go 418: Populated endpoint ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"ac77a038-6bc7-49b6-acb6-74bcc3a30baa", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-7sqj7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calibc0347e3635", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:16.497559 containerd[1525]: 2025-09-09 21:18:16.475 [INFO][4613] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.497559 containerd[1525]: 2025-09-09 21:18:16.475 [INFO][4613] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibc0347e3635 ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.497559 containerd[1525]: 2025-09-09 21:18:16.480 [INFO][4613] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.497618 containerd[1525]: 2025-09-09 21:18:16.480 [INFO][4613] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"ac77a038-6bc7-49b6-acb6-74bcc3a30baa", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8", Pod:"coredns-668d6bf9bc-7sqj7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calibc0347e3635", MAC:"be:71:9f:76:fe:bb", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:16.497618 containerd[1525]: 2025-09-09 21:18:16.491 [INFO][4613] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" Namespace="kube-system" Pod="coredns-668d6bf9bc-7sqj7" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--7sqj7-eth0" Sep 9 21:18:16.514931 kubelet[2660]: I0909 21:18:16.514730 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-f65bff7c7-9kk94" podStartSLOduration=23.492303202 podStartE2EDuration="25.514713215s" podCreationTimestamp="2025-09-09 21:17:51 +0000 UTC" firstStartedPulling="2025-09-09 21:18:13.560386184 +0000 UTC m=+38.312942103" lastFinishedPulling="2025-09-09 21:18:15.582796237 +0000 UTC m=+40.335352116" observedRunningTime="2025-09-09 21:18:16.50943689 +0000 UTC m=+41.261992809" watchObservedRunningTime="2025-09-09 21:18:16.514713215 +0000 UTC m=+41.267269134" Sep 9 21:18:16.523023 kubelet[2660]: I0909 21:18:16.522479 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-f65bff7c7-vqfpp" podStartSLOduration=25.522462404 podStartE2EDuration="25.522462404s" podCreationTimestamp="2025-09-09 21:17:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:18:16.522459924 +0000 UTC m=+41.275015843" watchObservedRunningTime="2025-09-09 21:18:16.522462404 +0000 UTC m=+41.275018283" Sep 9 21:18:16.540931 containerd[1525]: time="2025-09-09T21:18:16.540890282Z" level=info msg="connecting to shim 685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8" address="unix:///run/containerd/s/77c1fca1f2d7d405c14398d7237329e577a9ab76edcee57aa95e3558c0968443" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:16.572999 systemd[1]: Started cri-containerd-685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8.scope - libcontainer container 685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8. Sep 9 21:18:16.602323 systemd-networkd[1433]: cali1b960bf0488: Link UP Sep 9 21:18:16.608852 systemd-networkd[1433]: cali1b960bf0488: Gained carrier Sep 9 21:18:16.623231 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.380 [INFO][4626] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.399 [INFO][4626] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--54d579b49d--95rlw-eth0 goldmane-54d579b49d- calico-system 319137ec-563e-4c4c-ba44-c5fde18a3fdf 805 0 2025-09-09 21:17:53 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-54d579b49d-95rlw eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali1b960bf0488 [] [] }} ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.399 [INFO][4626] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.454 [INFO][4663] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" HandleID="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Workload="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.454 [INFO][4663] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" HandleID="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Workload="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c7d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-54d579b49d-95rlw", "timestamp":"2025-09-09 21:18:16.453961096 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.454 [INFO][4663] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.472 [INFO][4663] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.473 [INFO][4663] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.550 [INFO][4663] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.558 [INFO][4663] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.567 [INFO][4663] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.569 [INFO][4663] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.575 [INFO][4663] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.575 [INFO][4663] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.577 [INFO][4663] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909 Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.582 [INFO][4663] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.590 [INFO][4663] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.590 [INFO][4663] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" host="localhost" Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.590 [INFO][4663] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:16.629077 containerd[1525]: 2025-09-09 21:18:16.590 [INFO][4663] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" HandleID="k8s-pod-network.5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Workload="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.629572 containerd[1525]: 2025-09-09 21:18:16.595 [INFO][4626] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--95rlw-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"319137ec-563e-4c4c-ba44-c5fde18a3fdf", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-54d579b49d-95rlw", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali1b960bf0488", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:16.629572 containerd[1525]: 2025-09-09 21:18:16.595 [INFO][4626] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.629572 containerd[1525]: 2025-09-09 21:18:16.595 [INFO][4626] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1b960bf0488 ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.629572 containerd[1525]: 2025-09-09 21:18:16.607 [INFO][4626] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.629572 containerd[1525]: 2025-09-09 21:18:16.610 [INFO][4626] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--95rlw-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"319137ec-563e-4c4c-ba44-c5fde18a3fdf", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909", Pod:"goldmane-54d579b49d-95rlw", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali1b960bf0488", MAC:"1a:53:32:76:9f:4e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:16.629572 containerd[1525]: 2025-09-09 21:18:16.621 [INFO][4626] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" Namespace="calico-system" Pod="goldmane-54d579b49d-95rlw" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--95rlw-eth0" Sep 9 21:18:16.651977 containerd[1525]: time="2025-09-09T21:18:16.651700632Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sqj7,Uid:ac77a038-6bc7-49b6-acb6-74bcc3a30baa,Namespace:kube-system,Attempt:0,} returns sandbox id \"685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8\"" Sep 9 21:18:16.652801 kubelet[2660]: E0909 21:18:16.652756 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:16.656596 containerd[1525]: time="2025-09-09T21:18:16.656499120Z" level=info msg="CreateContainer within sandbox \"685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 9 21:18:16.658950 containerd[1525]: time="2025-09-09T21:18:16.658911624Z" level=info msg="connecting to shim 5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909" address="unix:///run/containerd/s/b3203956d03d24f75e9a009f2c522100d073b6923032390511605599b9224457" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:16.672399 containerd[1525]: time="2025-09-09T21:18:16.672014098Z" level=info msg="Container 4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:16.682575 containerd[1525]: time="2025-09-09T21:18:16.682495748Z" level=info msg="CreateContainer within sandbox \"685fe974531b2b2a40be155a648e1d75ca4133c28aaddeb438ad30048b74e8a8\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e\"" Sep 9 21:18:16.683457 containerd[1525]: time="2025-09-09T21:18:16.683423422Z" level=info msg="StartContainer for \"4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e\"" Sep 9 21:18:16.687425 containerd[1525]: time="2025-09-09T21:18:16.687181438Z" level=info msg="connecting to shim 4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e" address="unix:///run/containerd/s/77c1fca1f2d7d405c14398d7237329e577a9ab76edcee57aa95e3558c0968443" protocol=ttrpc version=3 Sep 9 21:18:16.698046 systemd[1]: Started cri-containerd-5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909.scope - libcontainer container 5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909. Sep 9 21:18:16.699907 systemd-networkd[1433]: cali252b1dbac29: Link UP Sep 9 21:18:16.701032 systemd-networkd[1433]: cali252b1dbac29: Gained carrier Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.379 [INFO][4623] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.400 [INFO][4623] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0 coredns-668d6bf9bc- kube-system caa9dc8e-593a-4c15-8b92-f1c9b4823e5d 810 0 2025-09-09 21:17:42 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-gr8qj eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali252b1dbac29 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.400 [INFO][4623] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.462 [INFO][4669] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" HandleID="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Workload="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.463 [INFO][4669] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" HandleID="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Workload="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000137a70), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-gr8qj", "timestamp":"2025-09-09 21:18:16.462866837 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.463 [INFO][4669] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.590 [INFO][4669] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.592 [INFO][4669] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.649 [INFO][4669] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.657 [INFO][4669] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.671 [INFO][4669] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.673 [INFO][4669] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.676 [INFO][4669] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.676 [INFO][4669] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.677 [INFO][4669] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48 Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.682 [INFO][4669] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.690 [INFO][4669] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.690 [INFO][4669] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" host="localhost" Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.690 [INFO][4669] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:16.717358 containerd[1525]: 2025-09-09 21:18:16.690 [INFO][4669] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" HandleID="k8s-pod-network.6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Workload="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.718174 containerd[1525]: 2025-09-09 21:18:16.695 [INFO][4623] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"caa9dc8e-593a-4c15-8b92-f1c9b4823e5d", ResourceVersion:"810", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-gr8qj", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali252b1dbac29", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:16.718174 containerd[1525]: 2025-09-09 21:18:16.696 [INFO][4623] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.718174 containerd[1525]: 2025-09-09 21:18:16.696 [INFO][4623] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali252b1dbac29 ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.718174 containerd[1525]: 2025-09-09 21:18:16.701 [INFO][4623] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.718174 containerd[1525]: 2025-09-09 21:18:16.702 [INFO][4623] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"caa9dc8e-593a-4c15-8b92-f1c9b4823e5d", ResourceVersion:"810", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48", Pod:"coredns-668d6bf9bc-gr8qj", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali252b1dbac29", MAC:"a6:df:11:55:48:3b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:16.718174 containerd[1525]: 2025-09-09 21:18:16.715 [INFO][4623] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" Namespace="kube-system" Pod="coredns-668d6bf9bc-gr8qj" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gr8qj-eth0" Sep 9 21:18:16.726981 systemd[1]: Started cri-containerd-4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e.scope - libcontainer container 4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e. Sep 9 21:18:16.734537 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:16.747533 containerd[1525]: time="2025-09-09T21:18:16.747436160Z" level=info msg="connecting to shim 6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48" address="unix:///run/containerd/s/4f887ef5ee918d003f621df02143296271ecffdc5f5a8396ff064ff265fd0711" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:16.768092 containerd[1525]: time="2025-09-09T21:18:16.768053704Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-95rlw,Uid:319137ec-563e-4c4c-ba44-c5fde18a3fdf,Namespace:calico-system,Attempt:0,} returns sandbox id \"5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909\"" Sep 9 21:18:16.775434 containerd[1525]: time="2025-09-09T21:18:16.775392456Z" level=info msg="StartContainer for \"4bd2fc9e1f18b5eb91914d92feaf7a9c1de371623767d20b90de40e5ce7a211e\" returns successfully" Sep 9 21:18:16.787940 systemd[1]: Started cri-containerd-6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48.scope - libcontainer container 6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48. Sep 9 21:18:16.802851 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:16.832278 containerd[1525]: time="2025-09-09T21:18:16.832230001Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gr8qj,Uid:caa9dc8e-593a-4c15-8b92-f1c9b4823e5d,Namespace:kube-system,Attempt:0,} returns sandbox id \"6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48\"" Sep 9 21:18:16.834957 kubelet[2660]: E0909 21:18:16.834924 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:16.840093 containerd[1525]: time="2025-09-09T21:18:16.840044429Z" level=info msg="CreateContainer within sandbox \"6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 9 21:18:16.856287 containerd[1525]: time="2025-09-09T21:18:16.856130403Z" level=info msg="Container db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:16.866835 containerd[1525]: time="2025-09-09T21:18:16.865426502Z" level=info msg="CreateContainer within sandbox \"6dccc8c514d1aa42b82c4929e6cb50c6811ed2a89dea17e8d2624099fad2de48\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144\"" Sep 9 21:18:16.868658 containerd[1525]: time="2025-09-09T21:18:16.868625841Z" level=info msg="StartContainer for \"db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144\"" Sep 9 21:18:16.869451 containerd[1525]: time="2025-09-09T21:18:16.869420036Z" level=info msg="connecting to shim db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144" address="unix:///run/containerd/s/4f887ef5ee918d003f621df02143296271ecffdc5f5a8396ff064ff265fd0711" protocol=ttrpc version=3 Sep 9 21:18:16.894981 systemd[1]: Started cri-containerd-db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144.scope - libcontainer container db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144. Sep 9 21:18:16.939272 containerd[1525]: time="2025-09-09T21:18:16.939211935Z" level=info msg="StartContainer for \"db6f778f04facf9b568d773f010973c85717a9dab58596854b40b9284abd9144\" returns successfully" Sep 9 21:18:17.154283 containerd[1525]: time="2025-09-09T21:18:17.154239863Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:17.155600 containerd[1525]: time="2025-09-09T21:18:17.155568694Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8227489" Sep 9 21:18:17.156167 containerd[1525]: time="2025-09-09T21:18:17.156130051Z" level=info msg="ImageCreate event name:\"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:17.158604 containerd[1525]: time="2025-09-09T21:18:17.158570275Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:17.159224 containerd[1525]: time="2025-09-09T21:18:17.159197351Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"9596730\" in 1.193792546s" Sep 9 21:18:17.159261 containerd[1525]: time="2025-09-09T21:18:17.159228711Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\"" Sep 9 21:18:17.160596 containerd[1525]: time="2025-09-09T21:18:17.160574182Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 9 21:18:17.163146 containerd[1525]: time="2025-09-09T21:18:17.163116046Z" level=info msg="CreateContainer within sandbox \"5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 9 21:18:17.173000 containerd[1525]: time="2025-09-09T21:18:17.172969142Z" level=info msg="Container 3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:17.192195 containerd[1525]: time="2025-09-09T21:18:17.192145379Z" level=info msg="CreateContainer within sandbox \"5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03\"" Sep 9 21:18:17.193815 containerd[1525]: time="2025-09-09T21:18:17.193656969Z" level=info msg="StartContainer for \"3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03\"" Sep 9 21:18:17.196841 containerd[1525]: time="2025-09-09T21:18:17.195602877Z" level=info msg="connecting to shim 3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03" address="unix:///run/containerd/s/38873af70b5bf5d47eec87069796b17c7ca083ad07bc514551eb44de3c90496a" protocol=ttrpc version=3 Sep 9 21:18:17.227048 systemd[1]: Started cri-containerd-3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03.scope - libcontainer container 3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03. Sep 9 21:18:17.298437 containerd[1525]: time="2025-09-09T21:18:17.298393736Z" level=info msg="StartContainer for \"3494349b14a243c28d144699dfc139ced46fcbd014ea26222b6a6e27b7772e03\" returns successfully" Sep 9 21:18:17.334284 containerd[1525]: time="2025-09-09T21:18:17.334199506Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6f9758d6db-9jkm7,Uid:1d112637-357c-40e9-9e80-6fa0222fdf7a,Namespace:calico-system,Attempt:0,}" Sep 9 21:18:17.478929 systemd-networkd[1433]: calic11378a124f: Link UP Sep 9 21:18:17.479539 systemd-networkd[1433]: calic11378a124f: Gained carrier Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.364 [INFO][4972] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.380 [INFO][4972] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0 calico-kube-controllers-6f9758d6db- calico-system 1d112637-357c-40e9-9e80-6fa0222fdf7a 809 0 2025-09-09 21:17:54 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6f9758d6db projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-6f9758d6db-9jkm7 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calic11378a124f [] [] }} ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.381 [INFO][4972] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.422 [INFO][4986] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" HandleID="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Workload="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.423 [INFO][4986] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" HandleID="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Workload="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001377a0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-6f9758d6db-9jkm7", "timestamp":"2025-09-09 21:18:17.42231614 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.423 [INFO][4986] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.423 [INFO][4986] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.423 [INFO][4986] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.437 [INFO][4986] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.447 [INFO][4986] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.452 [INFO][4986] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.454 [INFO][4986] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.457 [INFO][4986] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.457 [INFO][4986] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.459 [INFO][4986] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04 Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.466 [INFO][4986] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.472 [INFO][4986] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.472 [INFO][4986] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" host="localhost" Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.472 [INFO][4986] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 9 21:18:17.493351 containerd[1525]: 2025-09-09 21:18:17.472 [INFO][4986] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" HandleID="k8s-pod-network.cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Workload="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.494722 containerd[1525]: 2025-09-09 21:18:17.476 [INFO][4972] cni-plugin/k8s.go 418: Populated endpoint ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0", GenerateName:"calico-kube-controllers-6f9758d6db-", Namespace:"calico-system", SelfLink:"", UID:"1d112637-357c-40e9-9e80-6fa0222fdf7a", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6f9758d6db", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-6f9758d6db-9jkm7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic11378a124f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:17.494722 containerd[1525]: 2025-09-09 21:18:17.476 [INFO][4972] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.494722 containerd[1525]: 2025-09-09 21:18:17.476 [INFO][4972] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic11378a124f ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.494722 containerd[1525]: 2025-09-09 21:18:17.477 [INFO][4972] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.494722 containerd[1525]: 2025-09-09 21:18:17.478 [INFO][4972] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0", GenerateName:"calico-kube-controllers-6f9758d6db-", Namespace:"calico-system", SelfLink:"", UID:"1d112637-357c-40e9-9e80-6fa0222fdf7a", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.September, 9, 21, 17, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6f9758d6db", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04", Pod:"calico-kube-controllers-6f9758d6db-9jkm7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic11378a124f", MAC:"b2:9e:da:77:9c:5c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 9 21:18:17.494722 containerd[1525]: 2025-09-09 21:18:17.490 [INFO][4972] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" Namespace="calico-system" Pod="calico-kube-controllers-6f9758d6db-9jkm7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6f9758d6db--9jkm7-eth0" Sep 9 21:18:17.518328 kubelet[2660]: E0909 21:18:17.518294 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:17.521285 containerd[1525]: time="2025-09-09T21:18:17.521236744Z" level=info msg="connecting to shim cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04" address="unix:///run/containerd/s/0ac2d771646c83e48756dd07d785fc72470f9a189c25d02043dbc11c4c0ab0d5" namespace=k8s.io protocol=ttrpc version=3 Sep 9 21:18:17.530570 kubelet[2660]: E0909 21:18:17.530526 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:17.540568 kubelet[2660]: I0909 21:18:17.538188 2660 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 9 21:18:17.560002 systemd[1]: Started cri-containerd-cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04.scope - libcontainer container cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04. Sep 9 21:18:17.561590 kubelet[2660]: I0909 21:18:17.561525 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-gr8qj" podStartSLOduration=35.561381886 podStartE2EDuration="35.561381886s" podCreationTimestamp="2025-09-09 21:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:18:17.536211407 +0000 UTC m=+42.288767366" watchObservedRunningTime="2025-09-09 21:18:17.561381886 +0000 UTC m=+42.313937805" Sep 9 21:18:17.561873 kubelet[2660]: I0909 21:18:17.561835 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-7sqj7" podStartSLOduration=35.561824763 podStartE2EDuration="35.561824763s" podCreationTimestamp="2025-09-09 21:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 21:18:17.558505304 +0000 UTC m=+42.311061223" watchObservedRunningTime="2025-09-09 21:18:17.561824763 +0000 UTC m=+42.314380682" Sep 9 21:18:17.595164 systemd-resolved[1358]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 21:18:17.625905 containerd[1525]: time="2025-09-09T21:18:17.625768112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6f9758d6db-9jkm7,Uid:1d112637-357c-40e9-9e80-6fa0222fdf7a,Namespace:calico-system,Attempt:0,} returns sandbox id \"cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04\"" Sep 9 21:18:17.628007 systemd-networkd[1433]: cali00e65c2ffce: Gained IPv6LL Sep 9 21:18:17.691909 systemd-networkd[1433]: cali6a3a30f1f30: Gained IPv6LL Sep 9 21:18:17.756946 systemd-networkd[1433]: cali252b1dbac29: Gained IPv6LL Sep 9 21:18:18.011943 systemd-networkd[1433]: calibc0347e3635: Gained IPv6LL Sep 9 21:18:18.332990 systemd-networkd[1433]: cali1b960bf0488: Gained IPv6LL Sep 9 21:18:18.538999 kubelet[2660]: E0909 21:18:18.538961 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:18.540334 kubelet[2660]: E0909 21:18:18.539634 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:18.653177 systemd-networkd[1433]: calic11378a124f: Gained IPv6LL Sep 9 21:18:19.011336 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1948437342.mount: Deactivated successfully. Sep 9 21:18:19.411367 containerd[1525]: time="2025-09-09T21:18:19.411321518Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:19.412249 containerd[1525]: time="2025-09-09T21:18:19.412062673Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=61845332" Sep 9 21:18:19.413001 containerd[1525]: time="2025-09-09T21:18:19.412964668Z" level=info msg="ImageCreate event name:\"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:19.415346 containerd[1525]: time="2025-09-09T21:18:19.415319293Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:19.416113 containerd[1525]: time="2025-09-09T21:18:19.416084689Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"61845178\" in 2.255482387s" Sep 9 21:18:19.416176 containerd[1525]: time="2025-09-09T21:18:19.416115888Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\"" Sep 9 21:18:19.417079 containerd[1525]: time="2025-09-09T21:18:19.417057843Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 9 21:18:19.418636 containerd[1525]: time="2025-09-09T21:18:19.418588793Z" level=info msg="CreateContainer within sandbox \"5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 9 21:18:19.427969 containerd[1525]: time="2025-09-09T21:18:19.427026422Z" level=info msg="Container b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:19.429324 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount605735613.mount: Deactivated successfully. Sep 9 21:18:19.434486 containerd[1525]: time="2025-09-09T21:18:19.434458176Z" level=info msg="CreateContainer within sandbox \"5906d6e35f2d729f815927b094dd322eb2887f5949e957e9e8044780ccca1909\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec\"" Sep 9 21:18:19.434941 containerd[1525]: time="2025-09-09T21:18:19.434917093Z" level=info msg="StartContainer for \"b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec\"" Sep 9 21:18:19.436611 containerd[1525]: time="2025-09-09T21:18:19.436584763Z" level=info msg="connecting to shim b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec" address="unix:///run/containerd/s/b3203956d03d24f75e9a009f2c522100d073b6923032390511605599b9224457" protocol=ttrpc version=3 Sep 9 21:18:19.469975 systemd[1]: Started cri-containerd-b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec.scope - libcontainer container b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec. Sep 9 21:18:19.515518 containerd[1525]: time="2025-09-09T21:18:19.515483440Z" level=info msg="StartContainer for \"b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec\" returns successfully" Sep 9 21:18:19.544446 kubelet[2660]: E0909 21:18:19.544415 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:19.546061 kubelet[2660]: E0909 21:18:19.546032 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:19.557163 kubelet[2660]: I0909 21:18:19.557106 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-95rlw" podStartSLOduration=23.910101953 podStartE2EDuration="26.556924946s" podCreationTimestamp="2025-09-09 21:17:53 +0000 UTC" firstStartedPulling="2025-09-09 21:18:16.770055971 +0000 UTC m=+41.522611890" lastFinishedPulling="2025-09-09 21:18:19.416878964 +0000 UTC m=+44.169434883" observedRunningTime="2025-09-09 21:18:19.556613348 +0000 UTC m=+44.309169267" watchObservedRunningTime="2025-09-09 21:18:19.556924946 +0000 UTC m=+44.309480865" Sep 9 21:18:20.634602 containerd[1525]: time="2025-09-09T21:18:20.634544916Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec\" id:\"16e2c9dc8784e71c4044b0b27df3771a8b6fc532d67d53574ea477f3a2120854\" pid:5187 exit_status:1 exited_at:{seconds:1757452700 nanos:629498267}" Sep 9 21:18:20.636751 kubelet[2660]: I0909 21:18:20.636723 2660 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 9 21:18:20.638267 kubelet[2660]: E0909 21:18:20.638245 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:21.191712 systemd[1]: Started sshd@8-10.0.0.66:22-10.0.0.1:42842.service - OpenSSH per-connection server daemon (10.0.0.1:42842). Sep 9 21:18:21.269967 sshd[5214]: Accepted publickey for core from 10.0.0.1 port 42842 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:21.269863 sshd-session[5214]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:21.276700 systemd-logind[1509]: New session 9 of user core. Sep 9 21:18:21.280975 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 9 21:18:21.393776 containerd[1525]: time="2025-09-09T21:18:21.393717785Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:21.395096 containerd[1525]: time="2025-09-09T21:18:21.395037057Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=13761208" Sep 9 21:18:21.396589 containerd[1525]: time="2025-09-09T21:18:21.396537728Z" level=info msg="ImageCreate event name:\"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:21.398696 containerd[1525]: time="2025-09-09T21:18:21.398655076Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:21.399721 containerd[1525]: time="2025-09-09T21:18:21.399685350Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"15130401\" in 1.982598148s" Sep 9 21:18:21.399721 containerd[1525]: time="2025-09-09T21:18:21.399717470Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\"" Sep 9 21:18:21.401348 containerd[1525]: time="2025-09-09T21:18:21.401307980Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 9 21:18:21.406714 containerd[1525]: time="2025-09-09T21:18:21.406673589Z" level=info msg="CreateContainer within sandbox \"5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 9 21:18:21.428250 containerd[1525]: time="2025-09-09T21:18:21.428107463Z" level=info msg="Container d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:21.446305 containerd[1525]: time="2025-09-09T21:18:21.446192078Z" level=info msg="CreateContainer within sandbox \"5ad3511e51c5a0b2f47d704b1ac5b6dd4c1e2c2bc71fc04e11f31be120b50ce0\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b\"" Sep 9 21:18:21.447008 containerd[1525]: time="2025-09-09T21:18:21.446962233Z" level=info msg="StartContainer for \"d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b\"" Sep 9 21:18:21.451766 containerd[1525]: time="2025-09-09T21:18:21.451731525Z" level=info msg="connecting to shim d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b" address="unix:///run/containerd/s/38873af70b5bf5d47eec87069796b17c7ca083ad07bc514551eb44de3c90496a" protocol=ttrpc version=3 Sep 9 21:18:21.490998 systemd[1]: Started cri-containerd-d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b.scope - libcontainer container d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b. Sep 9 21:18:21.593044 containerd[1525]: time="2025-09-09T21:18:21.592981898Z" level=info msg="StartContainer for \"d664bcc9d86bb48bb839ebc30ec475e43c47f88145b7f96c64a94c5018bbf08b\" returns successfully" Sep 9 21:18:21.596557 kubelet[2660]: E0909 21:18:21.596528 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:21.618360 sshd[5234]: Connection closed by 10.0.0.1 port 42842 Sep 9 21:18:21.618632 sshd-session[5214]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:21.624092 systemd[1]: sshd@8-10.0.0.66:22-10.0.0.1:42842.service: Deactivated successfully. Sep 9 21:18:21.627335 systemd[1]: session-9.scope: Deactivated successfully. Sep 9 21:18:21.628635 systemd-logind[1509]: Session 9 logged out. Waiting for processes to exit. Sep 9 21:18:21.631154 systemd-logind[1509]: Removed session 9. Sep 9 21:18:21.713516 containerd[1525]: time="2025-09-09T21:18:21.713408633Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec\" id:\"8af89effaba82ac8b3653634093ee7c751003c626a06bac22b448e4d776c69a8\" pid:5315 exit_status:1 exited_at:{seconds:1757452701 nanos:713069235}" Sep 9 21:18:21.910824 systemd-networkd[1433]: vxlan.calico: Link UP Sep 9 21:18:21.910838 systemd-networkd[1433]: vxlan.calico: Gained carrier Sep 9 21:18:22.398665 kubelet[2660]: I0909 21:18:22.398634 2660 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 9 21:18:22.398665 kubelet[2660]: I0909 21:18:22.398668 2660 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 9 21:18:22.614615 kubelet[2660]: I0909 21:18:22.614375 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-q45ms" podStartSLOduration=23.178734057 podStartE2EDuration="28.614359274s" podCreationTimestamp="2025-09-09 21:17:54 +0000 UTC" firstStartedPulling="2025-09-09 21:18:15.964952208 +0000 UTC m=+40.717508127" lastFinishedPulling="2025-09-09 21:18:21.400577425 +0000 UTC m=+46.153133344" observedRunningTime="2025-09-09 21:18:22.613696758 +0000 UTC m=+47.366252717" watchObservedRunningTime="2025-09-09 21:18:22.614359274 +0000 UTC m=+47.366915193" Sep 9 21:18:23.005860 systemd-networkd[1433]: vxlan.calico: Gained IPv6LL Sep 9 21:18:23.527048 containerd[1525]: time="2025-09-09T21:18:23.526993102Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:23.527731 containerd[1525]: time="2025-09-09T21:18:23.527679139Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=48134957" Sep 9 21:18:23.528551 containerd[1525]: time="2025-09-09T21:18:23.528510814Z" level=info msg="ImageCreate event name:\"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:23.530396 containerd[1525]: time="2025-09-09T21:18:23.530363964Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 21:18:23.531137 containerd[1525]: time="2025-09-09T21:18:23.531109919Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"49504166\" in 2.129764259s" Sep 9 21:18:23.531232 containerd[1525]: time="2025-09-09T21:18:23.531142079Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\"" Sep 9 21:18:23.543756 containerd[1525]: time="2025-09-09T21:18:23.543724448Z" level=info msg="CreateContainer within sandbox \"cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 9 21:18:23.551181 containerd[1525]: time="2025-09-09T21:18:23.550959608Z" level=info msg="Container 4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79: CDI devices from CRI Config.CDIDevices: []" Sep 9 21:18:23.560091 containerd[1525]: time="2025-09-09T21:18:23.560038517Z" level=info msg="CreateContainer within sandbox \"cda152288791d676be4563ef744fb844882df1711ecf7393297b4bb60ec7bd04\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79\"" Sep 9 21:18:23.560606 containerd[1525]: time="2025-09-09T21:18:23.560581154Z" level=info msg="StartContainer for \"4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79\"" Sep 9 21:18:23.561655 containerd[1525]: time="2025-09-09T21:18:23.561625188Z" level=info msg="connecting to shim 4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79" address="unix:///run/containerd/s/0ac2d771646c83e48756dd07d785fc72470f9a189c25d02043dbc11c4c0ab0d5" protocol=ttrpc version=3 Sep 9 21:18:23.585485 systemd[1]: Started cri-containerd-4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79.scope - libcontainer container 4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79. Sep 9 21:18:23.729240 containerd[1525]: time="2025-09-09T21:18:23.729203446Z" level=info msg="StartContainer for \"4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79\" returns successfully" Sep 9 21:18:24.617805 kubelet[2660]: I0909 21:18:24.617682 2660 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6f9758d6db-9jkm7" podStartSLOduration=24.715209094 podStartE2EDuration="30.617662161s" podCreationTimestamp="2025-09-09 21:17:54 +0000 UTC" firstStartedPulling="2025-09-09 21:18:17.629401888 +0000 UTC m=+42.381957807" lastFinishedPulling="2025-09-09 21:18:23.531854955 +0000 UTC m=+48.284410874" observedRunningTime="2025-09-09 21:18:24.615918931 +0000 UTC m=+49.368474850" watchObservedRunningTime="2025-09-09 21:18:24.617662161 +0000 UTC m=+49.370218120" Sep 9 21:18:24.640631 containerd[1525]: time="2025-09-09T21:18:24.640579875Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79\" id:\"0e704abaf0b37e27198a6d52d5dea9151cc4f6d5ae5969df1eeeaa13a022fd90\" pid:5485 exited_at:{seconds:1757452704 nanos:640157357}" Sep 9 21:18:26.633179 systemd[1]: Started sshd@9-10.0.0.66:22-10.0.0.1:42848.service - OpenSSH per-connection server daemon (10.0.0.1:42848). Sep 9 21:18:26.685446 sshd[5501]: Accepted publickey for core from 10.0.0.1 port 42848 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:26.686744 sshd-session[5501]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:26.690831 systemd-logind[1509]: New session 10 of user core. Sep 9 21:18:26.697922 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 9 21:18:26.851466 sshd[5504]: Connection closed by 10.0.0.1 port 42848 Sep 9 21:18:26.851955 sshd-session[5501]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:26.865250 systemd[1]: sshd@9-10.0.0.66:22-10.0.0.1:42848.service: Deactivated successfully. Sep 9 21:18:26.868334 systemd[1]: session-10.scope: Deactivated successfully. Sep 9 21:18:26.868975 systemd-logind[1509]: Session 10 logged out. Waiting for processes to exit. Sep 9 21:18:26.871158 systemd[1]: Started sshd@10-10.0.0.66:22-10.0.0.1:42854.service - OpenSSH per-connection server daemon (10.0.0.1:42854). Sep 9 21:18:26.871915 systemd-logind[1509]: Removed session 10. Sep 9 21:18:26.930586 sshd[5519]: Accepted publickey for core from 10.0.0.1 port 42854 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:26.931828 sshd-session[5519]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:26.935683 systemd-logind[1509]: New session 11 of user core. Sep 9 21:18:26.944931 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 9 21:18:27.122133 sshd[5522]: Connection closed by 10.0.0.1 port 42854 Sep 9 21:18:27.121244 sshd-session[5519]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:27.134859 systemd[1]: sshd@10-10.0.0.66:22-10.0.0.1:42854.service: Deactivated successfully. Sep 9 21:18:27.139233 systemd[1]: session-11.scope: Deactivated successfully. Sep 9 21:18:27.140830 systemd-logind[1509]: Session 11 logged out. Waiting for processes to exit. Sep 9 21:18:27.146742 systemd[1]: Started sshd@11-10.0.0.66:22-10.0.0.1:42870.service - OpenSSH per-connection server daemon (10.0.0.1:42870). Sep 9 21:18:27.148098 systemd-logind[1509]: Removed session 11. Sep 9 21:18:27.203994 sshd[5535]: Accepted publickey for core from 10.0.0.1 port 42870 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:27.206295 sshd-session[5535]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:27.212566 systemd-logind[1509]: New session 12 of user core. Sep 9 21:18:27.225973 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 9 21:18:27.370733 sshd[5544]: Connection closed by 10.0.0.1 port 42870 Sep 9 21:18:27.371167 sshd-session[5535]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:27.374551 systemd-logind[1509]: Session 12 logged out. Waiting for processes to exit. Sep 9 21:18:27.374719 systemd[1]: sshd@11-10.0.0.66:22-10.0.0.1:42870.service: Deactivated successfully. Sep 9 21:18:27.376462 systemd[1]: session-12.scope: Deactivated successfully. Sep 9 21:18:27.379092 systemd-logind[1509]: Removed session 12. Sep 9 21:18:32.389069 systemd[1]: Started sshd@12-10.0.0.66:22-10.0.0.1:43522.service - OpenSSH per-connection server daemon (10.0.0.1:43522). Sep 9 21:18:32.430752 sshd[5571]: Accepted publickey for core from 10.0.0.1 port 43522 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:32.431854 sshd-session[5571]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:32.435301 systemd-logind[1509]: New session 13 of user core. Sep 9 21:18:32.449940 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 9 21:18:32.570259 sshd[5574]: Connection closed by 10.0.0.1 port 43522 Sep 9 21:18:32.570497 sshd-session[5571]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:32.574136 systemd[1]: sshd@12-10.0.0.66:22-10.0.0.1:43522.service: Deactivated successfully. Sep 9 21:18:32.577251 systemd[1]: session-13.scope: Deactivated successfully. Sep 9 21:18:32.577974 systemd-logind[1509]: Session 13 logged out. Waiting for processes to exit. Sep 9 21:18:32.578776 systemd-logind[1509]: Removed session 13. Sep 9 21:18:37.588271 systemd[1]: Started sshd@13-10.0.0.66:22-10.0.0.1:43524.service - OpenSSH per-connection server daemon (10.0.0.1:43524). Sep 9 21:18:37.635898 sshd[5590]: Accepted publickey for core from 10.0.0.1 port 43524 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:37.637129 sshd-session[5590]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:37.640705 systemd-logind[1509]: New session 14 of user core. Sep 9 21:18:37.649996 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 9 21:18:37.769831 sshd[5593]: Connection closed by 10.0.0.1 port 43524 Sep 9 21:18:37.769955 sshd-session[5590]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:37.773017 systemd-logind[1509]: Session 14 logged out. Waiting for processes to exit. Sep 9 21:18:37.773300 systemd[1]: sshd@13-10.0.0.66:22-10.0.0.1:43524.service: Deactivated successfully. Sep 9 21:18:37.774936 systemd[1]: session-14.scope: Deactivated successfully. Sep 9 21:18:37.776153 systemd-logind[1509]: Removed session 14. Sep 9 21:18:38.493946 containerd[1525]: time="2025-09-09T21:18:38.493901753Z" level=info msg="TaskExit event in podsandbox handler container_id:\"69a09a09b5616fc9fca1c83d4160a62ce172fee01a5ac64ec0233da538d3cb9a\" id:\"d88d72b26d76cddd28936554f632203cd74dcea6695ee4bfb5a206dfdcc23496\" pid:5618 exited_at:{seconds:1757452718 nanos:493561355}" Sep 9 21:18:42.786610 systemd[1]: Started sshd@14-10.0.0.66:22-10.0.0.1:44860.service - OpenSSH per-connection server daemon (10.0.0.1:44860). Sep 9 21:18:42.858506 sshd[5642]: Accepted publickey for core from 10.0.0.1 port 44860 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:42.859947 sshd-session[5642]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:42.865576 systemd-logind[1509]: New session 15 of user core. Sep 9 21:18:42.876985 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 9 21:18:43.027615 sshd[5645]: Connection closed by 10.0.0.1 port 44860 Sep 9 21:18:43.027976 sshd-session[5642]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:43.031351 systemd[1]: sshd@14-10.0.0.66:22-10.0.0.1:44860.service: Deactivated successfully. Sep 9 21:18:43.034435 systemd[1]: session-15.scope: Deactivated successfully. Sep 9 21:18:43.035296 systemd-logind[1509]: Session 15 logged out. Waiting for processes to exit. Sep 9 21:18:43.036446 systemd-logind[1509]: Removed session 15. Sep 9 21:18:44.532577 kubelet[2660]: I0909 21:18:44.532465 2660 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 9 21:18:48.040250 systemd[1]: Started sshd@15-10.0.0.66:22-10.0.0.1:44868.service - OpenSSH per-connection server daemon (10.0.0.1:44868). Sep 9 21:18:48.098659 sshd[5664]: Accepted publickey for core from 10.0.0.1 port 44868 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:48.099681 sshd-session[5664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:48.103857 systemd-logind[1509]: New session 16 of user core. Sep 9 21:18:48.109938 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 9 21:18:48.278583 sshd[5667]: Connection closed by 10.0.0.1 port 44868 Sep 9 21:18:48.280047 sshd-session[5664]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:48.288980 systemd[1]: sshd@15-10.0.0.66:22-10.0.0.1:44868.service: Deactivated successfully. Sep 9 21:18:48.291022 systemd[1]: session-16.scope: Deactivated successfully. Sep 9 21:18:48.294124 systemd-logind[1509]: Session 16 logged out. Waiting for processes to exit. Sep 9 21:18:48.296439 systemd[1]: Started sshd@16-10.0.0.66:22-10.0.0.1:44876.service - OpenSSH per-connection server daemon (10.0.0.1:44876). Sep 9 21:18:48.297747 systemd-logind[1509]: Removed session 16. Sep 9 21:18:48.325507 kubelet[2660]: E0909 21:18:48.325461 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:18:48.352436 sshd[5680]: Accepted publickey for core from 10.0.0.1 port 44876 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:48.354049 sshd-session[5680]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:48.358831 systemd-logind[1509]: New session 17 of user core. Sep 9 21:18:48.364928 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 9 21:18:48.555373 sshd[5683]: Connection closed by 10.0.0.1 port 44876 Sep 9 21:18:48.556761 sshd-session[5680]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:48.565814 systemd[1]: sshd@16-10.0.0.66:22-10.0.0.1:44876.service: Deactivated successfully. Sep 9 21:18:48.568503 systemd[1]: session-17.scope: Deactivated successfully. Sep 9 21:18:48.570266 systemd-logind[1509]: Session 17 logged out. Waiting for processes to exit. Sep 9 21:18:48.573757 systemd[1]: Started sshd@17-10.0.0.66:22-10.0.0.1:44878.service - OpenSSH per-connection server daemon (10.0.0.1:44878). Sep 9 21:18:48.575242 systemd-logind[1509]: Removed session 17. Sep 9 21:18:48.636616 sshd[5695]: Accepted publickey for core from 10.0.0.1 port 44878 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:48.637759 sshd-session[5695]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:48.641574 systemd-logind[1509]: New session 18 of user core. Sep 9 21:18:48.646934 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 9 21:18:49.257126 sshd[5698]: Connection closed by 10.0.0.1 port 44878 Sep 9 21:18:49.255910 sshd-session[5695]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:49.267618 systemd[1]: sshd@17-10.0.0.66:22-10.0.0.1:44878.service: Deactivated successfully. Sep 9 21:18:49.269196 systemd[1]: session-18.scope: Deactivated successfully. Sep 9 21:18:49.270495 systemd-logind[1509]: Session 18 logged out. Waiting for processes to exit. Sep 9 21:18:49.274660 systemd[1]: Started sshd@18-10.0.0.66:22-10.0.0.1:44892.service - OpenSSH per-connection server daemon (10.0.0.1:44892). Sep 9 21:18:49.276265 systemd-logind[1509]: Removed session 18. Sep 9 21:18:49.321954 sshd[5717]: Accepted publickey for core from 10.0.0.1 port 44892 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:49.323120 sshd-session[5717]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:49.327568 systemd-logind[1509]: New session 19 of user core. Sep 9 21:18:49.337020 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 9 21:18:49.665506 sshd[5721]: Connection closed by 10.0.0.1 port 44892 Sep 9 21:18:49.668659 sshd-session[5717]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:49.679745 systemd[1]: sshd@18-10.0.0.66:22-10.0.0.1:44892.service: Deactivated successfully. Sep 9 21:18:49.684472 systemd[1]: session-19.scope: Deactivated successfully. Sep 9 21:18:49.688995 systemd-logind[1509]: Session 19 logged out. Waiting for processes to exit. Sep 9 21:18:49.690953 systemd[1]: Started sshd@19-10.0.0.66:22-10.0.0.1:44904.service - OpenSSH per-connection server daemon (10.0.0.1:44904). Sep 9 21:18:49.696605 systemd-logind[1509]: Removed session 19. Sep 9 21:18:49.759772 sshd[5733]: Accepted publickey for core from 10.0.0.1 port 44904 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:49.761157 sshd-session[5733]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:49.767352 systemd-logind[1509]: New session 20 of user core. Sep 9 21:18:49.776982 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 9 21:18:49.928521 sshd[5736]: Connection closed by 10.0.0.1 port 44904 Sep 9 21:18:49.929420 sshd-session[5733]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:49.933502 systemd[1]: sshd@19-10.0.0.66:22-10.0.0.1:44904.service: Deactivated successfully. Sep 9 21:18:49.936677 systemd[1]: session-20.scope: Deactivated successfully. Sep 9 21:18:49.938251 systemd-logind[1509]: Session 20 logged out. Waiting for processes to exit. Sep 9 21:18:49.941164 systemd-logind[1509]: Removed session 20. Sep 9 21:18:51.704561 containerd[1525]: time="2025-09-09T21:18:51.704508789Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b37e4a2afd9a318d99006dcdf5984353b374596a4a1f2d294c2f92208c8858ec\" id:\"02cd1e281527ae5bd282756a8f0b4079cf5aa6ea01995d51cb0a32642b6680cc\" pid:5761 exited_at:{seconds:1757452731 nanos:704162705}" Sep 9 21:18:54.647694 containerd[1525]: time="2025-09-09T21:18:54.641464011Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4522e8bb53b5b1925442115fe6a0a63c81f6313885a72bd5e96e8198e154fd79\" id:\"b01819f656aaae778167b8119c30ba28b561f40581bfe7941ab8f341bdad7db8\" pid:5791 exited_at:{seconds:1757452734 nanos:641231409}" Sep 9 21:18:54.943992 systemd[1]: Started sshd@20-10.0.0.66:22-10.0.0.1:40366.service - OpenSSH per-connection server daemon (10.0.0.1:40366). Sep 9 21:18:55.004308 sshd[5802]: Accepted publickey for core from 10.0.0.1 port 40366 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:18:55.005710 sshd-session[5802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:18:55.009541 systemd-logind[1509]: New session 21 of user core. Sep 9 21:18:55.012929 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 9 21:18:55.168181 sshd[5805]: Connection closed by 10.0.0.1 port 40366 Sep 9 21:18:55.168685 sshd-session[5802]: pam_unix(sshd:session): session closed for user core Sep 9 21:18:55.172107 systemd[1]: sshd@20-10.0.0.66:22-10.0.0.1:40366.service: Deactivated successfully. Sep 9 21:18:55.174072 systemd[1]: session-21.scope: Deactivated successfully. Sep 9 21:18:55.174813 systemd-logind[1509]: Session 21 logged out. Waiting for processes to exit. Sep 9 21:18:55.175935 systemd-logind[1509]: Removed session 21. Sep 9 21:18:59.326398 kubelet[2660]: E0909 21:18:59.326358 2660 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 21:19:00.183188 systemd[1]: Started sshd@21-10.0.0.66:22-10.0.0.1:46078.service - OpenSSH per-connection server daemon (10.0.0.1:46078). Sep 9 21:19:00.249218 sshd[5818]: Accepted publickey for core from 10.0.0.1 port 46078 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:19:00.250445 sshd-session[5818]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:19:00.254452 systemd-logind[1509]: New session 22 of user core. Sep 9 21:19:00.267023 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 9 21:19:00.413221 sshd[5821]: Connection closed by 10.0.0.1 port 46078 Sep 9 21:19:00.413493 sshd-session[5818]: pam_unix(sshd:session): session closed for user core Sep 9 21:19:00.416923 systemd[1]: sshd@21-10.0.0.66:22-10.0.0.1:46078.service: Deactivated successfully. Sep 9 21:19:00.420329 systemd[1]: session-22.scope: Deactivated successfully. Sep 9 21:19:00.420925 systemd-logind[1509]: Session 22 logged out. Waiting for processes to exit. Sep 9 21:19:00.421891 systemd-logind[1509]: Removed session 22. Sep 9 21:19:05.428171 systemd[1]: Started sshd@22-10.0.0.66:22-10.0.0.1:46092.service - OpenSSH per-connection server daemon (10.0.0.1:46092). Sep 9 21:19:05.488871 sshd[5843]: Accepted publickey for core from 10.0.0.1 port 46092 ssh2: RSA SHA256:/os6YPp183JWsEVhW0evH0PAuBe7do22d4T7SoFOxUE Sep 9 21:19:05.489522 sshd-session[5843]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 21:19:05.493693 systemd-logind[1509]: New session 23 of user core. Sep 9 21:19:05.511002 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 9 21:19:05.752147 sshd[5846]: Connection closed by 10.0.0.1 port 46092 Sep 9 21:19:05.752413 sshd-session[5843]: pam_unix(sshd:session): session closed for user core Sep 9 21:19:05.759555 systemd-logind[1509]: Session 23 logged out. Waiting for processes to exit. Sep 9 21:19:05.759687 systemd[1]: sshd@22-10.0.0.66:22-10.0.0.1:46092.service: Deactivated successfully. Sep 9 21:19:05.763351 systemd[1]: session-23.scope: Deactivated successfully. Sep 9 21:19:05.766365 systemd-logind[1509]: Removed session 23.