Sep 11 23:30:41.768286 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Sep 11 23:30:41.768307 kernel: Linux version 6.12.46-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Thu Sep 11 22:16:14 -00 2025 Sep 11 23:30:41.768316 kernel: KASLR enabled Sep 11 23:30:41.768322 kernel: efi: EFI v2.7 by EDK II Sep 11 23:30:41.768327 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb832018 ACPI 2.0=0xdbfd0018 RNG=0xdbfd0a18 MEMRESERVE=0xdb838218 Sep 11 23:30:41.768333 kernel: random: crng init done Sep 11 23:30:41.768340 kernel: secureboot: Secure boot disabled Sep 11 23:30:41.768345 kernel: ACPI: Early table checksum verification disabled Sep 11 23:30:41.768351 kernel: ACPI: RSDP 0x00000000DBFD0018 000024 (v02 BOCHS ) Sep 11 23:30:41.768358 kernel: ACPI: XSDT 0x00000000DBFD0F18 000064 (v01 BOCHS BXPC 00000001 01000013) Sep 11 23:30:41.768364 kernel: ACPI: FACP 0x00000000DBFD0B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768370 kernel: ACPI: DSDT 0x00000000DBF0E018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768375 kernel: ACPI: APIC 0x00000000DBFD0C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768381 kernel: ACPI: PPTT 0x00000000DBFD0098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768388 kernel: ACPI: GTDT 0x00000000DBFD0818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768396 kernel: ACPI: MCFG 0x00000000DBFD0A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768402 kernel: ACPI: SPCR 0x00000000DBFD0918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768408 kernel: ACPI: DBG2 0x00000000DBFD0998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768414 kernel: ACPI: IORT 0x00000000DBFD0198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 23:30:41.768420 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Sep 11 23:30:41.768426 kernel: ACPI: Use ACPI SPCR as default console: No Sep 11 23:30:41.768440 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Sep 11 23:30:41.768447 kernel: NODE_DATA(0) allocated [mem 0xdc965a00-0xdc96cfff] Sep 11 23:30:41.768452 kernel: Zone ranges: Sep 11 23:30:41.768458 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Sep 11 23:30:41.768494 kernel: DMA32 empty Sep 11 23:30:41.768501 kernel: Normal empty Sep 11 23:30:41.768506 kernel: Device empty Sep 11 23:30:41.768512 kernel: Movable zone start for each node Sep 11 23:30:41.768518 kernel: Early memory node ranges Sep 11 23:30:41.768524 kernel: node 0: [mem 0x0000000040000000-0x00000000db81ffff] Sep 11 23:30:41.768530 kernel: node 0: [mem 0x00000000db820000-0x00000000db82ffff] Sep 11 23:30:41.768536 kernel: node 0: [mem 0x00000000db830000-0x00000000dc09ffff] Sep 11 23:30:41.768542 kernel: node 0: [mem 0x00000000dc0a0000-0x00000000dc2dffff] Sep 11 23:30:41.768548 kernel: node 0: [mem 0x00000000dc2e0000-0x00000000dc36ffff] Sep 11 23:30:41.768554 kernel: node 0: [mem 0x00000000dc370000-0x00000000dc45ffff] Sep 11 23:30:41.768560 kernel: node 0: [mem 0x00000000dc460000-0x00000000dc52ffff] Sep 11 23:30:41.768567 kernel: node 0: [mem 0x00000000dc530000-0x00000000dc5cffff] Sep 11 23:30:41.768574 kernel: node 0: [mem 0x00000000dc5d0000-0x00000000dce1ffff] Sep 11 23:30:41.768580 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Sep 11 23:30:41.768588 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Sep 11 23:30:41.768595 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Sep 11 23:30:41.768601 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Sep 11 23:30:41.768609 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Sep 11 23:30:41.768615 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Sep 11 23:30:41.768622 kernel: cma: Reserved 16 MiB at 0x00000000d8000000 on node -1 Sep 11 23:30:41.768628 kernel: psci: probing for conduit method from ACPI. Sep 11 23:30:41.768634 kernel: psci: PSCIv1.1 detected in firmware. Sep 11 23:30:41.768641 kernel: psci: Using standard PSCI v0.2 function IDs Sep 11 23:30:41.768647 kernel: psci: Trusted OS migration not required Sep 11 23:30:41.768653 kernel: psci: SMC Calling Convention v1.1 Sep 11 23:30:41.768660 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Sep 11 23:30:41.768666 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Sep 11 23:30:41.768674 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Sep 11 23:30:41.768681 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Sep 11 23:30:41.768687 kernel: Detected PIPT I-cache on CPU0 Sep 11 23:30:41.768693 kernel: CPU features: detected: GIC system register CPU interface Sep 11 23:30:41.768700 kernel: CPU features: detected: Spectre-v4 Sep 11 23:30:41.768706 kernel: CPU features: detected: Spectre-BHB Sep 11 23:30:41.768713 kernel: CPU features: kernel page table isolation forced ON by KASLR Sep 11 23:30:41.768719 kernel: CPU features: detected: Kernel page table isolation (KPTI) Sep 11 23:30:41.768726 kernel: CPU features: detected: ARM erratum 1418040 Sep 11 23:30:41.768732 kernel: CPU features: detected: SSBS not fully self-synchronizing Sep 11 23:30:41.768739 kernel: alternatives: applying boot alternatives Sep 11 23:30:41.768746 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=34cdae46b43e6281eb14909b07c5254135a938c8cecf4370cc2216c267809c7a Sep 11 23:30:41.768755 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 11 23:30:41.768762 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 11 23:30:41.768768 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 11 23:30:41.768775 kernel: Fallback order for Node 0: 0 Sep 11 23:30:41.768781 kernel: Built 1 zonelists, mobility grouping on. Total pages: 643072 Sep 11 23:30:41.768788 kernel: Policy zone: DMA Sep 11 23:30:41.768794 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 11 23:30:41.768801 kernel: software IO TLB: SWIOTLB bounce buffer size adjusted to 2MB Sep 11 23:30:41.768807 kernel: software IO TLB: area num 4. Sep 11 23:30:41.768813 kernel: software IO TLB: SWIOTLB bounce buffer size roundup to 4MB Sep 11 23:30:41.768820 kernel: software IO TLB: mapped [mem 0x00000000d7c00000-0x00000000d8000000] (4MB) Sep 11 23:30:41.768828 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 11 23:30:41.768834 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 11 23:30:41.768841 kernel: rcu: RCU event tracing is enabled. Sep 11 23:30:41.768848 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 11 23:30:41.768854 kernel: Trampoline variant of Tasks RCU enabled. Sep 11 23:30:41.768860 kernel: Tracing variant of Tasks RCU enabled. Sep 11 23:30:41.768867 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 11 23:30:41.768873 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 11 23:30:41.768880 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 23:30:41.768886 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 23:30:41.768893 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Sep 11 23:30:41.768900 kernel: GICv3: 256 SPIs implemented Sep 11 23:30:41.768907 kernel: GICv3: 0 Extended SPIs implemented Sep 11 23:30:41.768913 kernel: Root IRQ handler: gic_handle_irq Sep 11 23:30:41.768920 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Sep 11 23:30:41.768929 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Sep 11 23:30:41.768935 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Sep 11 23:30:41.768942 kernel: ITS [mem 0x08080000-0x0809ffff] Sep 11 23:30:41.768950 kernel: ITS@0x0000000008080000: allocated 8192 Devices @40110000 (indirect, esz 8, psz 64K, shr 1) Sep 11 23:30:41.768959 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @40120000 (flat, esz 8, psz 64K, shr 1) Sep 11 23:30:41.768969 kernel: GICv3: using LPI property table @0x0000000040130000 Sep 11 23:30:41.768976 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040140000 Sep 11 23:30:41.768983 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 11 23:30:41.768991 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 11 23:30:41.768997 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Sep 11 23:30:41.769004 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Sep 11 23:30:41.769010 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Sep 11 23:30:41.769017 kernel: arm-pv: using stolen time PV Sep 11 23:30:41.769024 kernel: Console: colour dummy device 80x25 Sep 11 23:30:41.769030 kernel: ACPI: Core revision 20240827 Sep 11 23:30:41.769037 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Sep 11 23:30:41.769044 kernel: pid_max: default: 32768 minimum: 301 Sep 11 23:30:41.769051 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 11 23:30:41.769058 kernel: landlock: Up and running. Sep 11 23:30:41.769065 kernel: SELinux: Initializing. Sep 11 23:30:41.769072 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 11 23:30:41.769078 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 11 23:30:41.769085 kernel: rcu: Hierarchical SRCU implementation. Sep 11 23:30:41.769092 kernel: rcu: Max phase no-delay instances is 400. Sep 11 23:30:41.769099 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 11 23:30:41.769105 kernel: Remapping and enabling EFI services. Sep 11 23:30:41.769112 kernel: smp: Bringing up secondary CPUs ... Sep 11 23:30:41.769125 kernel: Detected PIPT I-cache on CPU1 Sep 11 23:30:41.769132 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Sep 11 23:30:41.769139 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040150000 Sep 11 23:30:41.769148 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 11 23:30:41.769155 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Sep 11 23:30:41.769162 kernel: Detected PIPT I-cache on CPU2 Sep 11 23:30:41.769169 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Sep 11 23:30:41.769177 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040160000 Sep 11 23:30:41.769186 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 11 23:30:41.769193 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Sep 11 23:30:41.769200 kernel: Detected PIPT I-cache on CPU3 Sep 11 23:30:41.769208 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Sep 11 23:30:41.769215 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040170000 Sep 11 23:30:41.769222 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 11 23:30:41.769229 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Sep 11 23:30:41.769237 kernel: smp: Brought up 1 node, 4 CPUs Sep 11 23:30:41.769244 kernel: SMP: Total of 4 processors activated. Sep 11 23:30:41.769252 kernel: CPU: All CPU(s) started at EL1 Sep 11 23:30:41.769260 kernel: CPU features: detected: 32-bit EL0 Support Sep 11 23:30:41.769267 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Sep 11 23:30:41.769275 kernel: CPU features: detected: Common not Private translations Sep 11 23:30:41.769282 kernel: CPU features: detected: CRC32 instructions Sep 11 23:30:41.769289 kernel: CPU features: detected: Enhanced Virtualization Traps Sep 11 23:30:41.769296 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Sep 11 23:30:41.769303 kernel: CPU features: detected: LSE atomic instructions Sep 11 23:30:41.769311 kernel: CPU features: detected: Privileged Access Never Sep 11 23:30:41.769320 kernel: CPU features: detected: RAS Extension Support Sep 11 23:30:41.769327 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Sep 11 23:30:41.769342 kernel: alternatives: applying system-wide alternatives Sep 11 23:30:41.769357 kernel: CPU features: detected: Hardware dirty bit management on CPU0-3 Sep 11 23:30:41.769365 kernel: Memory: 2424544K/2572288K available (11136K kernel code, 2440K rwdata, 9068K rodata, 38912K init, 1038K bss, 125408K reserved, 16384K cma-reserved) Sep 11 23:30:41.769373 kernel: devtmpfs: initialized Sep 11 23:30:41.769380 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 11 23:30:41.769387 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 11 23:30:41.769394 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Sep 11 23:30:41.769402 kernel: 0 pages in range for non-PLT usage Sep 11 23:30:41.769410 kernel: 508576 pages in range for PLT usage Sep 11 23:30:41.769416 kernel: pinctrl core: initialized pinctrl subsystem Sep 11 23:30:41.769423 kernel: SMBIOS 3.0.0 present. Sep 11 23:30:41.769434 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Sep 11 23:30:41.769442 kernel: DMI: Memory slots populated: 1/1 Sep 11 23:30:41.769449 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 11 23:30:41.769456 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Sep 11 23:30:41.769518 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Sep 11 23:30:41.769529 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Sep 11 23:30:41.769536 kernel: audit: initializing netlink subsys (disabled) Sep 11 23:30:41.769543 kernel: audit: type=2000 audit(0.020:1): state=initialized audit_enabled=0 res=1 Sep 11 23:30:41.769550 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 11 23:30:41.769557 kernel: cpuidle: using governor menu Sep 11 23:30:41.769564 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Sep 11 23:30:41.769571 kernel: ASID allocator initialised with 32768 entries Sep 11 23:30:41.769578 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 11 23:30:41.769584 kernel: Serial: AMBA PL011 UART driver Sep 11 23:30:41.769593 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 11 23:30:41.769600 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Sep 11 23:30:41.769607 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Sep 11 23:30:41.769614 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Sep 11 23:30:41.769621 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 11 23:30:41.769628 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Sep 11 23:30:41.769635 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Sep 11 23:30:41.769642 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Sep 11 23:30:41.769649 kernel: ACPI: Added _OSI(Module Device) Sep 11 23:30:41.769657 kernel: ACPI: Added _OSI(Processor Device) Sep 11 23:30:41.769664 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 11 23:30:41.769671 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 11 23:30:41.769678 kernel: ACPI: Interpreter enabled Sep 11 23:30:41.769685 kernel: ACPI: Using GIC for interrupt routing Sep 11 23:30:41.769692 kernel: ACPI: MCFG table detected, 1 entries Sep 11 23:30:41.769700 kernel: ACPI: CPU0 has been hot-added Sep 11 23:30:41.769707 kernel: ACPI: CPU1 has been hot-added Sep 11 23:30:41.769714 kernel: ACPI: CPU2 has been hot-added Sep 11 23:30:41.769721 kernel: ACPI: CPU3 has been hot-added Sep 11 23:30:41.769730 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Sep 11 23:30:41.769737 kernel: printk: legacy console [ttyAMA0] enabled Sep 11 23:30:41.769744 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 11 23:30:41.769897 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 11 23:30:41.769972 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 11 23:30:41.770035 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 11 23:30:41.770093 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Sep 11 23:30:41.770153 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Sep 11 23:30:41.770162 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Sep 11 23:30:41.770169 kernel: PCI host bridge to bus 0000:00 Sep 11 23:30:41.770241 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Sep 11 23:30:41.770294 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Sep 11 23:30:41.770346 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Sep 11 23:30:41.770397 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 11 23:30:41.770516 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Sep 11 23:30:41.770604 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 11 23:30:41.770670 kernel: pci 0000:00:01.0: BAR 0 [io 0x0000-0x001f] Sep 11 23:30:41.770733 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff] Sep 11 23:30:41.770793 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Sep 11 23:30:41.770853 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Sep 11 23:30:41.770911 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff]: assigned Sep 11 23:30:41.770974 kernel: pci 0000:00:01.0: BAR 0 [io 0x1000-0x101f]: assigned Sep 11 23:30:41.771028 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Sep 11 23:30:41.771080 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Sep 11 23:30:41.771136 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Sep 11 23:30:41.771145 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Sep 11 23:30:41.771152 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Sep 11 23:30:41.771160 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Sep 11 23:30:41.771169 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Sep 11 23:30:41.771176 kernel: iommu: Default domain type: Translated Sep 11 23:30:41.771183 kernel: iommu: DMA domain TLB invalidation policy: strict mode Sep 11 23:30:41.771190 kernel: efivars: Registered efivars operations Sep 11 23:30:41.771209 kernel: vgaarb: loaded Sep 11 23:30:41.771216 kernel: clocksource: Switched to clocksource arch_sys_counter Sep 11 23:30:41.771223 kernel: VFS: Disk quotas dquot_6.6.0 Sep 11 23:30:41.771230 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 11 23:30:41.771237 kernel: pnp: PnP ACPI init Sep 11 23:30:41.771304 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Sep 11 23:30:41.771315 kernel: pnp: PnP ACPI: found 1 devices Sep 11 23:30:41.771322 kernel: NET: Registered PF_INET protocol family Sep 11 23:30:41.771329 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 11 23:30:41.771336 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 11 23:30:41.771343 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 11 23:30:41.771350 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 11 23:30:41.771357 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 11 23:30:41.771366 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 11 23:30:41.771373 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 11 23:30:41.771380 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 11 23:30:41.771387 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 11 23:30:41.771394 kernel: PCI: CLS 0 bytes, default 64 Sep 11 23:30:41.771401 kernel: kvm [1]: HYP mode not available Sep 11 23:30:41.771408 kernel: Initialise system trusted keyrings Sep 11 23:30:41.771415 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 11 23:30:41.771422 kernel: Key type asymmetric registered Sep 11 23:30:41.771438 kernel: Asymmetric key parser 'x509' registered Sep 11 23:30:41.771446 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 11 23:30:41.771453 kernel: io scheduler mq-deadline registered Sep 11 23:30:41.771481 kernel: io scheduler kyber registered Sep 11 23:30:41.771491 kernel: io scheduler bfq registered Sep 11 23:30:41.771499 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Sep 11 23:30:41.771506 kernel: ACPI: button: Power Button [PWRB] Sep 11 23:30:41.771514 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Sep 11 23:30:41.771588 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Sep 11 23:30:41.771602 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 11 23:30:41.771609 kernel: thunder_xcv, ver 1.0 Sep 11 23:30:41.771615 kernel: thunder_bgx, ver 1.0 Sep 11 23:30:41.771623 kernel: nicpf, ver 1.0 Sep 11 23:30:41.771629 kernel: nicvf, ver 1.0 Sep 11 23:30:41.771700 kernel: rtc-efi rtc-efi.0: registered as rtc0 Sep 11 23:30:41.771757 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-09-11T23:30:41 UTC (1757633441) Sep 11 23:30:41.771766 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 11 23:30:41.771776 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Sep 11 23:30:41.771784 kernel: watchdog: NMI not fully supported Sep 11 23:30:41.771791 kernel: watchdog: Hard watchdog permanently disabled Sep 11 23:30:41.771798 kernel: NET: Registered PF_INET6 protocol family Sep 11 23:30:41.771805 kernel: Segment Routing with IPv6 Sep 11 23:30:41.771812 kernel: In-situ OAM (IOAM) with IPv6 Sep 11 23:30:41.771819 kernel: NET: Registered PF_PACKET protocol family Sep 11 23:30:41.771826 kernel: Key type dns_resolver registered Sep 11 23:30:41.771833 kernel: registered taskstats version 1 Sep 11 23:30:41.771840 kernel: Loading compiled-in X.509 certificates Sep 11 23:30:41.771849 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.46-flatcar: c76a2532dfc607285c10ef525f008171185de1e8' Sep 11 23:30:41.771856 kernel: Demotion targets for Node 0: null Sep 11 23:30:41.771863 kernel: Key type .fscrypt registered Sep 11 23:30:41.771870 kernel: Key type fscrypt-provisioning registered Sep 11 23:30:41.771877 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 11 23:30:41.771884 kernel: ima: Allocated hash algorithm: sha1 Sep 11 23:30:41.771891 kernel: ima: No architecture policies found Sep 11 23:30:41.771898 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Sep 11 23:30:41.771906 kernel: clk: Disabling unused clocks Sep 11 23:30:41.771913 kernel: PM: genpd: Disabling unused power domains Sep 11 23:30:41.771920 kernel: Warning: unable to open an initial console. Sep 11 23:30:41.771927 kernel: Freeing unused kernel memory: 38912K Sep 11 23:30:41.771934 kernel: Run /init as init process Sep 11 23:30:41.771941 kernel: with arguments: Sep 11 23:30:41.771948 kernel: /init Sep 11 23:30:41.771955 kernel: with environment: Sep 11 23:30:41.771962 kernel: HOME=/ Sep 11 23:30:41.771970 kernel: TERM=linux Sep 11 23:30:41.771977 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 11 23:30:41.771985 systemd[1]: Successfully made /usr/ read-only. Sep 11 23:30:41.771995 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 11 23:30:41.772003 systemd[1]: Detected virtualization kvm. Sep 11 23:30:41.772010 systemd[1]: Detected architecture arm64. Sep 11 23:30:41.772017 systemd[1]: Running in initrd. Sep 11 23:30:41.772025 systemd[1]: No hostname configured, using default hostname. Sep 11 23:30:41.772034 systemd[1]: Hostname set to . Sep 11 23:30:41.772041 systemd[1]: Initializing machine ID from VM UUID. Sep 11 23:30:41.772049 systemd[1]: Queued start job for default target initrd.target. Sep 11 23:30:41.772056 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 23:30:41.772063 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 23:30:41.772072 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 11 23:30:41.772080 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 11 23:30:41.772088 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 11 23:30:41.772098 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 11 23:30:41.772106 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 11 23:30:41.772114 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 11 23:30:41.772122 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 23:30:41.772129 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 11 23:30:41.772137 systemd[1]: Reached target paths.target - Path Units. Sep 11 23:30:41.772145 systemd[1]: Reached target slices.target - Slice Units. Sep 11 23:30:41.772153 systemd[1]: Reached target swap.target - Swaps. Sep 11 23:30:41.772160 systemd[1]: Reached target timers.target - Timer Units. Sep 11 23:30:41.772168 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 11 23:30:41.772176 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 11 23:30:41.772183 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 11 23:30:41.772191 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 11 23:30:41.772198 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 11 23:30:41.772206 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 11 23:30:41.772215 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 23:30:41.772222 systemd[1]: Reached target sockets.target - Socket Units. Sep 11 23:30:41.772230 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 11 23:30:41.772237 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 11 23:30:41.772245 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 11 23:30:41.772253 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 11 23:30:41.772261 systemd[1]: Starting systemd-fsck-usr.service... Sep 11 23:30:41.772268 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 11 23:30:41.772276 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 11 23:30:41.772285 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 23:30:41.772292 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 11 23:30:41.772300 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 23:30:41.772308 systemd[1]: Finished systemd-fsck-usr.service. Sep 11 23:30:41.772333 systemd-journald[245]: Collecting audit messages is disabled. Sep 11 23:30:41.772353 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 11 23:30:41.772361 systemd-journald[245]: Journal started Sep 11 23:30:41.772382 systemd-journald[245]: Runtime Journal (/run/log/journal/5a88db8a42644295a0837de9915bbf06) is 6M, max 48.5M, 42.4M free. Sep 11 23:30:41.766051 systemd-modules-load[248]: Inserted module 'overlay' Sep 11 23:30:41.774024 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 23:30:41.778067 systemd[1]: Started systemd-journald.service - Journal Service. Sep 11 23:30:41.780491 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 11 23:30:41.780995 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 11 23:30:41.782442 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 11 23:30:41.787654 kernel: Bridge firewalling registered Sep 11 23:30:41.784037 systemd-modules-load[248]: Inserted module 'br_netfilter' Sep 11 23:30:41.785344 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 11 23:30:41.787596 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 11 23:30:41.790664 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 11 23:30:41.793746 systemd-tmpfiles[263]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 11 23:30:41.795395 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 11 23:30:41.797789 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 23:30:41.800450 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 11 23:30:41.806507 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 11 23:30:41.808317 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 11 23:30:41.810447 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 23:30:41.813600 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 11 23:30:41.836635 systemd-resolved[288]: Positive Trust Anchors: Sep 11 23:30:41.836660 systemd-resolved[288]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 11 23:30:41.836691 systemd-resolved[288]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 11 23:30:41.841499 systemd-resolved[288]: Defaulting to hostname 'linux'. Sep 11 23:30:41.845388 dracut-cmdline[291]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=34cdae46b43e6281eb14909b07c5254135a938c8cecf4370cc2216c267809c7a Sep 11 23:30:41.844262 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 11 23:30:41.850331 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 11 23:30:41.910505 kernel: SCSI subsystem initialized Sep 11 23:30:41.916523 kernel: Loading iSCSI transport class v2.0-870. Sep 11 23:30:41.924498 kernel: iscsi: registered transport (tcp) Sep 11 23:30:41.937499 kernel: iscsi: registered transport (qla4xxx) Sep 11 23:30:41.937525 kernel: QLogic iSCSI HBA Driver Sep 11 23:30:41.954007 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 11 23:30:41.975504 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 23:30:41.977393 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 11 23:30:42.021536 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 11 23:30:42.023601 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 11 23:30:42.087498 kernel: raid6: neonx8 gen() 15752 MB/s Sep 11 23:30:42.104483 kernel: raid6: neonx4 gen() 15801 MB/s Sep 11 23:30:42.121485 kernel: raid6: neonx2 gen() 13204 MB/s Sep 11 23:30:42.138504 kernel: raid6: neonx1 gen() 10461 MB/s Sep 11 23:30:42.155482 kernel: raid6: int64x8 gen() 6899 MB/s Sep 11 23:30:42.172490 kernel: raid6: int64x4 gen() 7353 MB/s Sep 11 23:30:42.189477 kernel: raid6: int64x2 gen() 6108 MB/s Sep 11 23:30:42.206501 kernel: raid6: int64x1 gen() 5055 MB/s Sep 11 23:30:42.206534 kernel: raid6: using algorithm neonx4 gen() 15801 MB/s Sep 11 23:30:42.223508 kernel: raid6: .... xor() 12331 MB/s, rmw enabled Sep 11 23:30:42.223563 kernel: raid6: using neon recovery algorithm Sep 11 23:30:42.228561 kernel: xor: measuring software checksum speed Sep 11 23:30:42.228592 kernel: 8regs : 21596 MB/sec Sep 11 23:30:42.229671 kernel: 32regs : 21687 MB/sec Sep 11 23:30:42.229688 kernel: arm64_neon : 28080 MB/sec Sep 11 23:30:42.229697 kernel: xor: using function: arm64_neon (28080 MB/sec) Sep 11 23:30:42.281518 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 11 23:30:42.287869 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 11 23:30:42.290167 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 23:30:42.317981 systemd-udevd[500]: Using default interface naming scheme 'v255'. Sep 11 23:30:42.322248 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 23:30:42.324044 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 11 23:30:42.347601 dracut-pre-trigger[507]: rd.md=0: removing MD RAID activation Sep 11 23:30:42.370364 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 11 23:30:42.372519 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 11 23:30:42.435121 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 23:30:42.437718 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 11 23:30:42.498756 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Sep 11 23:30:42.498927 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 11 23:30:42.503153 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 23:30:42.506046 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 11 23:30:42.506063 kernel: GPT:9289727 != 19775487 Sep 11 23:30:42.506073 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 11 23:30:42.503223 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 23:30:42.509558 kernel: GPT:9289727 != 19775487 Sep 11 23:30:42.509575 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 11 23:30:42.509584 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 23:30:42.508606 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 23:30:42.511149 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 23:30:42.533683 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 23:30:42.545527 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 11 23:30:42.546698 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 11 23:30:42.554966 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 11 23:30:42.563518 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 11 23:30:42.569548 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 11 23:30:42.570415 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 11 23:30:42.573192 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 11 23:30:42.575420 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 23:30:42.577363 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 11 23:30:42.580013 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 11 23:30:42.581613 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 11 23:30:42.595516 disk-uuid[593]: Primary Header is updated. Sep 11 23:30:42.595516 disk-uuid[593]: Secondary Entries is updated. Sep 11 23:30:42.595516 disk-uuid[593]: Secondary Header is updated. Sep 11 23:30:42.598476 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 23:30:42.599939 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 11 23:30:43.637486 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 23:30:43.637543 disk-uuid[598]: The operation has completed successfully. Sep 11 23:30:43.660596 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 11 23:30:43.661510 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 11 23:30:43.687344 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 11 23:30:43.716384 sh[613]: Success Sep 11 23:30:43.728858 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 11 23:30:43.728909 kernel: device-mapper: uevent: version 1.0.3 Sep 11 23:30:43.728920 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 11 23:30:43.736500 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Sep 11 23:30:43.762250 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 11 23:30:43.765065 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 11 23:30:43.783874 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 11 23:30:43.789767 kernel: BTRFS: device fsid 070f11bc-6881-4580-bbfd-8e1bd2605f24 devid 1 transid 38 /dev/mapper/usr (253:0) scanned by mount (625) Sep 11 23:30:43.789799 kernel: BTRFS info (device dm-0): first mount of filesystem 070f11bc-6881-4580-bbfd-8e1bd2605f24 Sep 11 23:30:43.789810 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Sep 11 23:30:43.794483 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 11 23:30:43.794520 kernel: BTRFS info (device dm-0): enabling free space tree Sep 11 23:30:43.795174 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 11 23:30:43.796313 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 11 23:30:43.797406 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 11 23:30:43.798291 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 11 23:30:43.799734 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 11 23:30:43.824483 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (656) Sep 11 23:30:43.826260 kernel: BTRFS info (device vda6): first mount of filesystem 2cbf2c8e-1b28-4a7c-a6d6-f07090d47234 Sep 11 23:30:43.826294 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 11 23:30:43.828846 kernel: BTRFS info (device vda6): turning on async discard Sep 11 23:30:43.828879 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 23:30:43.833508 kernel: BTRFS info (device vda6): last unmount of filesystem 2cbf2c8e-1b28-4a7c-a6d6-f07090d47234 Sep 11 23:30:43.834678 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 11 23:30:43.836618 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 11 23:30:43.902372 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 11 23:30:43.905012 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 11 23:30:43.939735 systemd-networkd[804]: lo: Link UP Sep 11 23:30:43.939749 systemd-networkd[804]: lo: Gained carrier Sep 11 23:30:43.940415 systemd-networkd[804]: Enumeration completed Sep 11 23:30:43.940921 systemd-networkd[804]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 23:30:43.940925 systemd-networkd[804]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 11 23:30:43.941871 systemd-networkd[804]: eth0: Link UP Sep 11 23:30:43.941970 systemd-networkd[804]: eth0: Gained carrier Sep 11 23:30:43.941979 systemd-networkd[804]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 23:30:43.943890 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 11 23:30:43.944940 systemd[1]: Reached target network.target - Network. Sep 11 23:30:43.954193 ignition[699]: Ignition 2.21.0 Sep 11 23:30:43.954210 ignition[699]: Stage: fetch-offline Sep 11 23:30:43.954244 ignition[699]: no configs at "/usr/lib/ignition/base.d" Sep 11 23:30:43.954252 ignition[699]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 23:30:43.954414 ignition[699]: parsed url from cmdline: "" Sep 11 23:30:43.954417 ignition[699]: no config URL provided Sep 11 23:30:43.954432 ignition[699]: reading system config file "/usr/lib/ignition/user.ign" Sep 11 23:30:43.954440 ignition[699]: no config at "/usr/lib/ignition/user.ign" Sep 11 23:30:43.954472 ignition[699]: op(1): [started] loading QEMU firmware config module Sep 11 23:30:43.954478 ignition[699]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 11 23:30:43.961431 ignition[699]: op(1): [finished] loading QEMU firmware config module Sep 11 23:30:43.961527 systemd-networkd[804]: eth0: DHCPv4 address 10.0.0.19/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 11 23:30:44.009871 ignition[699]: parsing config with SHA512: efb3e719672efba3ef7b3f488fbe2a6ab4dd04018a83943cfec380ffb5ad787817203ed269098ba21a23e4381b50f07009870ca1fb16d5482fee754a80186d70 Sep 11 23:30:44.013880 unknown[699]: fetched base config from "system" Sep 11 23:30:44.013891 unknown[699]: fetched user config from "qemu" Sep 11 23:30:44.014245 ignition[699]: fetch-offline: fetch-offline passed Sep 11 23:30:44.016145 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 11 23:30:44.014302 ignition[699]: Ignition finished successfully Sep 11 23:30:44.017301 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 11 23:30:44.018134 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 11 23:30:44.057664 ignition[814]: Ignition 2.21.0 Sep 11 23:30:44.057678 ignition[814]: Stage: kargs Sep 11 23:30:44.057868 ignition[814]: no configs at "/usr/lib/ignition/base.d" Sep 11 23:30:44.057881 ignition[814]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 23:30:44.059313 ignition[814]: kargs: kargs passed Sep 11 23:30:44.059368 ignition[814]: Ignition finished successfully Sep 11 23:30:44.063155 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 11 23:30:44.065434 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 11 23:30:44.093790 ignition[823]: Ignition 2.21.0 Sep 11 23:30:44.093957 ignition[823]: Stage: disks Sep 11 23:30:44.096997 ignition[823]: no configs at "/usr/lib/ignition/base.d" Sep 11 23:30:44.097013 ignition[823]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 23:30:44.100201 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 11 23:30:44.098084 ignition[823]: disks: disks passed Sep 11 23:30:44.098131 ignition[823]: Ignition finished successfully Sep 11 23:30:44.103713 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 11 23:30:44.104840 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 11 23:30:44.106399 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 11 23:30:44.108092 systemd[1]: Reached target sysinit.target - System Initialization. Sep 11 23:30:44.109956 systemd[1]: Reached target basic.target - Basic System. Sep 11 23:30:44.116633 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 11 23:30:44.148302 systemd-fsck[833]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 11 23:30:44.153454 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 11 23:30:44.156823 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 11 23:30:44.231498 kernel: EXT4-fs (vda9): mounted filesystem 358f7642-1e9a-4460-bcb4-1ef3d420e352 r/w with ordered data mode. Quota mode: none. Sep 11 23:30:44.231867 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 11 23:30:44.232937 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 11 23:30:44.235890 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 11 23:30:44.237622 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 11 23:30:44.238627 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 11 23:30:44.238673 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 11 23:30:44.238697 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 11 23:30:44.258905 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 11 23:30:44.261597 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 11 23:30:44.267564 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (841) Sep 11 23:30:44.267595 kernel: BTRFS info (device vda6): first mount of filesystem 2cbf2c8e-1b28-4a7c-a6d6-f07090d47234 Sep 11 23:30:44.267605 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 11 23:30:44.270515 kernel: BTRFS info (device vda6): turning on async discard Sep 11 23:30:44.270549 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 23:30:44.271908 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 11 23:30:44.304402 initrd-setup-root[865]: cut: /sysroot/etc/passwd: No such file or directory Sep 11 23:30:44.308353 initrd-setup-root[872]: cut: /sysroot/etc/group: No such file or directory Sep 11 23:30:44.312204 initrd-setup-root[879]: cut: /sysroot/etc/shadow: No such file or directory Sep 11 23:30:44.316852 initrd-setup-root[886]: cut: /sysroot/etc/gshadow: No such file or directory Sep 11 23:30:44.397171 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 11 23:30:44.399220 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 11 23:30:44.400959 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 11 23:30:44.427805 kernel: BTRFS info (device vda6): last unmount of filesystem 2cbf2c8e-1b28-4a7c-a6d6-f07090d47234 Sep 11 23:30:44.440620 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 11 23:30:44.454067 ignition[954]: INFO : Ignition 2.21.0 Sep 11 23:30:44.454067 ignition[954]: INFO : Stage: mount Sep 11 23:30:44.455495 ignition[954]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 23:30:44.455495 ignition[954]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 23:30:44.458432 ignition[954]: INFO : mount: mount passed Sep 11 23:30:44.458432 ignition[954]: INFO : Ignition finished successfully Sep 11 23:30:44.457793 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 11 23:30:44.460493 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 11 23:30:44.788870 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 11 23:30:44.790481 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 11 23:30:44.820486 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (968) Sep 11 23:30:44.822150 kernel: BTRFS info (device vda6): first mount of filesystem 2cbf2c8e-1b28-4a7c-a6d6-f07090d47234 Sep 11 23:30:44.822170 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 11 23:30:44.824637 kernel: BTRFS info (device vda6): turning on async discard Sep 11 23:30:44.824681 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 23:30:44.826308 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 11 23:30:44.859135 ignition[985]: INFO : Ignition 2.21.0 Sep 11 23:30:44.860006 ignition[985]: INFO : Stage: files Sep 11 23:30:44.860006 ignition[985]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 23:30:44.860006 ignition[985]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 23:30:44.862534 ignition[985]: DEBUG : files: compiled without relabeling support, skipping Sep 11 23:30:44.863828 ignition[985]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 11 23:30:44.863828 ignition[985]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 11 23:30:44.868798 ignition[985]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 11 23:30:44.870168 ignition[985]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 11 23:30:44.870168 ignition[985]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 11 23:30:44.869456 unknown[985]: wrote ssh authorized keys file for user: core Sep 11 23:30:44.873929 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Sep 11 23:30:44.873929 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Sep 11 23:30:44.954133 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 11 23:30:45.010640 systemd-networkd[804]: eth0: Gained IPv6LL Sep 11 23:30:46.051953 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Sep 11 23:30:46.051953 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 11 23:30:46.056004 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Sep 11 23:30:46.070483 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Sep 11 23:30:46.070483 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Sep 11 23:30:46.070483 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-arm64.raw: attempt #1 Sep 11 23:30:46.547536 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 11 23:30:47.196539 ignition[985]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Sep 11 23:30:47.196539 ignition[985]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 11 23:30:47.199982 ignition[985]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 11 23:30:47.201711 ignition[985]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 11 23:30:47.216897 ignition[985]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 11 23:30:47.220606 ignition[985]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 11 23:30:47.222100 ignition[985]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 11 23:30:47.222100 ignition[985]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 11 23:30:47.222100 ignition[985]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 11 23:30:47.222100 ignition[985]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 11 23:30:47.222100 ignition[985]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 11 23:30:47.222100 ignition[985]: INFO : files: files passed Sep 11 23:30:47.222100 ignition[985]: INFO : Ignition finished successfully Sep 11 23:30:47.225502 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 11 23:30:47.227687 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 11 23:30:47.231598 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 11 23:30:47.253627 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 11 23:30:47.253721 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 11 23:30:47.258823 initrd-setup-root-after-ignition[1012]: grep: /sysroot/oem/oem-release: No such file or directory Sep 11 23:30:47.260005 initrd-setup-root-after-ignition[1016]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 11 23:30:47.260005 initrd-setup-root-after-ignition[1016]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 11 23:30:47.262921 initrd-setup-root-after-ignition[1020]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 11 23:30:47.262865 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 11 23:30:47.264277 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 11 23:30:47.266927 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 11 23:30:47.313104 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 11 23:30:47.313247 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 11 23:30:47.315241 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 11 23:30:47.316710 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 11 23:30:47.318118 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 11 23:30:47.318937 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 11 23:30:47.357061 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 11 23:30:47.359361 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 11 23:30:47.381274 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 11 23:30:47.382354 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 23:30:47.384176 systemd[1]: Stopped target timers.target - Timer Units. Sep 11 23:30:47.385679 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 11 23:30:47.385808 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 11 23:30:47.387821 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 11 23:30:47.389279 systemd[1]: Stopped target basic.target - Basic System. Sep 11 23:30:47.391828 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 11 23:30:47.393221 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 11 23:30:47.394817 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 11 23:30:47.396395 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 11 23:30:47.397995 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 11 23:30:47.399384 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 11 23:30:47.400971 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 11 23:30:47.402471 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 11 23:30:47.403875 systemd[1]: Stopped target swap.target - Swaps. Sep 11 23:30:47.405049 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 11 23:30:47.405179 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 11 23:30:47.406991 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 11 23:30:47.408407 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 23:30:47.409944 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 11 23:30:47.410653 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 23:30:47.411636 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 11 23:30:47.411748 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 11 23:30:47.414006 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 11 23:30:47.414123 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 11 23:30:47.415780 systemd[1]: Stopped target paths.target - Path Units. Sep 11 23:30:47.416975 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 11 23:30:47.417608 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 23:30:47.418670 systemd[1]: Stopped target slices.target - Slice Units. Sep 11 23:30:47.420146 systemd[1]: Stopped target sockets.target - Socket Units. Sep 11 23:30:47.421424 systemd[1]: iscsid.socket: Deactivated successfully. Sep 11 23:30:47.421526 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 11 23:30:47.422946 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 11 23:30:47.423021 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 11 23:30:47.424777 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 11 23:30:47.424898 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 11 23:30:47.426327 systemd[1]: ignition-files.service: Deactivated successfully. Sep 11 23:30:47.426435 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 11 23:30:47.428489 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 11 23:30:47.430184 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 11 23:30:47.431729 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 11 23:30:47.431840 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 23:30:47.433835 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 11 23:30:47.433932 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 11 23:30:47.439165 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 11 23:30:47.441561 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 11 23:30:47.452657 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 11 23:30:47.456295 ignition[1040]: INFO : Ignition 2.21.0 Sep 11 23:30:47.456295 ignition[1040]: INFO : Stage: umount Sep 11 23:30:47.458696 ignition[1040]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 23:30:47.458696 ignition[1040]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 23:30:47.458696 ignition[1040]: INFO : umount: umount passed Sep 11 23:30:47.458696 ignition[1040]: INFO : Ignition finished successfully Sep 11 23:30:47.457236 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 11 23:30:47.458512 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 11 23:30:47.459798 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 11 23:30:47.459953 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 11 23:30:47.462282 systemd[1]: Stopped target network.target - Network. Sep 11 23:30:47.464009 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 11 23:30:47.464083 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 11 23:30:47.465724 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 11 23:30:47.465772 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 11 23:30:47.467303 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 11 23:30:47.467352 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 11 23:30:47.469315 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 11 23:30:47.469357 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 11 23:30:47.470980 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 11 23:30:47.471032 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 11 23:30:47.472671 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 11 23:30:47.474202 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 11 23:30:47.483223 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 11 23:30:47.483347 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 11 23:30:47.488316 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 11 23:30:47.489705 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 11 23:30:47.489761 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 23:30:47.493362 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 11 23:30:47.493656 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 11 23:30:47.493753 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 11 23:30:47.496847 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 11 23:30:47.497362 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 11 23:30:47.498839 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 11 23:30:47.498876 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 11 23:30:47.501373 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 11 23:30:47.502988 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 11 23:30:47.503045 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 11 23:30:47.504781 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 11 23:30:47.504822 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 11 23:30:47.507376 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 11 23:30:47.507432 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 11 23:30:47.509360 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 23:30:47.513687 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 11 23:30:47.527420 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 11 23:30:47.530956 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 23:30:47.533910 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 11 23:30:47.533953 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 11 23:30:47.535688 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 11 23:30:47.535725 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 23:30:47.538293 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 11 23:30:47.538356 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 11 23:30:47.541108 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 11 23:30:47.541168 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 11 23:30:47.543646 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 11 23:30:47.543705 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 11 23:30:47.546871 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 11 23:30:47.547770 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 11 23:30:47.547827 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 23:30:47.550555 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 11 23:30:47.550600 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 23:30:47.553330 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 11 23:30:47.553376 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 11 23:30:47.556480 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 11 23:30:47.556522 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 23:30:47.558457 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 23:30:47.558507 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 23:30:47.562259 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 11 23:30:47.562373 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 11 23:30:47.564700 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 11 23:30:47.564811 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 11 23:30:47.569065 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 11 23:30:47.570990 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 11 23:30:47.588811 systemd[1]: Switching root. Sep 11 23:30:47.621993 systemd-journald[245]: Journal stopped Sep 11 23:30:48.346485 systemd-journald[245]: Received SIGTERM from PID 1 (systemd). Sep 11 23:30:48.346540 kernel: SELinux: policy capability network_peer_controls=1 Sep 11 23:30:48.346560 kernel: SELinux: policy capability open_perms=1 Sep 11 23:30:48.346572 kernel: SELinux: policy capability extended_socket_class=1 Sep 11 23:30:48.346581 kernel: SELinux: policy capability always_check_network=0 Sep 11 23:30:48.346590 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 11 23:30:48.346599 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 11 23:30:48.346608 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 11 23:30:48.346617 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 11 23:30:48.346626 kernel: SELinux: policy capability userspace_initial_context=0 Sep 11 23:30:48.346635 kernel: audit: type=1403 audit(1757633447.767:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 11 23:30:48.346645 systemd[1]: Successfully loaded SELinux policy in 52.133ms. Sep 11 23:30:48.346666 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 5.468ms. Sep 11 23:30:48.346678 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 11 23:30:48.346689 systemd[1]: Detected virtualization kvm. Sep 11 23:30:48.346699 systemd[1]: Detected architecture arm64. Sep 11 23:30:48.346708 systemd[1]: Detected first boot. Sep 11 23:30:48.346718 systemd[1]: Initializing machine ID from VM UUID. Sep 11 23:30:48.346728 zram_generator::config[1087]: No configuration found. Sep 11 23:30:48.346738 kernel: NET: Registered PF_VSOCK protocol family Sep 11 23:30:48.346750 systemd[1]: Populated /etc with preset unit settings. Sep 11 23:30:48.346761 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 11 23:30:48.346771 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 11 23:30:48.346781 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 11 23:30:48.346790 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 11 23:30:48.346801 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 11 23:30:48.346811 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 11 23:30:48.346821 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 11 23:30:48.346831 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 11 23:30:48.346843 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 11 23:30:48.346852 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 11 23:30:48.346862 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 11 23:30:48.346872 systemd[1]: Created slice user.slice - User and Session Slice. Sep 11 23:30:48.346882 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 23:30:48.346892 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 23:30:48.346902 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 11 23:30:48.346912 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 11 23:30:48.346924 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 11 23:30:48.346934 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 11 23:30:48.346943 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Sep 11 23:30:48.346953 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 23:30:48.346963 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 11 23:30:48.346973 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 11 23:30:48.346983 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 11 23:30:48.346993 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 11 23:30:48.347004 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 11 23:30:48.347014 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 23:30:48.347024 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 11 23:30:48.347033 systemd[1]: Reached target slices.target - Slice Units. Sep 11 23:30:48.347043 systemd[1]: Reached target swap.target - Swaps. Sep 11 23:30:48.347053 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 11 23:30:48.347063 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 11 23:30:48.347073 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 11 23:30:48.347083 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 11 23:30:48.347094 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 11 23:30:48.347104 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 23:30:48.347115 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 11 23:30:48.347125 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 11 23:30:48.347135 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 11 23:30:48.347144 systemd[1]: Mounting media.mount - External Media Directory... Sep 11 23:30:48.347154 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 11 23:30:48.347164 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 11 23:30:48.347174 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 11 23:30:48.347185 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 11 23:30:48.347195 systemd[1]: Reached target machines.target - Containers. Sep 11 23:30:48.347205 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 11 23:30:48.347215 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 23:30:48.347225 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 11 23:30:48.347235 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 11 23:30:48.347245 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 23:30:48.347255 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 11 23:30:48.347265 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 23:30:48.347276 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 11 23:30:48.347287 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 23:30:48.347296 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 11 23:30:48.347306 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 11 23:30:48.347316 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 11 23:30:48.347326 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 11 23:30:48.347340 systemd[1]: Stopped systemd-fsck-usr.service. Sep 11 23:30:48.347350 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 23:30:48.347362 kernel: fuse: init (API version 7.41) Sep 11 23:30:48.347371 kernel: ACPI: bus type drm_connector registered Sep 11 23:30:48.347380 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 11 23:30:48.347390 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 11 23:30:48.347400 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 11 23:30:48.347418 kernel: loop: module loaded Sep 11 23:30:48.347430 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 11 23:30:48.347440 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 11 23:30:48.347450 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 11 23:30:48.347469 systemd[1]: verity-setup.service: Deactivated successfully. Sep 11 23:30:48.347480 systemd[1]: Stopped verity-setup.service. Sep 11 23:30:48.347489 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 11 23:30:48.347499 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 11 23:30:48.347508 systemd[1]: Mounted media.mount - External Media Directory. Sep 11 23:30:48.347541 systemd-journald[1162]: Collecting audit messages is disabled. Sep 11 23:30:48.347562 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 11 23:30:48.347573 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 11 23:30:48.347583 systemd-journald[1162]: Journal started Sep 11 23:30:48.347603 systemd-journald[1162]: Runtime Journal (/run/log/journal/5a88db8a42644295a0837de9915bbf06) is 6M, max 48.5M, 42.4M free. Sep 11 23:30:48.136204 systemd[1]: Queued start job for default target multi-user.target. Sep 11 23:30:48.157498 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 11 23:30:48.157902 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 11 23:30:48.350519 systemd[1]: Started systemd-journald.service - Journal Service. Sep 11 23:30:48.350971 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 11 23:30:48.352019 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 11 23:30:48.353184 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 23:30:48.354481 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 11 23:30:48.354644 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 11 23:30:48.355749 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 23:30:48.355911 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 23:30:48.357028 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 11 23:30:48.357179 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 11 23:30:48.358287 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 23:30:48.358477 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 23:30:48.359624 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 11 23:30:48.359786 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 11 23:30:48.360824 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 23:30:48.360976 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 23:30:48.362282 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 11 23:30:48.363533 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 23:30:48.364710 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 11 23:30:48.366160 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 11 23:30:48.378077 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 11 23:30:48.380304 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 11 23:30:48.384655 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 11 23:30:48.385526 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 11 23:30:48.385563 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 11 23:30:48.387162 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 11 23:30:48.391603 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 11 23:30:48.392585 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 23:30:48.394115 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 11 23:30:48.395947 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 11 23:30:48.397113 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 11 23:30:48.398053 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 11 23:30:48.399810 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 11 23:30:48.403253 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 11 23:30:48.405590 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 11 23:30:48.408642 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 11 23:30:48.411439 systemd-journald[1162]: Time spent on flushing to /var/log/journal/5a88db8a42644295a0837de9915bbf06 is 11.232ms for 887 entries. Sep 11 23:30:48.411439 systemd-journald[1162]: System Journal (/var/log/journal/5a88db8a42644295a0837de9915bbf06) is 8M, max 195.6M, 187.6M free. Sep 11 23:30:48.439730 systemd-journald[1162]: Received client request to flush runtime journal. Sep 11 23:30:48.439773 kernel: loop0: detected capacity change from 0 to 119320 Sep 11 23:30:48.439786 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 11 23:30:48.411324 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 23:30:48.414283 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 11 23:30:48.416855 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 11 23:30:48.426305 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 11 23:30:48.427685 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 11 23:30:48.436432 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 11 23:30:48.437687 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 11 23:30:48.449871 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 11 23:30:48.454492 systemd-tmpfiles[1204]: ACLs are not supported, ignoring. Sep 11 23:30:48.454508 systemd-tmpfiles[1204]: ACLs are not supported, ignoring. Sep 11 23:30:48.457885 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 11 23:30:48.461658 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 11 23:30:48.462489 kernel: loop1: detected capacity change from 0 to 100600 Sep 11 23:30:48.475605 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 11 23:30:48.490513 kernel: loop2: detected capacity change from 0 to 211168 Sep 11 23:30:48.494692 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 11 23:30:48.498643 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 11 23:30:48.513497 kernel: loop3: detected capacity change from 0 to 119320 Sep 11 23:30:48.520491 kernel: loop4: detected capacity change from 0 to 100600 Sep 11 23:30:48.521054 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Sep 11 23:30:48.521320 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Sep 11 23:30:48.524925 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 23:30:48.528540 kernel: loop5: detected capacity change from 0 to 211168 Sep 11 23:30:48.532770 (sd-merge)[1226]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 11 23:30:48.533164 (sd-merge)[1226]: Merged extensions into '/usr'. Sep 11 23:30:48.536735 systemd[1]: Reload requested from client PID 1203 ('systemd-sysext') (unit systemd-sysext.service)... Sep 11 23:30:48.536751 systemd[1]: Reloading... Sep 11 23:30:48.597801 zram_generator::config[1254]: No configuration found. Sep 11 23:30:48.684421 ldconfig[1198]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 11 23:30:48.741972 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 11 23:30:48.742217 systemd[1]: Reloading finished in 205 ms. Sep 11 23:30:48.764213 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 11 23:30:48.766547 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 11 23:30:48.781731 systemd[1]: Starting ensure-sysext.service... Sep 11 23:30:48.783440 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 11 23:30:48.792918 systemd[1]: Reload requested from client PID 1289 ('systemctl') (unit ensure-sysext.service)... Sep 11 23:30:48.792935 systemd[1]: Reloading... Sep 11 23:30:48.796074 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 11 23:30:48.796106 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 11 23:30:48.796318 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 11 23:30:48.796527 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 11 23:30:48.797100 systemd-tmpfiles[1290]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 11 23:30:48.797297 systemd-tmpfiles[1290]: ACLs are not supported, ignoring. Sep 11 23:30:48.797337 systemd-tmpfiles[1290]: ACLs are not supported, ignoring. Sep 11 23:30:48.800502 systemd-tmpfiles[1290]: Detected autofs mount point /boot during canonicalization of boot. Sep 11 23:30:48.800510 systemd-tmpfiles[1290]: Skipping /boot Sep 11 23:30:48.806207 systemd-tmpfiles[1290]: Detected autofs mount point /boot during canonicalization of boot. Sep 11 23:30:48.806223 systemd-tmpfiles[1290]: Skipping /boot Sep 11 23:30:48.840747 zram_generator::config[1320]: No configuration found. Sep 11 23:30:48.967244 systemd[1]: Reloading finished in 174 ms. Sep 11 23:30:48.988488 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 11 23:30:48.993678 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 23:30:49.004494 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 11 23:30:49.006517 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 11 23:30:49.008416 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 11 23:30:49.014614 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 11 23:30:49.016943 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 23:30:49.020702 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 11 23:30:49.029030 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 11 23:30:49.033203 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 23:30:49.035656 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 23:30:49.037723 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 23:30:49.040017 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 23:30:49.041009 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 23:30:49.041128 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 23:30:49.043495 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 11 23:30:49.049823 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 23:30:49.050014 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 23:30:49.050149 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 23:30:49.052357 systemd-udevd[1358]: Using default interface naming scheme 'v255'. Sep 11 23:30:49.053384 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 11 23:30:49.055681 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 23:30:49.055848 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 23:30:49.057864 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 23:30:49.058014 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 23:30:49.067732 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 11 23:30:49.070021 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 23:30:49.070231 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 23:30:49.074001 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 11 23:30:49.081253 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 23:30:49.085552 systemd[1]: Finished ensure-sysext.service. Sep 11 23:30:49.087072 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 11 23:30:49.090477 augenrules[1398]: No rules Sep 11 23:30:49.094887 systemd[1]: audit-rules.service: Deactivated successfully. Sep 11 23:30:49.095113 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 11 23:30:49.097165 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 11 23:30:49.098853 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 23:30:49.100248 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 23:30:49.104804 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 11 23:30:49.109900 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 23:30:49.112195 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 23:30:49.112239 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 23:30:49.117534 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 11 23:30:49.122710 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 11 23:30:49.123559 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 11 23:30:49.124083 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 23:30:49.124226 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 23:30:49.126930 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 11 23:30:49.129533 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 11 23:30:49.132736 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 23:30:49.133535 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 23:30:49.145516 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 11 23:30:49.145603 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 11 23:30:49.146010 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Sep 11 23:30:49.191280 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 11 23:30:49.196636 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 11 23:30:49.218026 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 11 23:30:49.264586 systemd-resolved[1356]: Positive Trust Anchors: Sep 11 23:30:49.264605 systemd-resolved[1356]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 11 23:30:49.264637 systemd-resolved[1356]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 11 23:30:49.271566 systemd-networkd[1433]: lo: Link UP Sep 11 23:30:49.271576 systemd-networkd[1433]: lo: Gained carrier Sep 11 23:30:49.272354 systemd-networkd[1433]: Enumeration completed Sep 11 23:30:49.272823 systemd-networkd[1433]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 23:30:49.272833 systemd-networkd[1433]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 11 23:30:49.273170 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 11 23:30:49.273766 systemd-resolved[1356]: Defaulting to hostname 'linux'. Sep 11 23:30:49.275527 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 11 23:30:49.277355 systemd-networkd[1433]: eth0: Link UP Sep 11 23:30:49.277528 systemd-networkd[1433]: eth0: Gained carrier Sep 11 23:30:49.277550 systemd-networkd[1433]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 23:30:49.278667 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 11 23:30:49.279601 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 11 23:30:49.281585 systemd[1]: Reached target network.target - Network. Sep 11 23:30:49.282305 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 11 23:30:49.304541 systemd-networkd[1433]: eth0: DHCPv4 address 10.0.0.19/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 11 23:30:49.306847 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 11 23:30:49.308036 systemd[1]: Reached target sysinit.target - System Initialization. Sep 11 23:30:49.309017 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 11 23:30:49.309639 systemd-timesyncd[1434]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 11 23:30:49.309693 systemd-timesyncd[1434]: Initial clock synchronization to Thu 2025-09-11 23:30:49.527807 UTC. Sep 11 23:30:49.310450 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 11 23:30:49.311654 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 11 23:30:49.313622 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 11 23:30:49.313658 systemd[1]: Reached target paths.target - Path Units. Sep 11 23:30:49.314555 systemd[1]: Reached target time-set.target - System Time Set. Sep 11 23:30:49.315545 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 11 23:30:49.316564 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 11 23:30:49.317870 systemd[1]: Reached target timers.target - Timer Units. Sep 11 23:30:49.319319 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 11 23:30:49.323164 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 11 23:30:49.325825 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 11 23:30:49.326997 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 11 23:30:49.328076 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 11 23:30:49.331452 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 11 23:30:49.332560 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 11 23:30:49.335509 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 11 23:30:49.336630 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 11 23:30:49.344968 systemd[1]: Reached target sockets.target - Socket Units. Sep 11 23:30:49.345782 systemd[1]: Reached target basic.target - Basic System. Sep 11 23:30:49.346522 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 11 23:30:49.346554 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 11 23:30:49.347647 systemd[1]: Starting containerd.service - containerd container runtime... Sep 11 23:30:49.349296 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 11 23:30:49.351027 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 11 23:30:49.363632 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 11 23:30:49.365283 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 11 23:30:49.366097 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 11 23:30:49.367179 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 11 23:30:49.368962 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 11 23:30:49.369782 jq[1474]: false Sep 11 23:30:49.372928 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 11 23:30:49.375880 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 11 23:30:49.380527 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 11 23:30:49.383745 extend-filesystems[1475]: Found /dev/vda6 Sep 11 23:30:49.387587 extend-filesystems[1475]: Found /dev/vda9 Sep 11 23:30:49.383770 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 23:30:49.385674 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 11 23:30:49.386833 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 11 23:30:49.387674 systemd[1]: Starting update-engine.service - Update Engine... Sep 11 23:30:49.390344 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 11 23:30:49.391574 extend-filesystems[1475]: Checking size of /dev/vda9 Sep 11 23:30:49.397957 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 11 23:30:49.398490 jq[1495]: true Sep 11 23:30:49.399760 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 11 23:30:49.399977 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 11 23:30:49.400228 systemd[1]: motdgen.service: Deactivated successfully. Sep 11 23:30:49.401504 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 11 23:30:49.403940 extend-filesystems[1475]: Resized partition /dev/vda9 Sep 11 23:30:49.404046 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 11 23:30:49.404261 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 11 23:30:49.408972 extend-filesystems[1504]: resize2fs 1.47.2 (1-Jan-2025) Sep 11 23:30:49.418686 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 11 23:30:49.424507 update_engine[1492]: I20250911 23:30:49.422938 1492 main.cc:92] Flatcar Update Engine starting Sep 11 23:30:49.440581 jq[1506]: true Sep 11 23:30:49.442018 (ntainerd)[1515]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 11 23:30:49.449326 tar[1503]: linux-arm64/LICENSE Sep 11 23:30:49.450315 tar[1503]: linux-arm64/helm Sep 11 23:30:49.452057 dbus-daemon[1472]: [system] SELinux support is enabled Sep 11 23:30:49.452248 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 11 23:30:49.456431 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 11 23:30:49.456509 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 11 23:30:49.457700 update_engine[1492]: I20250911 23:30:49.457638 1492 update_check_scheduler.cc:74] Next update check in 11m1s Sep 11 23:30:49.457779 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 11 23:30:49.457804 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 11 23:30:49.459756 systemd[1]: Started update-engine.service - Update Engine. Sep 11 23:30:49.468545 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 11 23:30:49.470640 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 11 23:30:49.485688 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 23:30:49.487581 extend-filesystems[1504]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 11 23:30:49.487581 extend-filesystems[1504]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 11 23:30:49.487581 extend-filesystems[1504]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 11 23:30:49.493640 extend-filesystems[1475]: Resized filesystem in /dev/vda9 Sep 11 23:30:49.489603 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 11 23:30:49.491773 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 11 23:30:49.501567 systemd-logind[1485]: Watching system buttons on /dev/input/event0 (Power Button) Sep 11 23:30:49.503176 systemd-logind[1485]: New seat seat0. Sep 11 23:30:49.508369 systemd[1]: Started systemd-logind.service - User Login Management. Sep 11 23:30:49.517985 bash[1541]: Updated "/home/core/.ssh/authorized_keys" Sep 11 23:30:49.522538 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 11 23:30:49.524987 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 11 23:30:49.552948 locksmithd[1523]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 11 23:30:49.613549 containerd[1515]: time="2025-09-11T23:30:49Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 11 23:30:49.614335 containerd[1515]: time="2025-09-11T23:30:49.614298720Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 11 23:30:49.627128 containerd[1515]: time="2025-09-11T23:30:49.627073240Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="10.88µs" Sep 11 23:30:49.627128 containerd[1515]: time="2025-09-11T23:30:49.627117440Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 11 23:30:49.627128 containerd[1515]: time="2025-09-11T23:30:49.627137120Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 11 23:30:49.627315 containerd[1515]: time="2025-09-11T23:30:49.627292240Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 11 23:30:49.627365 containerd[1515]: time="2025-09-11T23:30:49.627313520Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 11 23:30:49.627365 containerd[1515]: time="2025-09-11T23:30:49.627337640Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627429 containerd[1515]: time="2025-09-11T23:30:49.627398040Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627518 containerd[1515]: time="2025-09-11T23:30:49.627495320Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627842 containerd[1515]: time="2025-09-11T23:30:49.627812400Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627842 containerd[1515]: time="2025-09-11T23:30:49.627838200Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627899 containerd[1515]: time="2025-09-11T23:30:49.627855320Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627899 containerd[1515]: time="2025-09-11T23:30:49.627863400Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 11 23:30:49.627978 containerd[1515]: time="2025-09-11T23:30:49.627946680Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 11 23:30:49.628229 containerd[1515]: time="2025-09-11T23:30:49.628201320Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 11 23:30:49.628267 containerd[1515]: time="2025-09-11T23:30:49.628244120Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 11 23:30:49.628267 containerd[1515]: time="2025-09-11T23:30:49.628257720Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 11 23:30:49.628309 containerd[1515]: time="2025-09-11T23:30:49.628289280Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 11 23:30:49.628739 containerd[1515]: time="2025-09-11T23:30:49.628715600Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 11 23:30:49.628825 containerd[1515]: time="2025-09-11T23:30:49.628800720Z" level=info msg="metadata content store policy set" policy=shared Sep 11 23:30:49.632378 containerd[1515]: time="2025-09-11T23:30:49.632339400Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 11 23:30:49.632444 containerd[1515]: time="2025-09-11T23:30:49.632416400Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 11 23:30:49.632444 containerd[1515]: time="2025-09-11T23:30:49.632431680Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 11 23:30:49.632523 containerd[1515]: time="2025-09-11T23:30:49.632443160Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 11 23:30:49.632523 containerd[1515]: time="2025-09-11T23:30:49.632455840Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 11 23:30:49.632523 containerd[1515]: time="2025-09-11T23:30:49.632499440Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 11 23:30:49.632523 containerd[1515]: time="2025-09-11T23:30:49.632514440Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 11 23:30:49.632588 containerd[1515]: time="2025-09-11T23:30:49.632526920Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 11 23:30:49.632588 containerd[1515]: time="2025-09-11T23:30:49.632539240Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 11 23:30:49.632588 containerd[1515]: time="2025-09-11T23:30:49.632549200Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 11 23:30:49.632588 containerd[1515]: time="2025-09-11T23:30:49.632558640Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 11 23:30:49.632588 containerd[1515]: time="2025-09-11T23:30:49.632570960Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 11 23:30:49.632719 containerd[1515]: time="2025-09-11T23:30:49.632697120Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 11 23:30:49.632745 containerd[1515]: time="2025-09-11T23:30:49.632724000Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 11 23:30:49.632763 containerd[1515]: time="2025-09-11T23:30:49.632745840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 11 23:30:49.632763 containerd[1515]: time="2025-09-11T23:30:49.632756720Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 11 23:30:49.632800 containerd[1515]: time="2025-09-11T23:30:49.632767320Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 11 23:30:49.632800 containerd[1515]: time="2025-09-11T23:30:49.632778640Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 11 23:30:49.632800 containerd[1515]: time="2025-09-11T23:30:49.632790080Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 11 23:30:49.632851 containerd[1515]: time="2025-09-11T23:30:49.632800680Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 11 23:30:49.632851 containerd[1515]: time="2025-09-11T23:30:49.632811840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 11 23:30:49.632851 containerd[1515]: time="2025-09-11T23:30:49.632821880Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 11 23:30:49.632851 containerd[1515]: time="2025-09-11T23:30:49.632832160Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 11 23:30:49.633031 containerd[1515]: time="2025-09-11T23:30:49.633012000Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 11 23:30:49.633061 containerd[1515]: time="2025-09-11T23:30:49.633030600Z" level=info msg="Start snapshots syncer" Sep 11 23:30:49.633080 containerd[1515]: time="2025-09-11T23:30:49.633060400Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 11 23:30:49.633924 containerd[1515]: time="2025-09-11T23:30:49.633592520Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 11 23:30:49.634262 containerd[1515]: time="2025-09-11T23:30:49.634225920Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 11 23:30:49.634556 containerd[1515]: time="2025-09-11T23:30:49.634523760Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 11 23:30:49.635006 containerd[1515]: time="2025-09-11T23:30:49.634970840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 11 23:30:49.635143 containerd[1515]: time="2025-09-11T23:30:49.635107920Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 11 23:30:49.635205 containerd[1515]: time="2025-09-11T23:30:49.635183760Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 11 23:30:49.635235 containerd[1515]: time="2025-09-11T23:30:49.635205520Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 11 23:30:49.635235 containerd[1515]: time="2025-09-11T23:30:49.635221240Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 11 23:30:49.635235 containerd[1515]: time="2025-09-11T23:30:49.635232120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 11 23:30:49.635284 containerd[1515]: time="2025-09-11T23:30:49.635242960Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 11 23:30:49.635284 containerd[1515]: time="2025-09-11T23:30:49.635271840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 11 23:30:49.635318 containerd[1515]: time="2025-09-11T23:30:49.635303760Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 11 23:30:49.635342 containerd[1515]: time="2025-09-11T23:30:49.635315840Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 11 23:30:49.635376 containerd[1515]: time="2025-09-11T23:30:49.635360760Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 11 23:30:49.635412 containerd[1515]: time="2025-09-11T23:30:49.635378400Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 11 23:30:49.635412 containerd[1515]: time="2025-09-11T23:30:49.635387440Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 11 23:30:49.635412 containerd[1515]: time="2025-09-11T23:30:49.635396560Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 11 23:30:49.635486 containerd[1515]: time="2025-09-11T23:30:49.635415040Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 11 23:30:49.635486 containerd[1515]: time="2025-09-11T23:30:49.635425560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 11 23:30:49.635486 containerd[1515]: time="2025-09-11T23:30:49.635436560Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 11 23:30:49.635551 containerd[1515]: time="2025-09-11T23:30:49.635533840Z" level=info msg="runtime interface created" Sep 11 23:30:49.635551 containerd[1515]: time="2025-09-11T23:30:49.635540360Z" level=info msg="created NRI interface" Sep 11 23:30:49.635551 containerd[1515]: time="2025-09-11T23:30:49.635548400Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 11 23:30:49.635598 containerd[1515]: time="2025-09-11T23:30:49.635562680Z" level=info msg="Connect containerd service" Sep 11 23:30:49.635598 containerd[1515]: time="2025-09-11T23:30:49.635591440Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 11 23:30:49.636988 containerd[1515]: time="2025-09-11T23:30:49.636949720Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 11 23:30:49.706932 containerd[1515]: time="2025-09-11T23:30:49.706867280Z" level=info msg="Start subscribing containerd event" Sep 11 23:30:49.707036 containerd[1515]: time="2025-09-11T23:30:49.706945520Z" level=info msg="Start recovering state" Sep 11 23:30:49.707056 containerd[1515]: time="2025-09-11T23:30:49.707037480Z" level=info msg="Start event monitor" Sep 11 23:30:49.707056 containerd[1515]: time="2025-09-11T23:30:49.707052440Z" level=info msg="Start cni network conf syncer for default" Sep 11 23:30:49.707089 containerd[1515]: time="2025-09-11T23:30:49.707059720Z" level=info msg="Start streaming server" Sep 11 23:30:49.707089 containerd[1515]: time="2025-09-11T23:30:49.707068360Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 11 23:30:49.707143 containerd[1515]: time="2025-09-11T23:30:49.707074680Z" level=info msg="runtime interface starting up..." Sep 11 23:30:49.707143 containerd[1515]: time="2025-09-11T23:30:49.707099400Z" level=info msg="starting plugins..." Sep 11 23:30:49.707143 containerd[1515]: time="2025-09-11T23:30:49.707115000Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 11 23:30:49.707190 containerd[1515]: time="2025-09-11T23:30:49.707149280Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 11 23:30:49.707211 containerd[1515]: time="2025-09-11T23:30:49.707196640Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 11 23:30:49.707347 systemd[1]: Started containerd.service - containerd container runtime. Sep 11 23:30:49.708978 containerd[1515]: time="2025-09-11T23:30:49.708775520Z" level=info msg="containerd successfully booted in 0.095674s" Sep 11 23:30:49.762196 tar[1503]: linux-arm64/README.md Sep 11 23:30:49.780681 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 11 23:30:49.794220 sshd_keygen[1500]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 11 23:30:49.813500 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 11 23:30:49.815837 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 11 23:30:49.834814 systemd[1]: issuegen.service: Deactivated successfully. Sep 11 23:30:49.835035 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 11 23:30:49.837359 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 11 23:30:49.863641 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 11 23:30:49.866005 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 11 23:30:49.867772 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Sep 11 23:30:49.868791 systemd[1]: Reached target getty.target - Login Prompts. Sep 11 23:30:50.642810 systemd-networkd[1433]: eth0: Gained IPv6LL Sep 11 23:30:50.645188 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 11 23:30:50.646711 systemd[1]: Reached target network-online.target - Network is Online. Sep 11 23:30:50.648781 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 11 23:30:50.650850 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:30:50.652742 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 11 23:30:50.675512 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 11 23:30:50.677570 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 11 23:30:50.679397 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 11 23:30:50.681757 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 11 23:30:51.210688 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:30:51.212063 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 11 23:30:51.213506 systemd[1]: Startup finished in 2.010s (kernel) + 6.165s (initrd) + 3.498s (userspace) = 11.674s. Sep 11 23:30:51.215691 (kubelet)[1612]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 23:30:51.580310 kubelet[1612]: E0911 23:30:51.580195 1612 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 23:30:51.582773 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 23:30:51.582909 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 23:30:51.583276 systemd[1]: kubelet.service: Consumed 762ms CPU time, 258.1M memory peak. Sep 11 23:30:54.915565 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 11 23:30:54.918522 systemd[1]: Started sshd@0-10.0.0.19:22-10.0.0.1:46530.service - OpenSSH per-connection server daemon (10.0.0.1:46530). Sep 11 23:30:54.997304 sshd[1625]: Accepted publickey for core from 10.0.0.1 port 46530 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:54.999210 sshd-session[1625]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:55.007739 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 11 23:30:55.010083 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 11 23:30:55.019190 systemd-logind[1485]: New session 1 of user core. Sep 11 23:30:55.029554 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 11 23:30:55.033743 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 11 23:30:55.054589 (systemd)[1630]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 11 23:30:55.056839 systemd-logind[1485]: New session c1 of user core. Sep 11 23:30:55.174006 systemd[1630]: Queued start job for default target default.target. Sep 11 23:30:55.183951 systemd[1630]: Created slice app.slice - User Application Slice. Sep 11 23:30:55.183984 systemd[1630]: Reached target paths.target - Paths. Sep 11 23:30:55.184024 systemd[1630]: Reached target timers.target - Timers. Sep 11 23:30:55.185307 systemd[1630]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 11 23:30:55.197676 systemd[1630]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 11 23:30:55.197810 systemd[1630]: Reached target sockets.target - Sockets. Sep 11 23:30:55.197853 systemd[1630]: Reached target basic.target - Basic System. Sep 11 23:30:55.197880 systemd[1630]: Reached target default.target - Main User Target. Sep 11 23:30:55.197904 systemd[1630]: Startup finished in 135ms. Sep 11 23:30:55.198319 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 11 23:30:55.207668 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 11 23:30:55.269734 systemd[1]: Started sshd@1-10.0.0.19:22-10.0.0.1:46538.service - OpenSSH per-connection server daemon (10.0.0.1:46538). Sep 11 23:30:55.319881 sshd[1641]: Accepted publickey for core from 10.0.0.1 port 46538 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:55.321253 sshd-session[1641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:55.326814 systemd-logind[1485]: New session 2 of user core. Sep 11 23:30:55.333679 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 11 23:30:55.387460 sshd[1644]: Connection closed by 10.0.0.1 port 46538 Sep 11 23:30:55.387889 sshd-session[1641]: pam_unix(sshd:session): session closed for user core Sep 11 23:30:55.397677 systemd[1]: sshd@1-10.0.0.19:22-10.0.0.1:46538.service: Deactivated successfully. Sep 11 23:30:55.400967 systemd[1]: session-2.scope: Deactivated successfully. Sep 11 23:30:55.402560 systemd-logind[1485]: Session 2 logged out. Waiting for processes to exit. Sep 11 23:30:55.404313 systemd[1]: Started sshd@2-10.0.0.19:22-10.0.0.1:46542.service - OpenSSH per-connection server daemon (10.0.0.1:46542). Sep 11 23:30:55.405893 systemd-logind[1485]: Removed session 2. Sep 11 23:30:55.454318 sshd[1650]: Accepted publickey for core from 10.0.0.1 port 46542 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:55.455625 sshd-session[1650]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:55.460539 systemd-logind[1485]: New session 3 of user core. Sep 11 23:30:55.469687 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 11 23:30:55.519297 sshd[1653]: Connection closed by 10.0.0.1 port 46542 Sep 11 23:30:55.519703 sshd-session[1650]: pam_unix(sshd:session): session closed for user core Sep 11 23:30:55.532698 systemd[1]: sshd@2-10.0.0.19:22-10.0.0.1:46542.service: Deactivated successfully. Sep 11 23:30:55.534870 systemd[1]: session-3.scope: Deactivated successfully. Sep 11 23:30:55.535556 systemd-logind[1485]: Session 3 logged out. Waiting for processes to exit. Sep 11 23:30:55.538557 systemd[1]: Started sshd@3-10.0.0.19:22-10.0.0.1:46548.service - OpenSSH per-connection server daemon (10.0.0.1:46548). Sep 11 23:30:55.539387 systemd-logind[1485]: Removed session 3. Sep 11 23:30:55.587913 sshd[1659]: Accepted publickey for core from 10.0.0.1 port 46548 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:55.589187 sshd-session[1659]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:55.593722 systemd-logind[1485]: New session 4 of user core. Sep 11 23:30:55.603681 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 11 23:30:55.655453 sshd[1662]: Connection closed by 10.0.0.1 port 46548 Sep 11 23:30:55.655781 sshd-session[1659]: pam_unix(sshd:session): session closed for user core Sep 11 23:30:55.666781 systemd[1]: sshd@3-10.0.0.19:22-10.0.0.1:46548.service: Deactivated successfully. Sep 11 23:30:55.669822 systemd[1]: session-4.scope: Deactivated successfully. Sep 11 23:30:55.671230 systemd-logind[1485]: Session 4 logged out. Waiting for processes to exit. Sep 11 23:30:55.672493 systemd[1]: Started sshd@4-10.0.0.19:22-10.0.0.1:46558.service - OpenSSH per-connection server daemon (10.0.0.1:46558). Sep 11 23:30:55.673325 systemd-logind[1485]: Removed session 4. Sep 11 23:30:55.722565 sshd[1668]: Accepted publickey for core from 10.0.0.1 port 46558 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:55.723898 sshd-session[1668]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:55.727586 systemd-logind[1485]: New session 5 of user core. Sep 11 23:30:55.746816 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 11 23:30:55.803733 sudo[1672]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 11 23:30:55.803992 sudo[1672]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 23:30:55.817445 sudo[1672]: pam_unix(sudo:session): session closed for user root Sep 11 23:30:55.819104 sshd[1671]: Connection closed by 10.0.0.1 port 46558 Sep 11 23:30:55.819662 sshd-session[1668]: pam_unix(sshd:session): session closed for user core Sep 11 23:30:55.830637 systemd[1]: sshd@4-10.0.0.19:22-10.0.0.1:46558.service: Deactivated successfully. Sep 11 23:30:55.833866 systemd[1]: session-5.scope: Deactivated successfully. Sep 11 23:30:55.834618 systemd-logind[1485]: Session 5 logged out. Waiting for processes to exit. Sep 11 23:30:55.837050 systemd[1]: Started sshd@5-10.0.0.19:22-10.0.0.1:46564.service - OpenSSH per-connection server daemon (10.0.0.1:46564). Sep 11 23:30:55.837726 systemd-logind[1485]: Removed session 5. Sep 11 23:30:55.887843 sshd[1678]: Accepted publickey for core from 10.0.0.1 port 46564 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:55.889267 sshd-session[1678]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:55.893680 systemd-logind[1485]: New session 6 of user core. Sep 11 23:30:55.902668 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 11 23:30:55.954371 sudo[1683]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 11 23:30:55.954665 sudo[1683]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 23:30:56.033063 sudo[1683]: pam_unix(sudo:session): session closed for user root Sep 11 23:30:56.038458 sudo[1682]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 11 23:30:56.038993 sudo[1682]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 23:30:56.048308 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 11 23:30:56.101732 augenrules[1705]: No rules Sep 11 23:30:56.102666 systemd[1]: audit-rules.service: Deactivated successfully. Sep 11 23:30:56.102895 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 11 23:30:56.103882 sudo[1682]: pam_unix(sudo:session): session closed for user root Sep 11 23:30:56.105223 sshd[1681]: Connection closed by 10.0.0.1 port 46564 Sep 11 23:30:56.105790 sshd-session[1678]: pam_unix(sshd:session): session closed for user core Sep 11 23:30:56.114464 systemd[1]: sshd@5-10.0.0.19:22-10.0.0.1:46564.service: Deactivated successfully. Sep 11 23:30:56.116012 systemd[1]: session-6.scope: Deactivated successfully. Sep 11 23:30:56.117449 systemd-logind[1485]: Session 6 logged out. Waiting for processes to exit. Sep 11 23:30:56.118723 systemd[1]: Started sshd@6-10.0.0.19:22-10.0.0.1:46572.service - OpenSSH per-connection server daemon (10.0.0.1:46572). Sep 11 23:30:56.119589 systemd-logind[1485]: Removed session 6. Sep 11 23:30:56.178572 sshd[1714]: Accepted publickey for core from 10.0.0.1 port 46572 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:30:56.180274 sshd-session[1714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:30:56.184141 systemd-logind[1485]: New session 7 of user core. Sep 11 23:30:56.200654 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 11 23:30:56.251941 sudo[1718]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 11 23:30:56.252214 sudo[1718]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 23:30:56.529988 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 11 23:30:56.540895 (dockerd)[1739]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 11 23:30:56.744573 dockerd[1739]: time="2025-09-11T23:30:56.744500732Z" level=info msg="Starting up" Sep 11 23:30:56.745486 dockerd[1739]: time="2025-09-11T23:30:56.745452402Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 11 23:30:56.756739 dockerd[1739]: time="2025-09-11T23:30:56.756698344Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 11 23:30:56.788185 dockerd[1739]: time="2025-09-11T23:30:56.787900581Z" level=info msg="Loading containers: start." Sep 11 23:30:56.795496 kernel: Initializing XFRM netlink socket Sep 11 23:30:56.985429 systemd-networkd[1433]: docker0: Link UP Sep 11 23:30:56.989452 dockerd[1739]: time="2025-09-11T23:30:56.989404896Z" level=info msg="Loading containers: done." Sep 11 23:30:57.001164 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3891095375-merged.mount: Deactivated successfully. Sep 11 23:30:57.002967 dockerd[1739]: time="2025-09-11T23:30:57.002923394Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 11 23:30:57.003041 dockerd[1739]: time="2025-09-11T23:30:57.003004615Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 11 23:30:57.003098 dockerd[1739]: time="2025-09-11T23:30:57.003081915Z" level=info msg="Initializing buildkit" Sep 11 23:30:57.025199 dockerd[1739]: time="2025-09-11T23:30:57.025139025Z" level=info msg="Completed buildkit initialization" Sep 11 23:30:57.030062 dockerd[1739]: time="2025-09-11T23:30:57.030020930Z" level=info msg="Daemon has completed initialization" Sep 11 23:30:57.030251 dockerd[1739]: time="2025-09-11T23:30:57.030146784Z" level=info msg="API listen on /run/docker.sock" Sep 11 23:30:57.030468 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 11 23:30:57.659487 containerd[1515]: time="2025-09-11T23:30:57.659371825Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\"" Sep 11 23:30:58.300203 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount327547715.mount: Deactivated successfully. Sep 11 23:30:59.898510 containerd[1515]: time="2025-09-11T23:30:59.898449491Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:30:59.899737 containerd[1515]: time="2025-09-11T23:30:59.899699984Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.5: active requests=0, bytes read=27390230" Sep 11 23:30:59.900595 containerd[1515]: time="2025-09-11T23:30:59.900569296Z" level=info msg="ImageCreate event name:\"sha256:6a7fd297b49102b08dc3d8d4fd7f1538bcf21d3131eae8bf62ba26ce3283237f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:30:59.902999 containerd[1515]: time="2025-09-11T23:30:59.902973535Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:30:59.904850 containerd[1515]: time="2025-09-11T23:30:59.904681470Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.5\" with image id \"sha256:6a7fd297b49102b08dc3d8d4fd7f1538bcf21d3131eae8bf62ba26ce3283237f\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\", size \"27386827\" in 2.245267423s" Sep 11 23:30:59.904850 containerd[1515]: time="2025-09-11T23:30:59.904732364Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\" returns image reference \"sha256:6a7fd297b49102b08dc3d8d4fd7f1538bcf21d3131eae8bf62ba26ce3283237f\"" Sep 11 23:30:59.905925 containerd[1515]: time="2025-09-11T23:30:59.905896958Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\"" Sep 11 23:31:01.146434 containerd[1515]: time="2025-09-11T23:31:01.146387391Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:01.147284 containerd[1515]: time="2025-09-11T23:31:01.147010801Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.5: active requests=0, bytes read=23547919" Sep 11 23:31:01.148051 containerd[1515]: time="2025-09-11T23:31:01.148013177Z" level=info msg="ImageCreate event name:\"sha256:2dd4c25a937008b7b8a6cdca70d816403b5078b51550926721b7a7762139cd23\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:01.150276 containerd[1515]: time="2025-09-11T23:31:01.150249331Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:01.151246 containerd[1515]: time="2025-09-11T23:31:01.151209845Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.5\" with image id \"sha256:2dd4c25a937008b7b8a6cdca70d816403b5078b51550926721b7a7762139cd23\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\", size \"25135832\" in 1.24527927s" Sep 11 23:31:01.151292 containerd[1515]: time="2025-09-11T23:31:01.151252029Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\" returns image reference \"sha256:2dd4c25a937008b7b8a6cdca70d816403b5078b51550926721b7a7762139cd23\"" Sep 11 23:31:01.151676 containerd[1515]: time="2025-09-11T23:31:01.151653413Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\"" Sep 11 23:31:01.833491 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 11 23:31:01.836174 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:31:02.024706 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:02.028767 (kubelet)[2027]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 23:31:02.072516 kubelet[2027]: E0911 23:31:02.071131 2027 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 23:31:02.076003 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 23:31:02.076137 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 23:31:02.076653 systemd[1]: kubelet.service: Consumed 149ms CPU time, 108.5M memory peak. Sep 11 23:31:02.806594 containerd[1515]: time="2025-09-11T23:31:02.806531761Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:02.808346 containerd[1515]: time="2025-09-11T23:31:02.808296364Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.5: active requests=0, bytes read=18295979" Sep 11 23:31:02.810439 containerd[1515]: time="2025-09-11T23:31:02.810410236Z" level=info msg="ImageCreate event name:\"sha256:5e600beaed8620718e0650dd2721266869ce1d737488c004a869333273e6ec15\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:02.813099 containerd[1515]: time="2025-09-11T23:31:02.813063012Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:02.814050 containerd[1515]: time="2025-09-11T23:31:02.814008338Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.5\" with image id \"sha256:5e600beaed8620718e0650dd2721266869ce1d737488c004a869333273e6ec15\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\", size \"19883910\" in 1.662324462s" Sep 11 23:31:02.814050 containerd[1515]: time="2025-09-11T23:31:02.814045622Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\" returns image reference \"sha256:5e600beaed8620718e0650dd2721266869ce1d737488c004a869333273e6ec15\"" Sep 11 23:31:02.814760 containerd[1515]: time="2025-09-11T23:31:02.814711863Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\"" Sep 11 23:31:03.820762 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount511439575.mount: Deactivated successfully. Sep 11 23:31:04.254875 containerd[1515]: time="2025-09-11T23:31:04.254818431Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:04.255964 containerd[1515]: time="2025-09-11T23:31:04.255526036Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.5: active requests=0, bytes read=28240108" Sep 11 23:31:04.256835 containerd[1515]: time="2025-09-11T23:31:04.256793265Z" level=info msg="ImageCreate event name:\"sha256:021a8d45ab0c346664e47d95595ff5180ce90a22a681ea27904c65ae90788e70\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:04.258796 containerd[1515]: time="2025-09-11T23:31:04.258765647Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:04.259542 containerd[1515]: time="2025-09-11T23:31:04.259283979Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.5\" with image id \"sha256:021a8d45ab0c346664e47d95595ff5180ce90a22a681ea27904c65ae90788e70\", repo tag \"registry.k8s.io/kube-proxy:v1.33.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\", size \"28239125\" in 1.444538585s" Sep 11 23:31:04.259542 containerd[1515]: time="2025-09-11T23:31:04.259310088Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\" returns image reference \"sha256:021a8d45ab0c346664e47d95595ff5180ce90a22a681ea27904c65ae90788e70\"" Sep 11 23:31:04.259817 containerd[1515]: time="2025-09-11T23:31:04.259765476Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 11 23:31:04.738081 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1442344733.mount: Deactivated successfully. Sep 11 23:31:05.695041 containerd[1515]: time="2025-09-11T23:31:05.694974979Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:05.696275 containerd[1515]: time="2025-09-11T23:31:05.696011982Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=19152119" Sep 11 23:31:05.697016 containerd[1515]: time="2025-09-11T23:31:05.696990652Z" level=info msg="ImageCreate event name:\"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:05.701843 containerd[1515]: time="2025-09-11T23:31:05.701789451Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:05.702949 containerd[1515]: time="2025-09-11T23:31:05.702914217Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"19148915\" in 1.443107331s" Sep 11 23:31:05.702998 containerd[1515]: time="2025-09-11T23:31:05.702953802Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\"" Sep 11 23:31:05.703704 containerd[1515]: time="2025-09-11T23:31:05.703646502Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 11 23:31:06.137033 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1759940015.mount: Deactivated successfully. Sep 11 23:31:06.141970 containerd[1515]: time="2025-09-11T23:31:06.141919224Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 23:31:06.142431 containerd[1515]: time="2025-09-11T23:31:06.142397279Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Sep 11 23:31:06.144017 containerd[1515]: time="2025-09-11T23:31:06.143450380Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 23:31:06.145906 containerd[1515]: time="2025-09-11T23:31:06.145865333Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 23:31:06.146730 containerd[1515]: time="2025-09-11T23:31:06.146696602Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 443.015136ms" Sep 11 23:31:06.146730 containerd[1515]: time="2025-09-11T23:31:06.146726217Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Sep 11 23:31:06.147231 containerd[1515]: time="2025-09-11T23:31:06.147199737Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 11 23:31:06.604398 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3603362488.mount: Deactivated successfully. Sep 11 23:31:08.729656 containerd[1515]: time="2025-09-11T23:31:08.729594445Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:08.730782 containerd[1515]: time="2025-09-11T23:31:08.730449749Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=69465859" Sep 11 23:31:08.731688 containerd[1515]: time="2025-09-11T23:31:08.731655274Z" level=info msg="ImageCreate event name:\"sha256:31747a36ce712f0bf61b50a0c06e99768522025e7b8daedd6dc63d1ae84837b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:08.734319 containerd[1515]: time="2025-09-11T23:31:08.734273515Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:08.736288 containerd[1515]: time="2025-09-11T23:31:08.736213246Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:31747a36ce712f0bf61b50a0c06e99768522025e7b8daedd6dc63d1ae84837b5\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"70026017\" in 2.588975877s" Sep 11 23:31:08.736339 containerd[1515]: time="2025-09-11T23:31:08.736295008Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:31747a36ce712f0bf61b50a0c06e99768522025e7b8daedd6dc63d1ae84837b5\"" Sep 11 23:31:12.226061 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 11 23:31:12.227632 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:31:12.374226 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:12.390834 (kubelet)[2192]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 23:31:12.424138 kubelet[2192]: E0911 23:31:12.424070 2192 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 23:31:12.426694 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 23:31:12.426836 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 23:31:12.428532 systemd[1]: kubelet.service: Consumed 139ms CPU time, 105.2M memory peak. Sep 11 23:31:13.833757 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:13.834070 systemd[1]: kubelet.service: Consumed 139ms CPU time, 105.2M memory peak. Sep 11 23:31:13.836150 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:31:13.860406 systemd[1]: Reload requested from client PID 2208 ('systemctl') (unit session-7.scope)... Sep 11 23:31:13.860428 systemd[1]: Reloading... Sep 11 23:31:13.941491 zram_generator::config[2251]: No configuration found. Sep 11 23:31:14.124763 systemd[1]: Reloading finished in 263 ms. Sep 11 23:31:14.175095 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 11 23:31:14.175176 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 11 23:31:14.176515 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:14.176569 systemd[1]: kubelet.service: Consumed 96ms CPU time, 95M memory peak. Sep 11 23:31:14.178259 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:31:14.304343 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:14.312816 (kubelet)[2294]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 11 23:31:14.347840 kubelet[2294]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 23:31:14.347840 kubelet[2294]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 11 23:31:14.347840 kubelet[2294]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 23:31:14.348177 kubelet[2294]: I0911 23:31:14.347881 2294 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 11 23:31:15.127487 kubelet[2294]: I0911 23:31:15.126658 2294 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 11 23:31:15.127487 kubelet[2294]: I0911 23:31:15.126693 2294 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 11 23:31:15.127487 kubelet[2294]: I0911 23:31:15.127060 2294 server.go:956] "Client rotation is on, will bootstrap in background" Sep 11 23:31:15.147172 kubelet[2294]: E0911 23:31:15.147131 2294 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.19:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.19:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 11 23:31:15.148338 kubelet[2294]: I0911 23:31:15.148316 2294 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 23:31:15.155503 kubelet[2294]: I0911 23:31:15.155476 2294 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 11 23:31:15.158291 kubelet[2294]: I0911 23:31:15.158265 2294 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 11 23:31:15.159532 kubelet[2294]: I0911 23:31:15.159488 2294 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 11 23:31:15.159810 kubelet[2294]: I0911 23:31:15.159628 2294 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 11 23:31:15.160022 kubelet[2294]: I0911 23:31:15.160007 2294 topology_manager.go:138] "Creating topology manager with none policy" Sep 11 23:31:15.160094 kubelet[2294]: I0911 23:31:15.160084 2294 container_manager_linux.go:303] "Creating device plugin manager" Sep 11 23:31:15.160346 kubelet[2294]: I0911 23:31:15.160330 2294 state_mem.go:36] "Initialized new in-memory state store" Sep 11 23:31:15.162935 kubelet[2294]: I0911 23:31:15.162912 2294 kubelet.go:480] "Attempting to sync node with API server" Sep 11 23:31:15.163037 kubelet[2294]: I0911 23:31:15.163026 2294 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 11 23:31:15.163145 kubelet[2294]: I0911 23:31:15.163133 2294 kubelet.go:386] "Adding apiserver pod source" Sep 11 23:31:15.163200 kubelet[2294]: I0911 23:31:15.163191 2294 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 11 23:31:15.164348 kubelet[2294]: I0911 23:31:15.164328 2294 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 11 23:31:15.165139 kubelet[2294]: I0911 23:31:15.165117 2294 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 11 23:31:15.166239 kubelet[2294]: W0911 23:31:15.166220 2294 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 11 23:31:15.169511 kubelet[2294]: I0911 23:31:15.168835 2294 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 11 23:31:15.169511 kubelet[2294]: I0911 23:31:15.168884 2294 server.go:1289] "Started kubelet" Sep 11 23:31:15.170013 kubelet[2294]: I0911 23:31:15.169967 2294 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 11 23:31:15.170241 kubelet[2294]: E0911 23:31:15.170202 2294 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.19:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.19:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 11 23:31:15.170308 kubelet[2294]: E0911 23:31:15.170290 2294 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.19:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.19:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 11 23:31:15.171204 kubelet[2294]: I0911 23:31:15.171169 2294 server.go:317] "Adding debug handlers to kubelet server" Sep 11 23:31:15.172484 kubelet[2294]: I0911 23:31:15.172388 2294 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 11 23:31:15.172934 kubelet[2294]: I0911 23:31:15.172908 2294 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 11 23:31:15.173435 kubelet[2294]: I0911 23:31:15.173413 2294 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 11 23:31:15.174219 kubelet[2294]: I0911 23:31:15.174189 2294 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 11 23:31:15.175003 kubelet[2294]: E0911 23:31:15.174978 2294 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 23:31:15.175124 kubelet[2294]: I0911 23:31:15.175112 2294 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 11 23:31:15.175384 kubelet[2294]: I0911 23:31:15.175363 2294 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 11 23:31:15.175536 kubelet[2294]: I0911 23:31:15.175522 2294 reconciler.go:26] "Reconciler: start to sync state" Sep 11 23:31:15.176025 kubelet[2294]: I0911 23:31:15.175999 2294 factory.go:223] Registration of the systemd container factory successfully Sep 11 23:31:15.176133 kubelet[2294]: I0911 23:31:15.176109 2294 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 11 23:31:15.176267 kubelet[2294]: E0911 23:31:15.176240 2294 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.19:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.19:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 11 23:31:15.176859 kubelet[2294]: E0911 23:31:15.176824 2294 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.19:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.19:6443: connect: connection refused" interval="200ms" Sep 11 23:31:15.177081 kubelet[2294]: E0911 23:31:15.176934 2294 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 11 23:31:15.177434 kubelet[2294]: I0911 23:31:15.177415 2294 factory.go:223] Registration of the containerd container factory successfully Sep 11 23:31:15.177851 kubelet[2294]: E0911 23:31:15.176796 2294 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.19:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.19:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18645e5554c653b0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-11 23:31:15.168850864 +0000 UTC m=+0.852682982,LastTimestamp:2025-09-11 23:31:15.168850864 +0000 UTC m=+0.852682982,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 11 23:31:15.188891 kubelet[2294]: I0911 23:31:15.188868 2294 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 11 23:31:15.188891 kubelet[2294]: I0911 23:31:15.188883 2294 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 11 23:31:15.188891 kubelet[2294]: I0911 23:31:15.188901 2294 state_mem.go:36] "Initialized new in-memory state store" Sep 11 23:31:15.192966 kubelet[2294]: I0911 23:31:15.192918 2294 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 11 23:31:15.194412 kubelet[2294]: I0911 23:31:15.194378 2294 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 11 23:31:15.194412 kubelet[2294]: I0911 23:31:15.194411 2294 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 11 23:31:15.194508 kubelet[2294]: I0911 23:31:15.194437 2294 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 11 23:31:15.194508 kubelet[2294]: I0911 23:31:15.194444 2294 kubelet.go:2436] "Starting kubelet main sync loop" Sep 11 23:31:15.194650 kubelet[2294]: E0911 23:31:15.194577 2294 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 11 23:31:15.275373 kubelet[2294]: E0911 23:31:15.275291 2294 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 23:31:15.295712 kubelet[2294]: E0911 23:31:15.295668 2294 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 11 23:31:15.306669 kubelet[2294]: I0911 23:31:15.306582 2294 policy_none.go:49] "None policy: Start" Sep 11 23:31:15.306669 kubelet[2294]: I0911 23:31:15.306607 2294 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 11 23:31:15.306669 kubelet[2294]: I0911 23:31:15.306620 2294 state_mem.go:35] "Initializing new in-memory state store" Sep 11 23:31:15.306796 kubelet[2294]: E0911 23:31:15.306774 2294 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.19:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.19:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 11 23:31:15.312635 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 11 23:31:15.323805 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 11 23:31:15.327807 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 11 23:31:15.346496 kubelet[2294]: E0911 23:31:15.346382 2294 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 11 23:31:15.346853 kubelet[2294]: I0911 23:31:15.346641 2294 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 11 23:31:15.346853 kubelet[2294]: I0911 23:31:15.346653 2294 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 11 23:31:15.346950 kubelet[2294]: I0911 23:31:15.346881 2294 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 11 23:31:15.348591 kubelet[2294]: E0911 23:31:15.348536 2294 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 11 23:31:15.348869 kubelet[2294]: E0911 23:31:15.348608 2294 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 11 23:31:15.378066 kubelet[2294]: E0911 23:31:15.377949 2294 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.19:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.19:6443: connect: connection refused" interval="400ms" Sep 11 23:31:15.448475 kubelet[2294]: I0911 23:31:15.448385 2294 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 23:31:15.448957 kubelet[2294]: E0911 23:31:15.448933 2294 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.19:6443/api/v1/nodes\": dial tcp 10.0.0.19:6443: connect: connection refused" node="localhost" Sep 11 23:31:15.505660 systemd[1]: Created slice kubepods-burstable-pod768e42a9509eb750dc68239187637f92.slice - libcontainer container kubepods-burstable-pod768e42a9509eb750dc68239187637f92.slice. Sep 11 23:31:15.516249 kubelet[2294]: E0911 23:31:15.516203 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:15.518803 systemd[1]: Created slice kubepods-burstable-podb678d5c6713e936e66aa5bb73166297e.slice - libcontainer container kubepods-burstable-podb678d5c6713e936e66aa5bb73166297e.slice. Sep 11 23:31:15.535692 kubelet[2294]: E0911 23:31:15.535574 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:15.537833 systemd[1]: Created slice kubepods-burstable-pod7b968cf906b2d9d713a362c43868bef2.slice - libcontainer container kubepods-burstable-pod7b968cf906b2d9d713a362c43868bef2.slice. Sep 11 23:31:15.539409 kubelet[2294]: E0911 23:31:15.539366 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:15.651251 kubelet[2294]: I0911 23:31:15.651151 2294 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 23:31:15.651605 kubelet[2294]: E0911 23:31:15.651492 2294 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.19:6443/api/v1/nodes\": dial tcp 10.0.0.19:6443: connect: connection refused" node="localhost" Sep 11 23:31:15.678320 kubelet[2294]: I0911 23:31:15.678279 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/768e42a9509eb750dc68239187637f92-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"768e42a9509eb750dc68239187637f92\") " pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:15.678320 kubelet[2294]: I0911 23:31:15.678318 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:15.678644 kubelet[2294]: I0911 23:31:15.678482 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:15.678690 kubelet[2294]: I0911 23:31:15.678656 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/768e42a9509eb750dc68239187637f92-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"768e42a9509eb750dc68239187637f92\") " pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:15.678712 kubelet[2294]: I0911 23:31:15.678692 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:15.678734 kubelet[2294]: I0911 23:31:15.678714 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:15.678757 kubelet[2294]: I0911 23:31:15.678733 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:15.678808 kubelet[2294]: I0911 23:31:15.678794 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7b968cf906b2d9d713a362c43868bef2-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7b968cf906b2d9d713a362c43868bef2\") " pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:15.678886 kubelet[2294]: I0911 23:31:15.678835 2294 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/768e42a9509eb750dc68239187637f92-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"768e42a9509eb750dc68239187637f92\") " pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:15.779411 kubelet[2294]: E0911 23:31:15.779363 2294 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.19:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.19:6443: connect: connection refused" interval="800ms" Sep 11 23:31:15.817944 kubelet[2294]: E0911 23:31:15.817893 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:15.818576 containerd[1515]: time="2025-09-11T23:31:15.818522552Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:768e42a9509eb750dc68239187637f92,Namespace:kube-system,Attempt:0,}" Sep 11 23:31:15.835988 kubelet[2294]: E0911 23:31:15.835948 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:15.836427 containerd[1515]: time="2025-09-11T23:31:15.836394322Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b678d5c6713e936e66aa5bb73166297e,Namespace:kube-system,Attempt:0,}" Sep 11 23:31:15.839323 containerd[1515]: time="2025-09-11T23:31:15.839288603Z" level=info msg="connecting to shim bea751ad6d9df803ca036ea5256733f6c42fe789b8dede70fd4e86c79d67e808" address="unix:///run/containerd/s/ddde5c45ba3e30586b01aaa3057852b6a4280e86c881758bb3626cb73ca4b51e" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:15.839894 kubelet[2294]: E0911 23:31:15.839872 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:15.840504 containerd[1515]: time="2025-09-11T23:31:15.840354153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7b968cf906b2d9d713a362c43868bef2,Namespace:kube-system,Attempt:0,}" Sep 11 23:31:15.864671 systemd[1]: Started cri-containerd-bea751ad6d9df803ca036ea5256733f6c42fe789b8dede70fd4e86c79d67e808.scope - libcontainer container bea751ad6d9df803ca036ea5256733f6c42fe789b8dede70fd4e86c79d67e808. Sep 11 23:31:15.922315 containerd[1515]: time="2025-09-11T23:31:15.922156934Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:768e42a9509eb750dc68239187637f92,Namespace:kube-system,Attempt:0,} returns sandbox id \"bea751ad6d9df803ca036ea5256733f6c42fe789b8dede70fd4e86c79d67e808\"" Sep 11 23:31:15.924451 kubelet[2294]: E0911 23:31:15.924406 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:15.929002 containerd[1515]: time="2025-09-11T23:31:15.928962158Z" level=info msg="CreateContainer within sandbox \"bea751ad6d9df803ca036ea5256733f6c42fe789b8dede70fd4e86c79d67e808\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 11 23:31:15.945840 containerd[1515]: time="2025-09-11T23:31:15.945789958Z" level=info msg="connecting to shim 87a29919c06273ddb3f774de11feba7f2673957202d4171ef550f1f0cba17807" address="unix:///run/containerd/s/b4fc7f8d5b369f5abd90ad3f0a38a57a73949f162568026d48195af6effd4225" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:15.950916 containerd[1515]: time="2025-09-11T23:31:15.950774580Z" level=info msg="Container 3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:15.956694 containerd[1515]: time="2025-09-11T23:31:15.956647342Z" level=info msg="connecting to shim fba118a5212a7bda757ec49974f38a76be5682bc6e062bc719f97fc126188cd8" address="unix:///run/containerd/s/af3972af76fcbb2eb9e303e8c0693ee7116b4514ff25a05e0fa329ebd4f36cc9" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:15.963951 containerd[1515]: time="2025-09-11T23:31:15.963283002Z" level=info msg="CreateContainer within sandbox \"bea751ad6d9df803ca036ea5256733f6c42fe789b8dede70fd4e86c79d67e808\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad\"" Sep 11 23:31:15.964634 containerd[1515]: time="2025-09-11T23:31:15.964562119Z" level=info msg="StartContainer for \"3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad\"" Sep 11 23:31:15.966146 containerd[1515]: time="2025-09-11T23:31:15.965659140Z" level=info msg="connecting to shim 3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad" address="unix:///run/containerd/s/ddde5c45ba3e30586b01aaa3057852b6a4280e86c881758bb3626cb73ca4b51e" protocol=ttrpc version=3 Sep 11 23:31:15.968655 systemd[1]: Started cri-containerd-87a29919c06273ddb3f774de11feba7f2673957202d4171ef550f1f0cba17807.scope - libcontainer container 87a29919c06273ddb3f774de11feba7f2673957202d4171ef550f1f0cba17807. Sep 11 23:31:15.991667 systemd[1]: Started cri-containerd-3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad.scope - libcontainer container 3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad. Sep 11 23:31:15.993608 systemd[1]: Started cri-containerd-fba118a5212a7bda757ec49974f38a76be5682bc6e062bc719f97fc126188cd8.scope - libcontainer container fba118a5212a7bda757ec49974f38a76be5682bc6e062bc719f97fc126188cd8. Sep 11 23:31:16.006299 containerd[1515]: time="2025-09-11T23:31:16.006134615Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7b968cf906b2d9d713a362c43868bef2,Namespace:kube-system,Attempt:0,} returns sandbox id \"87a29919c06273ddb3f774de11feba7f2673957202d4171ef550f1f0cba17807\"" Sep 11 23:31:16.008003 kubelet[2294]: E0911 23:31:16.007955 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:16.016934 containerd[1515]: time="2025-09-11T23:31:16.016897647Z" level=info msg="CreateContainer within sandbox \"87a29919c06273ddb3f774de11feba7f2673957202d4171ef550f1f0cba17807\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 11 23:31:16.029137 containerd[1515]: time="2025-09-11T23:31:16.028508837Z" level=info msg="Container 52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:16.036084 containerd[1515]: time="2025-09-11T23:31:16.035948656Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b678d5c6713e936e66aa5bb73166297e,Namespace:kube-system,Attempt:0,} returns sandbox id \"fba118a5212a7bda757ec49974f38a76be5682bc6e062bc719f97fc126188cd8\"" Sep 11 23:31:16.036712 kubelet[2294]: E0911 23:31:16.036682 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:16.041489 containerd[1515]: time="2025-09-11T23:31:16.040489540Z" level=info msg="CreateContainer within sandbox \"fba118a5212a7bda757ec49974f38a76be5682bc6e062bc719f97fc126188cd8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 11 23:31:16.042301 containerd[1515]: time="2025-09-11T23:31:16.042260400Z" level=info msg="CreateContainer within sandbox \"87a29919c06273ddb3f774de11feba7f2673957202d4171ef550f1f0cba17807\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2\"" Sep 11 23:31:16.047667 containerd[1515]: time="2025-09-11T23:31:16.047638313Z" level=info msg="StartContainer for \"3426b5caea4b5c0a28624123643c9ee1ce2149f9e529dd4a426f90e2c99d8aad\" returns successfully" Sep 11 23:31:16.048450 containerd[1515]: time="2025-09-11T23:31:16.047980042Z" level=info msg="StartContainer for \"52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2\"" Sep 11 23:31:16.049920 containerd[1515]: time="2025-09-11T23:31:16.049874766Z" level=info msg="connecting to shim 52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2" address="unix:///run/containerd/s/b4fc7f8d5b369f5abd90ad3f0a38a57a73949f162568026d48195af6effd4225" protocol=ttrpc version=3 Sep 11 23:31:16.050415 containerd[1515]: time="2025-09-11T23:31:16.050387841Z" level=info msg="Container b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:16.053667 kubelet[2294]: I0911 23:31:16.053631 2294 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 23:31:16.054259 kubelet[2294]: E0911 23:31:16.054184 2294 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.19:6443/api/v1/nodes\": dial tcp 10.0.0.19:6443: connect: connection refused" node="localhost" Sep 11 23:31:16.059849 containerd[1515]: time="2025-09-11T23:31:16.059798768Z" level=info msg="CreateContainer within sandbox \"fba118a5212a7bda757ec49974f38a76be5682bc6e062bc719f97fc126188cd8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d\"" Sep 11 23:31:16.060331 containerd[1515]: time="2025-09-11T23:31:16.060303916Z" level=info msg="StartContainer for \"b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d\"" Sep 11 23:31:16.061416 containerd[1515]: time="2025-09-11T23:31:16.061389395Z" level=info msg="connecting to shim b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d" address="unix:///run/containerd/s/af3972af76fcbb2eb9e303e8c0693ee7116b4514ff25a05e0fa329ebd4f36cc9" protocol=ttrpc version=3 Sep 11 23:31:16.071611 systemd[1]: Started cri-containerd-52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2.scope - libcontainer container 52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2. Sep 11 23:31:16.089631 systemd[1]: Started cri-containerd-b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d.scope - libcontainer container b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d. Sep 11 23:31:16.133650 containerd[1515]: time="2025-09-11T23:31:16.133608057Z" level=info msg="StartContainer for \"b9fba92a0dabd5d8f27e3e79fce5878df61d76110fc0918b78594b0a5fa45a9d\" returns successfully" Sep 11 23:31:16.141788 containerd[1515]: time="2025-09-11T23:31:16.141756956Z" level=info msg="StartContainer for \"52fe8b73d6a4176cf1ef8bcfcdef61bc52c0583dfbd9d364fc0d58c42ac893b2\" returns successfully" Sep 11 23:31:16.204026 kubelet[2294]: E0911 23:31:16.203272 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:16.205082 kubelet[2294]: E0911 23:31:16.205026 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:16.205959 kubelet[2294]: E0911 23:31:16.205932 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:16.206101 kubelet[2294]: E0911 23:31:16.206079 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:16.207196 kubelet[2294]: E0911 23:31:16.207172 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:16.207294 kubelet[2294]: E0911 23:31:16.207281 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:16.836968 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4000366523.mount: Deactivated successfully. Sep 11 23:31:16.856651 kubelet[2294]: I0911 23:31:16.855713 2294 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 23:31:17.210161 kubelet[2294]: E0911 23:31:17.210061 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:17.210250 kubelet[2294]: E0911 23:31:17.210203 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:17.210962 kubelet[2294]: E0911 23:31:17.210531 2294 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 23:31:17.210962 kubelet[2294]: E0911 23:31:17.210644 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:17.990489 kubelet[2294]: E0911 23:31:17.989650 2294 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 11 23:31:18.167675 kubelet[2294]: I0911 23:31:18.167639 2294 apiserver.go:52] "Watching apiserver" Sep 11 23:31:18.176957 kubelet[2294]: I0911 23:31:18.176917 2294 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 11 23:31:18.178663 kubelet[2294]: I0911 23:31:18.177301 2294 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:18.179735 kubelet[2294]: I0911 23:31:18.177305 2294 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 11 23:31:18.188366 kubelet[2294]: E0911 23:31:18.188320 2294 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:18.188517 kubelet[2294]: I0911 23:31:18.188505 2294 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:18.192026 kubelet[2294]: E0911 23:31:18.191993 2294 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:18.192451 kubelet[2294]: I0911 23:31:18.192209 2294 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:18.196567 kubelet[2294]: E0911 23:31:18.196538 2294 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:18.210480 kubelet[2294]: I0911 23:31:18.210225 2294 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:18.215069 kubelet[2294]: E0911 23:31:18.214976 2294 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:18.215777 kubelet[2294]: E0911 23:31:18.215434 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:18.576126 kubelet[2294]: I0911 23:31:18.576074 2294 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:18.578554 kubelet[2294]: E0911 23:31:18.578512 2294 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:18.578777 kubelet[2294]: E0911 23:31:18.578744 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:20.018784 kubelet[2294]: I0911 23:31:20.018741 2294 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:20.025330 kubelet[2294]: E0911 23:31:20.025291 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:20.214794 kubelet[2294]: E0911 23:31:20.214740 2294 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:20.250316 systemd[1]: Reload requested from client PID 2583 ('systemctl') (unit session-7.scope)... Sep 11 23:31:20.250335 systemd[1]: Reloading... Sep 11 23:31:20.336588 kernel: hrtimer: interrupt took 11527883 ns Sep 11 23:31:20.336694 zram_generator::config[2625]: No configuration found. Sep 11 23:31:20.512917 systemd[1]: Reloading finished in 262 ms. Sep 11 23:31:20.545710 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:31:20.561517 systemd[1]: kubelet.service: Deactivated successfully. Sep 11 23:31:20.561755 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:20.561819 systemd[1]: kubelet.service: Consumed 1.239s CPU time, 126.3M memory peak. Sep 11 23:31:20.563612 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 23:31:20.741431 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 23:31:20.746311 (kubelet)[2668]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 11 23:31:20.809578 kubelet[2668]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 23:31:20.809578 kubelet[2668]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 11 23:31:20.809578 kubelet[2668]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 23:31:20.809578 kubelet[2668]: I0911 23:31:20.808314 2668 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 11 23:31:20.816094 kubelet[2668]: I0911 23:31:20.816046 2668 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 11 23:31:20.816094 kubelet[2668]: I0911 23:31:20.816085 2668 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 11 23:31:20.816322 kubelet[2668]: I0911 23:31:20.816305 2668 server.go:956] "Client rotation is on, will bootstrap in background" Sep 11 23:31:20.817727 kubelet[2668]: I0911 23:31:20.817689 2668 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 11 23:31:20.820081 kubelet[2668]: I0911 23:31:20.820028 2668 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 23:31:20.823967 kubelet[2668]: I0911 23:31:20.823917 2668 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 11 23:31:20.827515 kubelet[2668]: I0911 23:31:20.827038 2668 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 11 23:31:20.827515 kubelet[2668]: I0911 23:31:20.827322 2668 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 11 23:31:20.827624 kubelet[2668]: I0911 23:31:20.827354 2668 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 11 23:31:20.827624 kubelet[2668]: I0911 23:31:20.827561 2668 topology_manager.go:138] "Creating topology manager with none policy" Sep 11 23:31:20.827624 kubelet[2668]: I0911 23:31:20.827569 2668 container_manager_linux.go:303] "Creating device plugin manager" Sep 11 23:31:20.827624 kubelet[2668]: I0911 23:31:20.827616 2668 state_mem.go:36] "Initialized new in-memory state store" Sep 11 23:31:20.827793 kubelet[2668]: I0911 23:31:20.827775 2668 kubelet.go:480] "Attempting to sync node with API server" Sep 11 23:31:20.827793 kubelet[2668]: I0911 23:31:20.827792 2668 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 11 23:31:20.827840 kubelet[2668]: I0911 23:31:20.827816 2668 kubelet.go:386] "Adding apiserver pod source" Sep 11 23:31:20.827840 kubelet[2668]: I0911 23:31:20.827828 2668 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 11 23:31:20.828838 kubelet[2668]: I0911 23:31:20.828811 2668 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 11 23:31:20.833970 kubelet[2668]: I0911 23:31:20.833920 2668 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 11 23:31:20.837764 kubelet[2668]: I0911 23:31:20.837733 2668 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 11 23:31:20.837876 kubelet[2668]: I0911 23:31:20.837785 2668 server.go:1289] "Started kubelet" Sep 11 23:31:20.841560 kubelet[2668]: I0911 23:31:20.839811 2668 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 11 23:31:20.843286 kubelet[2668]: I0911 23:31:20.843176 2668 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 11 23:31:20.843698 kubelet[2668]: I0911 23:31:20.843636 2668 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 11 23:31:20.843935 kubelet[2668]: I0911 23:31:20.843908 2668 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 11 23:31:20.845026 kubelet[2668]: I0911 23:31:20.845008 2668 server.go:317] "Adding debug handlers to kubelet server" Sep 11 23:31:20.845825 kubelet[2668]: I0911 23:31:20.845798 2668 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 11 23:31:20.847963 kubelet[2668]: I0911 23:31:20.847942 2668 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 11 23:31:20.848213 kubelet[2668]: I0911 23:31:20.848195 2668 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 11 23:31:20.848392 kubelet[2668]: I0911 23:31:20.848380 2668 reconciler.go:26] "Reconciler: start to sync state" Sep 11 23:31:20.851921 kubelet[2668]: I0911 23:31:20.851885 2668 factory.go:223] Registration of the systemd container factory successfully Sep 11 23:31:20.852007 kubelet[2668]: I0911 23:31:20.851981 2668 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 11 23:31:20.858586 kubelet[2668]: I0911 23:31:20.858550 2668 factory.go:223] Registration of the containerd container factory successfully Sep 11 23:31:20.860836 kubelet[2668]: I0911 23:31:20.860802 2668 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 11 23:31:20.862833 kubelet[2668]: I0911 23:31:20.862808 2668 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 11 23:31:20.863341 kubelet[2668]: I0911 23:31:20.863323 2668 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 11 23:31:20.863444 kubelet[2668]: I0911 23:31:20.863431 2668 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 11 23:31:20.863555 kubelet[2668]: I0911 23:31:20.863545 2668 kubelet.go:2436] "Starting kubelet main sync loop" Sep 11 23:31:20.863676 kubelet[2668]: E0911 23:31:20.863649 2668 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 11 23:31:20.869234 kubelet[2668]: E0911 23:31:20.858557 2668 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 11 23:31:20.897758 kubelet[2668]: I0911 23:31:20.897725 2668 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 11 23:31:20.897758 kubelet[2668]: I0911 23:31:20.897746 2668 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 11 23:31:20.897758 kubelet[2668]: I0911 23:31:20.897766 2668 state_mem.go:36] "Initialized new in-memory state store" Sep 11 23:31:20.897993 kubelet[2668]: I0911 23:31:20.897904 2668 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 11 23:31:20.897993 kubelet[2668]: I0911 23:31:20.897915 2668 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 11 23:31:20.897993 kubelet[2668]: I0911 23:31:20.897932 2668 policy_none.go:49] "None policy: Start" Sep 11 23:31:20.897993 kubelet[2668]: I0911 23:31:20.897942 2668 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 11 23:31:20.897993 kubelet[2668]: I0911 23:31:20.897951 2668 state_mem.go:35] "Initializing new in-memory state store" Sep 11 23:31:20.898132 kubelet[2668]: I0911 23:31:20.898031 2668 state_mem.go:75] "Updated machine memory state" Sep 11 23:31:20.902386 kubelet[2668]: E0911 23:31:20.901801 2668 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 11 23:31:20.902386 kubelet[2668]: I0911 23:31:20.901990 2668 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 11 23:31:20.902386 kubelet[2668]: I0911 23:31:20.902003 2668 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 11 23:31:20.902386 kubelet[2668]: I0911 23:31:20.902304 2668 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 11 23:31:20.907270 kubelet[2668]: E0911 23:31:20.907234 2668 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 11 23:31:20.964770 kubelet[2668]: I0911 23:31:20.964731 2668 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:20.965015 kubelet[2668]: I0911 23:31:20.964758 2668 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:20.965162 kubelet[2668]: I0911 23:31:20.964816 2668 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:20.972871 kubelet[2668]: E0911 23:31:20.972819 2668 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.009182 kubelet[2668]: I0911 23:31:21.009146 2668 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 23:31:21.016189 kubelet[2668]: I0911 23:31:21.016138 2668 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 11 23:31:21.016366 kubelet[2668]: I0911 23:31:21.016265 2668 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 11 23:31:21.149816 kubelet[2668]: I0911 23:31:21.149683 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.149816 kubelet[2668]: I0911 23:31:21.149724 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.149816 kubelet[2668]: I0911 23:31:21.149745 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/768e42a9509eb750dc68239187637f92-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"768e42a9509eb750dc68239187637f92\") " pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:21.149816 kubelet[2668]: I0911 23:31:21.149764 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/768e42a9509eb750dc68239187637f92-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"768e42a9509eb750dc68239187637f92\") " pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:21.149816 kubelet[2668]: I0911 23:31:21.149784 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/768e42a9509eb750dc68239187637f92-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"768e42a9509eb750dc68239187637f92\") " pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:21.150024 kubelet[2668]: I0911 23:31:21.149801 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.150024 kubelet[2668]: I0911 23:31:21.149818 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.150024 kubelet[2668]: I0911 23:31:21.149833 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7b968cf906b2d9d713a362c43868bef2-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7b968cf906b2d9d713a362c43868bef2\") " pod="kube-system/kube-scheduler-localhost" Sep 11 23:31:21.150024 kubelet[2668]: I0911 23:31:21.149858 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.271307 kubelet[2668]: E0911 23:31:21.271205 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:21.273504 kubelet[2668]: E0911 23:31:21.273460 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:21.273504 kubelet[2668]: E0911 23:31:21.273499 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:21.828760 kubelet[2668]: I0911 23:31:21.828687 2668 apiserver.go:52] "Watching apiserver" Sep 11 23:31:21.849107 kubelet[2668]: I0911 23:31:21.848405 2668 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 11 23:31:21.880492 kubelet[2668]: I0911 23:31:21.880440 2668 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.880492 kubelet[2668]: I0911 23:31:21.880482 2668 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:21.881240 kubelet[2668]: E0911 23:31:21.881149 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:21.888538 kubelet[2668]: E0911 23:31:21.888179 2668 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 11 23:31:21.888538 kubelet[2668]: E0911 23:31:21.888432 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:21.890568 kubelet[2668]: E0911 23:31:21.889113 2668 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 11 23:31:21.890778 kubelet[2668]: E0911 23:31:21.890728 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:21.917245 kubelet[2668]: I0911 23:31:21.917179 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.9171609699999999 podStartE2EDuration="1.91716097s" podCreationTimestamp="2025-09-11 23:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:31:21.906797543 +0000 UTC m=+1.156317663" watchObservedRunningTime="2025-09-11 23:31:21.91716097 +0000 UTC m=+1.166681090" Sep 11 23:31:21.917419 kubelet[2668]: I0911 23:31:21.917296 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.917292793 podStartE2EDuration="1.917292793s" podCreationTimestamp="2025-09-11 23:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:31:21.916981763 +0000 UTC m=+1.166501883" watchObservedRunningTime="2025-09-11 23:31:21.917292793 +0000 UTC m=+1.166812873" Sep 11 23:31:21.927677 kubelet[2668]: I0911 23:31:21.927558 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.9275422039999999 podStartE2EDuration="1.927542204s" podCreationTimestamp="2025-09-11 23:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:31:21.927458684 +0000 UTC m=+1.176978804" watchObservedRunningTime="2025-09-11 23:31:21.927542204 +0000 UTC m=+1.177062284" Sep 11 23:31:22.882963 kubelet[2668]: E0911 23:31:22.882607 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:22.882963 kubelet[2668]: E0911 23:31:22.882752 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:22.883803 kubelet[2668]: E0911 23:31:22.883780 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:23.884673 kubelet[2668]: E0911 23:31:23.884572 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:23.884673 kubelet[2668]: E0911 23:31:23.884661 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:27.289335 kubelet[2668]: I0911 23:31:27.289249 2668 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 11 23:31:27.290491 kubelet[2668]: I0911 23:31:27.289715 2668 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 11 23:31:27.290570 containerd[1515]: time="2025-09-11T23:31:27.289535579Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 11 23:31:27.891139 systemd[1]: Created slice kubepods-besteffort-pod8d40bb21_bb07_4329_bf3a_b0e1c279e728.slice - libcontainer container kubepods-besteffort-pod8d40bb21_bb07_4329_bf3a_b0e1c279e728.slice. Sep 11 23:31:27.893682 kubelet[2668]: I0911 23:31:27.893244 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d40bb21-bb07-4329-bf3a-b0e1c279e728-lib-modules\") pod \"kube-proxy-sh9jt\" (UID: \"8d40bb21-bb07-4329-bf3a-b0e1c279e728\") " pod="kube-system/kube-proxy-sh9jt" Sep 11 23:31:27.893682 kubelet[2668]: I0911 23:31:27.893334 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/8d40bb21-bb07-4329-bf3a-b0e1c279e728-kube-proxy\") pod \"kube-proxy-sh9jt\" (UID: \"8d40bb21-bb07-4329-bf3a-b0e1c279e728\") " pod="kube-system/kube-proxy-sh9jt" Sep 11 23:31:27.893682 kubelet[2668]: I0911 23:31:27.893357 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dgl5\" (UniqueName: \"kubernetes.io/projected/8d40bb21-bb07-4329-bf3a-b0e1c279e728-kube-api-access-2dgl5\") pod \"kube-proxy-sh9jt\" (UID: \"8d40bb21-bb07-4329-bf3a-b0e1c279e728\") " pod="kube-system/kube-proxy-sh9jt" Sep 11 23:31:27.893682 kubelet[2668]: I0911 23:31:27.893378 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8d40bb21-bb07-4329-bf3a-b0e1c279e728-xtables-lock\") pod \"kube-proxy-sh9jt\" (UID: \"8d40bb21-bb07-4329-bf3a-b0e1c279e728\") " pod="kube-system/kube-proxy-sh9jt" Sep 11 23:31:28.206251 kubelet[2668]: E0911 23:31:28.206105 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:28.206873 containerd[1515]: time="2025-09-11T23:31:28.206824952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-sh9jt,Uid:8d40bb21-bb07-4329-bf3a-b0e1c279e728,Namespace:kube-system,Attempt:0,}" Sep 11 23:31:28.232177 containerd[1515]: time="2025-09-11T23:31:28.232133641Z" level=info msg="connecting to shim ffb95309b3119817be08ecc6fb90f8cec473b95e763852cb73ab81e78bafaefd" address="unix:///run/containerd/s/556c43ed8571e4fee9f6c3937003e190bb15b44287b53003aefbb8f5b3ae7b81" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:28.273924 systemd[1]: Started cri-containerd-ffb95309b3119817be08ecc6fb90f8cec473b95e763852cb73ab81e78bafaefd.scope - libcontainer container ffb95309b3119817be08ecc6fb90f8cec473b95e763852cb73ab81e78bafaefd. Sep 11 23:31:28.300534 containerd[1515]: time="2025-09-11T23:31:28.300455981Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-sh9jt,Uid:8d40bb21-bb07-4329-bf3a-b0e1c279e728,Namespace:kube-system,Attempt:0,} returns sandbox id \"ffb95309b3119817be08ecc6fb90f8cec473b95e763852cb73ab81e78bafaefd\"" Sep 11 23:31:28.301836 kubelet[2668]: E0911 23:31:28.301637 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:28.308924 containerd[1515]: time="2025-09-11T23:31:28.308887396Z" level=info msg="CreateContainer within sandbox \"ffb95309b3119817be08ecc6fb90f8cec473b95e763852cb73ab81e78bafaefd\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 11 23:31:28.323646 containerd[1515]: time="2025-09-11T23:31:28.323600659Z" level=info msg="Container eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:28.332768 containerd[1515]: time="2025-09-11T23:31:28.330806905Z" level=info msg="CreateContainer within sandbox \"ffb95309b3119817be08ecc6fb90f8cec473b95e763852cb73ab81e78bafaefd\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f\"" Sep 11 23:31:28.332768 containerd[1515]: time="2025-09-11T23:31:28.331664123Z" level=info msg="StartContainer for \"eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f\"" Sep 11 23:31:28.333359 containerd[1515]: time="2025-09-11T23:31:28.333326263Z" level=info msg="connecting to shim eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f" address="unix:///run/containerd/s/556c43ed8571e4fee9f6c3937003e190bb15b44287b53003aefbb8f5b3ae7b81" protocol=ttrpc version=3 Sep 11 23:31:28.355702 systemd[1]: Started cri-containerd-eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f.scope - libcontainer container eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f. Sep 11 23:31:28.395780 containerd[1515]: time="2025-09-11T23:31:28.395737986Z" level=info msg="StartContainer for \"eba6b0c96325e8ec7aa1c8befc1a19435662580917233fbfe117a3c53cfa8b2f\" returns successfully" Sep 11 23:31:28.528595 systemd[1]: Created slice kubepods-besteffort-pod514dffb5_c294_4238_9bbc_9cea6d258d66.slice - libcontainer container kubepods-besteffort-pod514dffb5_c294_4238_9bbc_9cea6d258d66.slice. Sep 11 23:31:28.598095 kubelet[2668]: I0911 23:31:28.597981 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/514dffb5-c294-4238-9bbc-9cea6d258d66-var-lib-calico\") pod \"tigera-operator-755d956888-2sn9g\" (UID: \"514dffb5-c294-4238-9bbc-9cea6d258d66\") " pod="tigera-operator/tigera-operator-755d956888-2sn9g" Sep 11 23:31:28.598095 kubelet[2668]: I0911 23:31:28.598042 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4fsz\" (UniqueName: \"kubernetes.io/projected/514dffb5-c294-4238-9bbc-9cea6d258d66-kube-api-access-f4fsz\") pod \"tigera-operator-755d956888-2sn9g\" (UID: \"514dffb5-c294-4238-9bbc-9cea6d258d66\") " pod="tigera-operator/tigera-operator-755d956888-2sn9g" Sep 11 23:31:28.832449 containerd[1515]: time="2025-09-11T23:31:28.832320396Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-2sn9g,Uid:514dffb5-c294-4238-9bbc-9cea6d258d66,Namespace:tigera-operator,Attempt:0,}" Sep 11 23:31:28.850100 containerd[1515]: time="2025-09-11T23:31:28.850053568Z" level=info msg="connecting to shim 7a72ac6737479dc3fc662ed30caeb630cf289ba6cc0f69f7ce0088f080c3803b" address="unix:///run/containerd/s/5d2ef6b8bd1a85d4b8260ce1eaf36a710aaaad39ceff7aba6966fa8aa382e173" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:28.878634 systemd[1]: Started cri-containerd-7a72ac6737479dc3fc662ed30caeb630cf289ba6cc0f69f7ce0088f080c3803b.scope - libcontainer container 7a72ac6737479dc3fc662ed30caeb630cf289ba6cc0f69f7ce0088f080c3803b. Sep 11 23:31:28.896182 kubelet[2668]: E0911 23:31:28.896148 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:28.913591 containerd[1515]: time="2025-09-11T23:31:28.913544455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-2sn9g,Uid:514dffb5-c294-4238-9bbc-9cea6d258d66,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"7a72ac6737479dc3fc662ed30caeb630cf289ba6cc0f69f7ce0088f080c3803b\"" Sep 11 23:31:28.915312 containerd[1515]: time="2025-09-11T23:31:28.915279897Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 11 23:31:30.254048 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1591318729.mount: Deactivated successfully. Sep 11 23:31:30.618515 kubelet[2668]: E0911 23:31:30.617978 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:30.636215 kubelet[2668]: I0911 23:31:30.636094 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-sh9jt" podStartSLOduration=3.636077125 podStartE2EDuration="3.636077125s" podCreationTimestamp="2025-09-11 23:31:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:31:28.908828197 +0000 UTC m=+8.158348357" watchObservedRunningTime="2025-09-11 23:31:30.636077125 +0000 UTC m=+9.885597245" Sep 11 23:31:30.901289 kubelet[2668]: E0911 23:31:30.901179 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:30.998507 containerd[1515]: time="2025-09-11T23:31:30.998021613Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:30.998892 containerd[1515]: time="2025-09-11T23:31:30.998609771Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=22152365" Sep 11 23:31:30.999355 containerd[1515]: time="2025-09-11T23:31:30.999299397Z" level=info msg="ImageCreate event name:\"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:31.001766 containerd[1515]: time="2025-09-11T23:31:31.001617327Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:31.002437 containerd[1515]: time="2025-09-11T23:31:31.002408970Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"22148360\" in 2.087094143s" Sep 11 23:31:31.002571 containerd[1515]: time="2025-09-11T23:31:31.002554167Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\"" Sep 11 23:31:31.016061 containerd[1515]: time="2025-09-11T23:31:31.015991446Z" level=info msg="CreateContainer within sandbox \"7a72ac6737479dc3fc662ed30caeb630cf289ba6cc0f69f7ce0088f080c3803b\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 11 23:31:31.024140 containerd[1515]: time="2025-09-11T23:31:31.023717864Z" level=info msg="Container 0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:31.025062 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3576671463.mount: Deactivated successfully. Sep 11 23:31:31.031288 containerd[1515]: time="2025-09-11T23:31:31.031238989Z" level=info msg="CreateContainer within sandbox \"7a72ac6737479dc3fc662ed30caeb630cf289ba6cc0f69f7ce0088f080c3803b\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda\"" Sep 11 23:31:31.032069 containerd[1515]: time="2025-09-11T23:31:31.032043675Z" level=info msg="StartContainer for \"0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda\"" Sep 11 23:31:31.032962 containerd[1515]: time="2025-09-11T23:31:31.032932462Z" level=info msg="connecting to shim 0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda" address="unix:///run/containerd/s/5d2ef6b8bd1a85d4b8260ce1eaf36a710aaaad39ceff7aba6966fa8aa382e173" protocol=ttrpc version=3 Sep 11 23:31:31.055632 systemd[1]: Started cri-containerd-0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda.scope - libcontainer container 0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda. Sep 11 23:31:31.079597 containerd[1515]: time="2025-09-11T23:31:31.079527468Z" level=info msg="StartContainer for \"0534e86ffbf57136f1e8691e6c6acfce5a925305f65311cb90e905224b7cbdda\" returns successfully" Sep 11 23:31:32.736114 kubelet[2668]: E0911 23:31:32.736072 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:32.747492 kubelet[2668]: I0911 23:31:32.746364 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-2sn9g" podStartSLOduration=2.648587672 podStartE2EDuration="4.74634912s" podCreationTimestamp="2025-09-11 23:31:28 +0000 UTC" firstStartedPulling="2025-09-11 23:31:28.914968123 +0000 UTC m=+8.164488243" lastFinishedPulling="2025-09-11 23:31:31.012729571 +0000 UTC m=+10.262249691" observedRunningTime="2025-09-11 23:31:31.91312631 +0000 UTC m=+11.162646470" watchObservedRunningTime="2025-09-11 23:31:32.74634912 +0000 UTC m=+11.995869240" Sep 11 23:31:33.497142 kubelet[2668]: E0911 23:31:33.497070 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:33.908506 kubelet[2668]: E0911 23:31:33.908382 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:34.543211 update_engine[1492]: I20250911 23:31:34.542527 1492 update_attempter.cc:509] Updating boot flags... Sep 11 23:31:36.256028 sudo[1718]: pam_unix(sudo:session): session closed for user root Sep 11 23:31:36.258306 sshd[1717]: Connection closed by 10.0.0.1 port 46572 Sep 11 23:31:36.258710 sshd-session[1714]: pam_unix(sshd:session): session closed for user core Sep 11 23:31:36.262122 systemd[1]: sshd@6-10.0.0.19:22-10.0.0.1:46572.service: Deactivated successfully. Sep 11 23:31:36.265139 systemd[1]: session-7.scope: Deactivated successfully. Sep 11 23:31:36.265414 systemd[1]: session-7.scope: Consumed 6.835s CPU time, 223.2M memory peak. Sep 11 23:31:36.266958 systemd-logind[1485]: Session 7 logged out. Waiting for processes to exit. Sep 11 23:31:36.268626 systemd-logind[1485]: Removed session 7. Sep 11 23:31:41.412048 systemd[1]: Created slice kubepods-besteffort-pod5d753a1a_802b_4081_b80e_b3cf6781ce15.slice - libcontainer container kubepods-besteffort-pod5d753a1a_802b_4081_b80e_b3cf6781ce15.slice. Sep 11 23:31:41.488455 kubelet[2668]: I0911 23:31:41.488393 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d753a1a-802b-4081-b80e-b3cf6781ce15-tigera-ca-bundle\") pod \"calico-typha-58f86f9779-g9r62\" (UID: \"5d753a1a-802b-4081-b80e-b3cf6781ce15\") " pod="calico-system/calico-typha-58f86f9779-g9r62" Sep 11 23:31:41.488455 kubelet[2668]: I0911 23:31:41.488450 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/5d753a1a-802b-4081-b80e-b3cf6781ce15-typha-certs\") pod \"calico-typha-58f86f9779-g9r62\" (UID: \"5d753a1a-802b-4081-b80e-b3cf6781ce15\") " pod="calico-system/calico-typha-58f86f9779-g9r62" Sep 11 23:31:41.488831 kubelet[2668]: I0911 23:31:41.488542 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmfpb\" (UniqueName: \"kubernetes.io/projected/5d753a1a-802b-4081-b80e-b3cf6781ce15-kube-api-access-lmfpb\") pod \"calico-typha-58f86f9779-g9r62\" (UID: \"5d753a1a-802b-4081-b80e-b3cf6781ce15\") " pod="calico-system/calico-typha-58f86f9779-g9r62" Sep 11 23:31:41.716244 kubelet[2668]: E0911 23:31:41.715779 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:41.716610 containerd[1515]: time="2025-09-11T23:31:41.716543907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-58f86f9779-g9r62,Uid:5d753a1a-802b-4081-b80e-b3cf6781ce15,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:41.740115 systemd[1]: Created slice kubepods-besteffort-pod9dbc6c1f_6cae_46de_81c9_c37c09295b70.slice - libcontainer container kubepods-besteffort-pod9dbc6c1f_6cae_46de_81c9_c37c09295b70.slice. Sep 11 23:31:41.789613 kubelet[2668]: I0911 23:31:41.789575 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-flexvol-driver-host\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.789892 kubelet[2668]: I0911 23:31:41.789794 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-var-run-calico\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.789892 kubelet[2668]: I0911 23:31:41.789823 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-cni-bin-dir\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.789892 kubelet[2668]: I0911 23:31:41.789840 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-var-lib-calico\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.789892 kubelet[2668]: I0911 23:31:41.789860 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4bcp\" (UniqueName: \"kubernetes.io/projected/9dbc6c1f-6cae-46de-81c9-c37c09295b70-kube-api-access-b4bcp\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790572 kubelet[2668]: I0911 23:31:41.789903 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-lib-modules\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790572 kubelet[2668]: I0911 23:31:41.789953 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/9dbc6c1f-6cae-46de-81c9-c37c09295b70-node-certs\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790572 kubelet[2668]: I0911 23:31:41.789970 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-xtables-lock\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790572 kubelet[2668]: I0911 23:31:41.789997 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-cni-log-dir\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790572 kubelet[2668]: I0911 23:31:41.790028 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dbc6c1f-6cae-46de-81c9-c37c09295b70-tigera-ca-bundle\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790732 kubelet[2668]: I0911 23:31:41.790092 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-cni-net-dir\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.790732 kubelet[2668]: I0911 23:31:41.790126 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/9dbc6c1f-6cae-46de-81c9-c37c09295b70-policysync\") pod \"calico-node-mxfjl\" (UID: \"9dbc6c1f-6cae-46de-81c9-c37c09295b70\") " pod="calico-system/calico-node-mxfjl" Sep 11 23:31:41.791487 containerd[1515]: time="2025-09-11T23:31:41.790991397Z" level=info msg="connecting to shim 9055445a17d97088c0cde50c49abe84344d6ac10f6ee976b175372137c235314" address="unix:///run/containerd/s/b7dc757dbd2eb9d06db5299cf588881b1d46d23499627c88476a64610bca6ec2" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:41.840642 systemd[1]: Started cri-containerd-9055445a17d97088c0cde50c49abe84344d6ac10f6ee976b175372137c235314.scope - libcontainer container 9055445a17d97088c0cde50c49abe84344d6ac10f6ee976b175372137c235314. Sep 11 23:31:41.894496 kubelet[2668]: E0911 23:31:41.893351 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:41.894496 kubelet[2668]: W0911 23:31:41.893386 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:41.894496 kubelet[2668]: E0911 23:31:41.893410 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:41.901692 kubelet[2668]: E0911 23:31:41.895070 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:41.901853 kubelet[2668]: W0911 23:31:41.901829 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:41.902075 kubelet[2668]: E0911 23:31:41.901923 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:41.904536 kubelet[2668]: E0911 23:31:41.904363 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:41.904536 kubelet[2668]: W0911 23:31:41.904381 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:41.904536 kubelet[2668]: E0911 23:31:41.904398 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:41.905724 kubelet[2668]: E0911 23:31:41.905705 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:41.905904 kubelet[2668]: W0911 23:31:41.905837 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:41.905904 kubelet[2668]: E0911 23:31:41.905858 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:41.910187 kubelet[2668]: E0911 23:31:41.909303 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:41.910187 kubelet[2668]: W0911 23:31:41.909323 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:41.910187 kubelet[2668]: E0911 23:31:41.909338 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:41.910836 containerd[1515]: time="2025-09-11T23:31:41.910796985Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-58f86f9779-g9r62,Uid:5d753a1a-802b-4081-b80e-b3cf6781ce15,Namespace:calico-system,Attempt:0,} returns sandbox id \"9055445a17d97088c0cde50c49abe84344d6ac10f6ee976b175372137c235314\"" Sep 11 23:31:41.914032 kubelet[2668]: E0911 23:31:41.912693 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:41.915277 containerd[1515]: time="2025-09-11T23:31:41.915236079Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 11 23:31:42.044723 containerd[1515]: time="2025-09-11T23:31:42.044672838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-mxfjl,Uid:9dbc6c1f-6cae-46de-81c9-c37c09295b70,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:42.052502 kubelet[2668]: E0911 23:31:42.052240 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-lgxh5" podUID="d65db1fa-9a8f-403b-a056-75584e2ffe6b" Sep 11 23:31:42.075761 kubelet[2668]: E0911 23:31:42.075472 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.075761 kubelet[2668]: W0911 23:31:42.075637 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.075761 kubelet[2668]: E0911 23:31:42.075663 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.076614 kubelet[2668]: E0911 23:31:42.076533 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.076683 kubelet[2668]: W0911 23:31:42.076558 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.076683 kubelet[2668]: E0911 23:31:42.076649 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.076989 kubelet[2668]: E0911 23:31:42.076844 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.076989 kubelet[2668]: W0911 23:31:42.076858 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.076989 kubelet[2668]: E0911 23:31:42.076869 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.078102 kubelet[2668]: E0911 23:31:42.078077 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.078102 kubelet[2668]: W0911 23:31:42.078102 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.078173 kubelet[2668]: E0911 23:31:42.078115 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.078514 kubelet[2668]: E0911 23:31:42.078492 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.078514 kubelet[2668]: W0911 23:31:42.078509 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.078588 kubelet[2668]: E0911 23:31:42.078521 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.079011 kubelet[2668]: E0911 23:31:42.078900 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.079011 kubelet[2668]: W0911 23:31:42.078915 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.079011 kubelet[2668]: E0911 23:31:42.078926 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.079428 kubelet[2668]: E0911 23:31:42.079396 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.079428 kubelet[2668]: W0911 23:31:42.079415 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.079428 kubelet[2668]: E0911 23:31:42.079427 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.080220 kubelet[2668]: E0911 23:31:42.080065 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.080220 kubelet[2668]: W0911 23:31:42.080084 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.080220 kubelet[2668]: E0911 23:31:42.080101 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.080686 kubelet[2668]: E0911 23:31:42.080646 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.080686 kubelet[2668]: W0911 23:31:42.080664 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.080686 kubelet[2668]: E0911 23:31:42.080676 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.081020 kubelet[2668]: E0911 23:31:42.081001 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.081020 kubelet[2668]: W0911 23:31:42.081012 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.081020 kubelet[2668]: E0911 23:31:42.081022 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.081341 kubelet[2668]: E0911 23:31:42.081316 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.081341 kubelet[2668]: W0911 23:31:42.081333 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.081341 kubelet[2668]: E0911 23:31:42.081343 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.081945 kubelet[2668]: E0911 23:31:42.081924 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.081945 kubelet[2668]: W0911 23:31:42.081941 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.082197 kubelet[2668]: E0911 23:31:42.081953 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.082790 kubelet[2668]: E0911 23:31:42.082763 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.082790 kubelet[2668]: W0911 23:31:42.082781 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.082790 kubelet[2668]: E0911 23:31:42.082792 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.083285 kubelet[2668]: E0911 23:31:42.083260 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.083285 kubelet[2668]: W0911 23:31:42.083277 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.083285 kubelet[2668]: E0911 23:31:42.083289 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.083547 kubelet[2668]: E0911 23:31:42.083442 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.083547 kubelet[2668]: W0911 23:31:42.083456 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.083547 kubelet[2668]: E0911 23:31:42.083475 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.084934 kubelet[2668]: E0911 23:31:42.084548 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.084934 kubelet[2668]: W0911 23:31:42.084590 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.084934 kubelet[2668]: E0911 23:31:42.084605 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.084934 kubelet[2668]: E0911 23:31:42.084906 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.084934 kubelet[2668]: W0911 23:31:42.084918 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.084934 kubelet[2668]: E0911 23:31:42.084929 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.085182 kubelet[2668]: E0911 23:31:42.085152 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.085182 kubelet[2668]: W0911 23:31:42.085166 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.085182 kubelet[2668]: E0911 23:31:42.085176 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.085409 kubelet[2668]: E0911 23:31:42.085392 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.085409 kubelet[2668]: W0911 23:31:42.085406 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.085456 kubelet[2668]: E0911 23:31:42.085415 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.086711 kubelet[2668]: E0911 23:31:42.086352 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.086711 kubelet[2668]: W0911 23:31:42.086368 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.086711 kubelet[2668]: E0911 23:31:42.086456 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.092956 kubelet[2668]: E0911 23:31:42.092934 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.092956 kubelet[2668]: W0911 23:31:42.092951 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.093049 kubelet[2668]: E0911 23:31:42.092967 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.093049 kubelet[2668]: I0911 23:31:42.092989 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4qf4\" (UniqueName: \"kubernetes.io/projected/d65db1fa-9a8f-403b-a056-75584e2ffe6b-kube-api-access-f4qf4\") pod \"csi-node-driver-lgxh5\" (UID: \"d65db1fa-9a8f-403b-a056-75584e2ffe6b\") " pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:42.093138 kubelet[2668]: E0911 23:31:42.093123 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.093167 kubelet[2668]: W0911 23:31:42.093143 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.093167 kubelet[2668]: E0911 23:31:42.093152 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.093203 kubelet[2668]: I0911 23:31:42.093168 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d65db1fa-9a8f-403b-a056-75584e2ffe6b-registration-dir\") pod \"csi-node-driver-lgxh5\" (UID: \"d65db1fa-9a8f-403b-a056-75584e2ffe6b\") " pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:42.093322 kubelet[2668]: E0911 23:31:42.093309 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.093322 kubelet[2668]: W0911 23:31:42.093321 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.093373 kubelet[2668]: E0911 23:31:42.093329 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.093373 kubelet[2668]: I0911 23:31:42.093344 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d65db1fa-9a8f-403b-a056-75584e2ffe6b-kubelet-dir\") pod \"csi-node-driver-lgxh5\" (UID: \"d65db1fa-9a8f-403b-a056-75584e2ffe6b\") " pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:42.093490 kubelet[2668]: E0911 23:31:42.093480 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.093516 kubelet[2668]: W0911 23:31:42.093492 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.093516 kubelet[2668]: E0911 23:31:42.093500 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.093552 kubelet[2668]: I0911 23:31:42.093516 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d65db1fa-9a8f-403b-a056-75584e2ffe6b-socket-dir\") pod \"csi-node-driver-lgxh5\" (UID: \"d65db1fa-9a8f-403b-a056-75584e2ffe6b\") " pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:42.093658 kubelet[2668]: E0911 23:31:42.093645 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.093682 kubelet[2668]: W0911 23:31:42.093657 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.093682 kubelet[2668]: E0911 23:31:42.093665 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.093722 kubelet[2668]: I0911 23:31:42.093694 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/d65db1fa-9a8f-403b-a056-75584e2ffe6b-varrun\") pod \"csi-node-driver-lgxh5\" (UID: \"d65db1fa-9a8f-403b-a056-75584e2ffe6b\") " pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:42.094154 kubelet[2668]: E0911 23:31:42.094073 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.094154 kubelet[2668]: W0911 23:31:42.094138 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.094154 kubelet[2668]: E0911 23:31:42.094151 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.094535 kubelet[2668]: E0911 23:31:42.094378 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.094535 kubelet[2668]: W0911 23:31:42.094391 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.094535 kubelet[2668]: E0911 23:31:42.094401 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.095348 kubelet[2668]: E0911 23:31:42.094614 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.095348 kubelet[2668]: W0911 23:31:42.094623 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.095348 kubelet[2668]: E0911 23:31:42.094643 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.095348 kubelet[2668]: E0911 23:31:42.094824 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.095348 kubelet[2668]: W0911 23:31:42.094832 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.095348 kubelet[2668]: E0911 23:31:42.094840 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.095348 kubelet[2668]: E0911 23:31:42.094986 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.095348 kubelet[2668]: W0911 23:31:42.094994 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.095348 kubelet[2668]: E0911 23:31:42.095001 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.095718 kubelet[2668]: E0911 23:31:42.095692 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.095718 kubelet[2668]: W0911 23:31:42.095707 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.095850 kubelet[2668]: E0911 23:31:42.095828 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.096835 kubelet[2668]: E0911 23:31:42.096675 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.097013 kubelet[2668]: W0911 23:31:42.096982 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.097013 kubelet[2668]: E0911 23:31:42.097011 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.098265 kubelet[2668]: E0911 23:31:42.097636 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.098265 kubelet[2668]: W0911 23:31:42.097654 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.098265 kubelet[2668]: E0911 23:31:42.097931 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.098409 kubelet[2668]: E0911 23:31:42.098388 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.098441 kubelet[2668]: W0911 23:31:42.098406 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.098488 kubelet[2668]: E0911 23:31:42.098439 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.098758 kubelet[2668]: E0911 23:31:42.098738 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.098758 kubelet[2668]: W0911 23:31:42.098750 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.098758 kubelet[2668]: E0911 23:31:42.098759 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.099335 containerd[1515]: time="2025-09-11T23:31:42.099295891Z" level=info msg="connecting to shim ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b" address="unix:///run/containerd/s/de97ef20ae042bfae0358d80ebda5abb21291fb4d0003c51ff50cc3a289a7ac2" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:42.120663 systemd[1]: Started cri-containerd-ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b.scope - libcontainer container ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b. Sep 11 23:31:42.166843 containerd[1515]: time="2025-09-11T23:31:42.166698456Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-mxfjl,Uid:9dbc6c1f-6cae-46de-81c9-c37c09295b70,Namespace:calico-system,Attempt:0,} returns sandbox id \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\"" Sep 11 23:31:42.194873 kubelet[2668]: E0911 23:31:42.194844 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.195023 kubelet[2668]: W0911 23:31:42.194991 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.195023 kubelet[2668]: E0911 23:31:42.195016 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.195377 kubelet[2668]: E0911 23:31:42.195356 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.195377 kubelet[2668]: W0911 23:31:42.195371 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.196331 kubelet[2668]: E0911 23:31:42.195383 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.196331 kubelet[2668]: E0911 23:31:42.195657 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.196331 kubelet[2668]: W0911 23:31:42.195680 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.196331 kubelet[2668]: E0911 23:31:42.195689 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.196331 kubelet[2668]: E0911 23:31:42.196101 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.196331 kubelet[2668]: W0911 23:31:42.196111 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.196331 kubelet[2668]: E0911 23:31:42.196125 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.197058 kubelet[2668]: E0911 23:31:42.197035 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.197058 kubelet[2668]: W0911 23:31:42.197053 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.197180 kubelet[2668]: E0911 23:31:42.197066 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.197715 kubelet[2668]: E0911 23:31:42.197700 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.197777 kubelet[2668]: W0911 23:31:42.197758 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.197777 kubelet[2668]: E0911 23:31:42.197773 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.198015 kubelet[2668]: E0911 23:31:42.198003 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.198015 kubelet[2668]: W0911 23:31:42.198015 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.198117 kubelet[2668]: E0911 23:31:42.198024 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.198271 kubelet[2668]: E0911 23:31:42.198258 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.198271 kubelet[2668]: W0911 23:31:42.198271 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.198324 kubelet[2668]: E0911 23:31:42.198281 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.198701 kubelet[2668]: E0911 23:31:42.198684 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.199543 kubelet[2668]: W0911 23:31:42.199512 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.199543 kubelet[2668]: E0911 23:31:42.199543 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.200680 kubelet[2668]: E0911 23:31:42.200660 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.200680 kubelet[2668]: W0911 23:31:42.200678 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.200856 kubelet[2668]: E0911 23:31:42.200691 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.200883 kubelet[2668]: E0911 23:31:42.200873 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.200883 kubelet[2668]: W0911 23:31:42.200881 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.200983 kubelet[2668]: E0911 23:31:42.200890 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.201057 kubelet[2668]: E0911 23:31:42.201046 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.201561 kubelet[2668]: W0911 23:31:42.201057 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.201561 kubelet[2668]: E0911 23:31:42.201066 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.201561 kubelet[2668]: E0911 23:31:42.201276 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.201561 kubelet[2668]: W0911 23:31:42.201285 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.201561 kubelet[2668]: E0911 23:31:42.201294 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.201561 kubelet[2668]: E0911 23:31:42.201500 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.201561 kubelet[2668]: W0911 23:31:42.201510 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.201561 kubelet[2668]: E0911 23:31:42.201519 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.201778 kubelet[2668]: E0911 23:31:42.201645 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.201778 kubelet[2668]: W0911 23:31:42.201653 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.201778 kubelet[2668]: E0911 23:31:42.201663 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.201778 kubelet[2668]: E0911 23:31:42.201772 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.201778 kubelet[2668]: W0911 23:31:42.201779 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.201888 kubelet[2668]: E0911 23:31:42.201786 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.202166 kubelet[2668]: E0911 23:31:42.201945 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.202166 kubelet[2668]: W0911 23:31:42.201964 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.202166 kubelet[2668]: E0911 23:31:42.201972 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.202166 kubelet[2668]: E0911 23:31:42.202230 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.202166 kubelet[2668]: W0911 23:31:42.202246 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.202166 kubelet[2668]: E0911 23:31:42.202258 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.203231 kubelet[2668]: E0911 23:31:42.202789 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.203231 kubelet[2668]: W0911 23:31:42.202802 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.203231 kubelet[2668]: E0911 23:31:42.202844 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.203873 kubelet[2668]: E0911 23:31:42.203848 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.203873 kubelet[2668]: W0911 23:31:42.203866 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.203882 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.204065 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.204528 kubelet[2668]: W0911 23:31:42.204075 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.204083 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.204270 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.204528 kubelet[2668]: W0911 23:31:42.204278 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.204288 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.204443 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.204528 kubelet[2668]: W0911 23:31:42.204451 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.204528 kubelet[2668]: E0911 23:31:42.204458 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.204759 kubelet[2668]: E0911 23:31:42.204745 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.204782 kubelet[2668]: W0911 23:31:42.204758 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.204782 kubelet[2668]: E0911 23:31:42.204769 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.204989 kubelet[2668]: E0911 23:31:42.204974 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.204989 kubelet[2668]: W0911 23:31:42.204986 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.204989 kubelet[2668]: E0911 23:31:42.204995 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:42.218072 kubelet[2668]: E0911 23:31:42.218030 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:42.218072 kubelet[2668]: W0911 23:31:42.218053 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:42.218072 kubelet[2668]: E0911 23:31:42.218072 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:43.028545 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3271393184.mount: Deactivated successfully. Sep 11 23:31:43.864836 kubelet[2668]: E0911 23:31:43.864781 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-lgxh5" podUID="d65db1fa-9a8f-403b-a056-75584e2ffe6b" Sep 11 23:31:44.820005 containerd[1515]: time="2025-09-11T23:31:44.819956826Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:44.820982 containerd[1515]: time="2025-09-11T23:31:44.820730052Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=33105775" Sep 11 23:31:44.823178 containerd[1515]: time="2025-09-11T23:31:44.823141903Z" level=info msg="ImageCreate event name:\"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:44.826488 containerd[1515]: time="2025-09-11T23:31:44.826434235Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"33105629\" in 2.911151628s" Sep 11 23:31:44.826645 containerd[1515]: time="2025-09-11T23:31:44.826597497Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\"" Sep 11 23:31:44.828033 containerd[1515]: time="2025-09-11T23:31:44.827959284Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 11 23:31:44.830847 containerd[1515]: time="2025-09-11T23:31:44.830799074Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:44.844345 containerd[1515]: time="2025-09-11T23:31:44.844303726Z" level=info msg="CreateContainer within sandbox \"9055445a17d97088c0cde50c49abe84344d6ac10f6ee976b175372137c235314\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 11 23:31:44.853745 containerd[1515]: time="2025-09-11T23:31:44.853681412Z" level=info msg="Container f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:44.873965 containerd[1515]: time="2025-09-11T23:31:44.873918668Z" level=info msg="CreateContainer within sandbox \"9055445a17d97088c0cde50c49abe84344d6ac10f6ee976b175372137c235314\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278\"" Sep 11 23:31:44.875340 containerd[1515]: time="2025-09-11T23:31:44.875073547Z" level=info msg="StartContainer for \"f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278\"" Sep 11 23:31:44.877838 containerd[1515]: time="2025-09-11T23:31:44.877790159Z" level=info msg="connecting to shim f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278" address="unix:///run/containerd/s/b7dc757dbd2eb9d06db5299cf588881b1d46d23499627c88476a64610bca6ec2" protocol=ttrpc version=3 Sep 11 23:31:44.897745 systemd[1]: Started cri-containerd-f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278.scope - libcontainer container f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278. Sep 11 23:31:44.935750 containerd[1515]: time="2025-09-11T23:31:44.935710344Z" level=info msg="StartContainer for \"f6a8077ae1d289cce651e7f5ae47dfcf185b61713904fba4d33fd5815083a278\" returns successfully" Sep 11 23:31:45.864187 kubelet[2668]: E0911 23:31:45.864109 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-lgxh5" podUID="d65db1fa-9a8f-403b-a056-75584e2ffe6b" Sep 11 23:31:45.937202 kubelet[2668]: E0911 23:31:45.937030 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:45.953597 kubelet[2668]: I0911 23:31:45.953507 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-58f86f9779-g9r62" podStartSLOduration=2.040485796 podStartE2EDuration="4.953491688s" podCreationTimestamp="2025-09-11 23:31:41 +0000 UTC" firstStartedPulling="2025-09-11 23:31:41.91479357 +0000 UTC m=+21.164313690" lastFinishedPulling="2025-09-11 23:31:44.827799462 +0000 UTC m=+24.077319582" observedRunningTime="2025-09-11 23:31:45.952036176 +0000 UTC m=+25.201556296" watchObservedRunningTime="2025-09-11 23:31:45.953491688 +0000 UTC m=+25.203011808" Sep 11 23:31:46.015682 kubelet[2668]: E0911 23:31:46.015650 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.015682 kubelet[2668]: W0911 23:31:46.015675 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.015850 kubelet[2668]: E0911 23:31:46.015697 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.015902 kubelet[2668]: E0911 23:31:46.015885 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.015947 kubelet[2668]: W0911 23:31:46.015898 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016002 kubelet[2668]: E0911 23:31:46.015947 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.016124 kubelet[2668]: E0911 23:31:46.016112 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.016124 kubelet[2668]: W0911 23:31:46.016123 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016181 kubelet[2668]: E0911 23:31:46.016132 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.016272 kubelet[2668]: E0911 23:31:46.016261 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.016272 kubelet[2668]: W0911 23:31:46.016272 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016328 kubelet[2668]: E0911 23:31:46.016280 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.016443 kubelet[2668]: E0911 23:31:46.016431 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.016443 kubelet[2668]: W0911 23:31:46.016442 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016523 kubelet[2668]: E0911 23:31:46.016451 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.016605 kubelet[2668]: E0911 23:31:46.016594 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.016605 kubelet[2668]: W0911 23:31:46.016605 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016714 kubelet[2668]: E0911 23:31:46.016616 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.016741 kubelet[2668]: E0911 23:31:46.016730 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.016761 kubelet[2668]: W0911 23:31:46.016739 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016761 kubelet[2668]: E0911 23:31:46.016747 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.016873 kubelet[2668]: E0911 23:31:46.016863 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.016873 kubelet[2668]: W0911 23:31:46.016872 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.016936 kubelet[2668]: E0911 23:31:46.016880 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017013 kubelet[2668]: E0911 23:31:46.017002 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017013 kubelet[2668]: W0911 23:31:46.017012 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017074 kubelet[2668]: E0911 23:31:46.017021 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017146 kubelet[2668]: E0911 23:31:46.017137 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017146 kubelet[2668]: W0911 23:31:46.017146 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017206 kubelet[2668]: E0911 23:31:46.017153 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017272 kubelet[2668]: E0911 23:31:46.017262 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017272 kubelet[2668]: W0911 23:31:46.017271 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017329 kubelet[2668]: E0911 23:31:46.017279 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017407 kubelet[2668]: E0911 23:31:46.017397 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017407 kubelet[2668]: W0911 23:31:46.017407 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017482 kubelet[2668]: E0911 23:31:46.017414 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017567 kubelet[2668]: E0911 23:31:46.017555 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017567 kubelet[2668]: W0911 23:31:46.017566 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017626 kubelet[2668]: E0911 23:31:46.017574 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017696 kubelet[2668]: E0911 23:31:46.017686 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017696 kubelet[2668]: W0911 23:31:46.017695 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017758 kubelet[2668]: E0911 23:31:46.017702 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.017832 kubelet[2668]: E0911 23:31:46.017821 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.017859 kubelet[2668]: W0911 23:31:46.017833 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.017859 kubelet[2668]: E0911 23:31:46.017841 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.023490 kubelet[2668]: E0911 23:31:46.023323 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.023490 kubelet[2668]: W0911 23:31:46.023352 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.023490 kubelet[2668]: E0911 23:31:46.023371 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.023621 kubelet[2668]: E0911 23:31:46.023594 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.023621 kubelet[2668]: W0911 23:31:46.023602 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.023621 kubelet[2668]: E0911 23:31:46.023611 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.023845 kubelet[2668]: E0911 23:31:46.023831 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.023845 kubelet[2668]: W0911 23:31:46.023841 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.023911 kubelet[2668]: E0911 23:31:46.023850 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.024054 kubelet[2668]: E0911 23:31:46.024028 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.024054 kubelet[2668]: W0911 23:31:46.024040 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.024054 kubelet[2668]: E0911 23:31:46.024048 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.024210 kubelet[2668]: E0911 23:31:46.024194 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.024210 kubelet[2668]: W0911 23:31:46.024205 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.024263 kubelet[2668]: E0911 23:31:46.024213 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.024358 kubelet[2668]: E0911 23:31:46.024338 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.024358 kubelet[2668]: W0911 23:31:46.024355 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.024401 kubelet[2668]: E0911 23:31:46.024365 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.024595 kubelet[2668]: E0911 23:31:46.024571 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.024595 kubelet[2668]: W0911 23:31:46.024585 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.024595 kubelet[2668]: E0911 23:31:46.024594 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.024858 kubelet[2668]: E0911 23:31:46.024831 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.024858 kubelet[2668]: W0911 23:31:46.024848 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.024906 kubelet[2668]: E0911 23:31:46.024867 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.025038 kubelet[2668]: E0911 23:31:46.025026 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.025064 kubelet[2668]: W0911 23:31:46.025037 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.025064 kubelet[2668]: E0911 23:31:46.025046 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.025193 kubelet[2668]: E0911 23:31:46.025182 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.025217 kubelet[2668]: W0911 23:31:46.025193 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.025217 kubelet[2668]: E0911 23:31:46.025201 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.025335 kubelet[2668]: E0911 23:31:46.025325 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.025367 kubelet[2668]: W0911 23:31:46.025335 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.025367 kubelet[2668]: E0911 23:31:46.025352 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.025543 kubelet[2668]: E0911 23:31:46.025532 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.025567 kubelet[2668]: W0911 23:31:46.025542 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.025567 kubelet[2668]: E0911 23:31:46.025550 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.025712 kubelet[2668]: E0911 23:31:46.025701 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.025712 kubelet[2668]: W0911 23:31:46.025711 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.025754 kubelet[2668]: E0911 23:31:46.025719 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.025960 kubelet[2668]: E0911 23:31:46.025941 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.025981 kubelet[2668]: W0911 23:31:46.025959 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.025981 kubelet[2668]: E0911 23:31:46.025972 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.026102 kubelet[2668]: E0911 23:31:46.026093 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.026127 kubelet[2668]: W0911 23:31:46.026102 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.026127 kubelet[2668]: E0911 23:31:46.026110 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.026267 kubelet[2668]: E0911 23:31:46.026257 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.026288 kubelet[2668]: W0911 23:31:46.026267 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.026288 kubelet[2668]: E0911 23:31:46.026276 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.026568 kubelet[2668]: E0911 23:31:46.026550 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.026600 kubelet[2668]: W0911 23:31:46.026568 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.026600 kubelet[2668]: E0911 23:31:46.026581 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.026783 kubelet[2668]: E0911 23:31:46.026768 2668 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 23:31:46.026810 kubelet[2668]: W0911 23:31:46.026782 2668 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 23:31:46.026810 kubelet[2668]: E0911 23:31:46.026792 2668 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 23:31:46.214824 containerd[1515]: time="2025-09-11T23:31:46.214709948Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:46.215691 containerd[1515]: time="2025-09-11T23:31:46.215480925Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4266814" Sep 11 23:31:46.216693 containerd[1515]: time="2025-09-11T23:31:46.216561102Z" level=info msg="ImageCreate event name:\"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:46.218398 containerd[1515]: time="2025-09-11T23:31:46.218358689Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:46.219125 containerd[1515]: time="2025-09-11T23:31:46.219098462Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5636015\" in 1.390971715s" Sep 11 23:31:46.219182 containerd[1515]: time="2025-09-11T23:31:46.219131186Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\"" Sep 11 23:31:46.223403 containerd[1515]: time="2025-09-11T23:31:46.223356359Z" level=info msg="CreateContainer within sandbox \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 11 23:31:46.230823 containerd[1515]: time="2025-09-11T23:31:46.230777776Z" level=info msg="Container 84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:46.239117 containerd[1515]: time="2025-09-11T23:31:46.239069983Z" level=info msg="CreateContainer within sandbox \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\"" Sep 11 23:31:46.239611 containerd[1515]: time="2025-09-11T23:31:46.239571726Z" level=info msg="StartContainer for \"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\"" Sep 11 23:31:46.240972 containerd[1515]: time="2025-09-11T23:31:46.240935378Z" level=info msg="connecting to shim 84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622" address="unix:///run/containerd/s/de97ef20ae042bfae0358d80ebda5abb21291fb4d0003c51ff50cc3a289a7ac2" protocol=ttrpc version=3 Sep 11 23:31:46.262634 systemd[1]: Started cri-containerd-84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622.scope - libcontainer container 84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622. Sep 11 23:31:46.298654 containerd[1515]: time="2025-09-11T23:31:46.298356986Z" level=info msg="StartContainer for \"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\" returns successfully" Sep 11 23:31:46.315423 systemd[1]: cri-containerd-84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622.scope: Deactivated successfully. Sep 11 23:31:46.335666 containerd[1515]: time="2025-09-11T23:31:46.335616409Z" level=info msg="received exit event container_id:\"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\" id:\"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\" pid:3383 exited_at:{seconds:1757633506 nanos:332044838}" Sep 11 23:31:46.335791 containerd[1515]: time="2025-09-11T23:31:46.335703540Z" level=info msg="TaskExit event in podsandbox handler container_id:\"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\" id:\"84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622\" pid:3383 exited_at:{seconds:1757633506 nanos:332044838}" Sep 11 23:31:46.377864 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-84ae6585b5efa42822e82e15e0b51c2428c5bb421446544f2acaacfa94f0c622-rootfs.mount: Deactivated successfully. Sep 11 23:31:46.941685 containerd[1515]: time="2025-09-11T23:31:46.941630821Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 11 23:31:47.822301 kubelet[2668]: I0911 23:31:47.822167 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:31:47.822734 kubelet[2668]: E0911 23:31:47.822710 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:47.864541 kubelet[2668]: E0911 23:31:47.864489 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-lgxh5" podUID="d65db1fa-9a8f-403b-a056-75584e2ffe6b" Sep 11 23:31:47.943035 kubelet[2668]: E0911 23:31:47.942667 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:49.865183 kubelet[2668]: E0911 23:31:49.864800 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-lgxh5" podUID="d65db1fa-9a8f-403b-a056-75584e2ffe6b" Sep 11 23:31:50.251848 containerd[1515]: time="2025-09-11T23:31:50.251795484Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:50.252478 containerd[1515]: time="2025-09-11T23:31:50.252345743Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=65913477" Sep 11 23:31:50.253408 containerd[1515]: time="2025-09-11T23:31:50.253380735Z" level=info msg="ImageCreate event name:\"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:50.256504 containerd[1515]: time="2025-09-11T23:31:50.256444747Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:50.257136 containerd[1515]: time="2025-09-11T23:31:50.257104138Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"67282718\" in 3.315250049s" Sep 11 23:31:50.257203 containerd[1515]: time="2025-09-11T23:31:50.257142182Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\"" Sep 11 23:31:50.261601 containerd[1515]: time="2025-09-11T23:31:50.261556780Z" level=info msg="CreateContainer within sandbox \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 11 23:31:50.270896 containerd[1515]: time="2025-09-11T23:31:50.269781829Z" level=info msg="Container 3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:50.278950 containerd[1515]: time="2025-09-11T23:31:50.278903335Z" level=info msg="CreateContainer within sandbox \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\"" Sep 11 23:31:50.279542 containerd[1515]: time="2025-09-11T23:31:50.279512361Z" level=info msg="StartContainer for \"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\"" Sep 11 23:31:50.281149 containerd[1515]: time="2025-09-11T23:31:50.281117215Z" level=info msg="connecting to shim 3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3" address="unix:///run/containerd/s/de97ef20ae042bfae0358d80ebda5abb21291fb4d0003c51ff50cc3a289a7ac2" protocol=ttrpc version=3 Sep 11 23:31:50.306685 systemd[1]: Started cri-containerd-3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3.scope - libcontainer container 3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3. Sep 11 23:31:50.359504 containerd[1515]: time="2025-09-11T23:31:50.359448366Z" level=info msg="StartContainer for \"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\" returns successfully" Sep 11 23:31:50.845634 systemd[1]: cri-containerd-3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3.scope: Deactivated successfully. Sep 11 23:31:50.845958 systemd[1]: cri-containerd-3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3.scope: Consumed 438ms CPU time, 176.9M memory peak, 2.9M read from disk, 165.8M written to disk. Sep 11 23:31:50.847219 containerd[1515]: time="2025-09-11T23:31:50.846264811Z" level=info msg="received exit event container_id:\"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\" id:\"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\" pid:3447 exited_at:{seconds:1757633510 nanos:845899252}" Sep 11 23:31:50.847319 containerd[1515]: time="2025-09-11T23:31:50.846318257Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\" id:\"3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3\" pid:3447 exited_at:{seconds:1757633510 nanos:845899252}" Sep 11 23:31:50.873054 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3a1c512d2c5e5c824537f2e3342a081bd6bc4a1fcc38c7e8094a162059bdf0c3-rootfs.mount: Deactivated successfully. Sep 11 23:31:50.954524 kubelet[2668]: I0911 23:31:50.953649 2668 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 11 23:31:51.085666 systemd[1]: Created slice kubepods-besteffort-pod341f9b5c_0ff7_44c6_9add_ef9f11e8780b.slice - libcontainer container kubepods-besteffort-pod341f9b5c_0ff7_44c6_9add_ef9f11e8780b.slice. Sep 11 23:31:51.104410 systemd[1]: Created slice kubepods-burstable-pod32dd7b29_d990_4d4c_8903_e51a4239fdc5.slice - libcontainer container kubepods-burstable-pod32dd7b29_d990_4d4c_8903_e51a4239fdc5.slice. Sep 11 23:31:51.167869 systemd[1]: Created slice kubepods-besteffort-podc0e77bb6_9a42_4b41_98bb_3d6bcd21104b.slice - libcontainer container kubepods-besteffort-podc0e77bb6_9a42_4b41_98bb_3d6bcd21104b.slice. Sep 11 23:31:51.171801 systemd[1]: Created slice kubepods-besteffort-podcd9cb3e9_4e82_450d_850d_add7a5cdeee3.slice - libcontainer container kubepods-besteffort-podcd9cb3e9_4e82_450d_850d_add7a5cdeee3.slice. Sep 11 23:31:51.192707 systemd[1]: Created slice kubepods-besteffort-podcb91f87b_09d0_4126_b7eb_549a67e8f409.slice - libcontainer container kubepods-besteffort-podcb91f87b_09d0_4126_b7eb_549a67e8f409.slice. Sep 11 23:31:51.254327 systemd[1]: Created slice kubepods-besteffort-pod6252d405_10a5_459a_ab90_82f9637e1a5a.slice - libcontainer container kubepods-besteffort-pod6252d405_10a5_459a_ab90_82f9637e1a5a.slice. Sep 11 23:31:51.257459 kubelet[2668]: I0911 23:31:51.257381 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32dd7b29-d990-4d4c-8903-e51a4239fdc5-config-volume\") pod \"coredns-674b8bbfcf-d725f\" (UID: \"32dd7b29-d990-4d4c-8903-e51a4239fdc5\") " pod="kube-system/coredns-674b8bbfcf-d725f" Sep 11 23:31:51.258078 kubelet[2668]: I0911 23:31:51.258055 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvhws\" (UniqueName: \"kubernetes.io/projected/341f9b5c-0ff7-44c6-9add-ef9f11e8780b-kube-api-access-nvhws\") pod \"calico-apiserver-56f49b666f-ckb7g\" (UID: \"341f9b5c-0ff7-44c6-9add-ef9f11e8780b\") " pod="calico-apiserver/calico-apiserver-56f49b666f-ckb7g" Sep 11 23:31:51.258330 kubelet[2668]: I0911 23:31:51.258315 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c0e77bb6-9a42-4b41-98bb-3d6bcd21104b-tigera-ca-bundle\") pod \"calico-kube-controllers-cb484f97b-jkcgr\" (UID: \"c0e77bb6-9a42-4b41-98bb-3d6bcd21104b\") " pod="calico-system/calico-kube-controllers-cb484f97b-jkcgr" Sep 11 23:31:51.258555 kubelet[2668]: I0911 23:31:51.258397 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msklq\" (UniqueName: \"kubernetes.io/projected/c0e77bb6-9a42-4b41-98bb-3d6bcd21104b-kube-api-access-msklq\") pod \"calico-kube-controllers-cb484f97b-jkcgr\" (UID: \"c0e77bb6-9a42-4b41-98bb-3d6bcd21104b\") " pod="calico-system/calico-kube-controllers-cb484f97b-jkcgr" Sep 11 23:31:51.258555 kubelet[2668]: I0911 23:31:51.258451 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h9jb\" (UniqueName: \"kubernetes.io/projected/32dd7b29-d990-4d4c-8903-e51a4239fdc5-kube-api-access-7h9jb\") pod \"coredns-674b8bbfcf-d725f\" (UID: \"32dd7b29-d990-4d4c-8903-e51a4239fdc5\") " pod="kube-system/coredns-674b8bbfcf-d725f" Sep 11 23:31:51.258555 kubelet[2668]: I0911 23:31:51.258521 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/cd9cb3e9-4e82-450d-850d-add7a5cdeee3-calico-apiserver-certs\") pod \"calico-apiserver-56f49b666f-zjzkc\" (UID: \"cd9cb3e9-4e82-450d-850d-add7a5cdeee3\") " pod="calico-apiserver/calico-apiserver-56f49b666f-zjzkc" Sep 11 23:31:51.258734 kubelet[2668]: I0911 23:31:51.258598 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvdjw\" (UniqueName: \"kubernetes.io/projected/cd9cb3e9-4e82-450d-850d-add7a5cdeee3-kube-api-access-rvdjw\") pod \"calico-apiserver-56f49b666f-zjzkc\" (UID: \"cd9cb3e9-4e82-450d-850d-add7a5cdeee3\") " pod="calico-apiserver/calico-apiserver-56f49b666f-zjzkc" Sep 11 23:31:51.258734 kubelet[2668]: I0911 23:31:51.258619 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/341f9b5c-0ff7-44c6-9add-ef9f11e8780b-calico-apiserver-certs\") pod \"calico-apiserver-56f49b666f-ckb7g\" (UID: \"341f9b5c-0ff7-44c6-9add-ef9f11e8780b\") " pod="calico-apiserver/calico-apiserver-56f49b666f-ckb7g" Sep 11 23:31:51.277066 systemd[1]: Created slice kubepods-burstable-podaf6fbb68_721d_41a3_86cc_0d3211856c13.slice - libcontainer container kubepods-burstable-podaf6fbb68_721d_41a3_86cc_0d3211856c13.slice. Sep 11 23:31:51.359921 kubelet[2668]: I0911 23:31:51.359529 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2r2b\" (UniqueName: \"kubernetes.io/projected/6252d405-10a5-459a-ab90-82f9637e1a5a-kube-api-access-r2r2b\") pod \"whisker-7cdb87984d-s8t98\" (UID: \"6252d405-10a5-459a-ab90-82f9637e1a5a\") " pod="calico-system/whisker-7cdb87984d-s8t98" Sep 11 23:31:51.359921 kubelet[2668]: I0911 23:31:51.359576 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-backend-key-pair\") pod \"whisker-7cdb87984d-s8t98\" (UID: \"6252d405-10a5-459a-ab90-82f9637e1a5a\") " pod="calico-system/whisker-7cdb87984d-s8t98" Sep 11 23:31:51.359921 kubelet[2668]: I0911 23:31:51.359595 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/cb91f87b-09d0-4126-b7eb-549a67e8f409-goldmane-key-pair\") pod \"goldmane-54d579b49d-s852n\" (UID: \"cb91f87b-09d0-4126-b7eb-549a67e8f409\") " pod="calico-system/goldmane-54d579b49d-s852n" Sep 11 23:31:51.359921 kubelet[2668]: I0911 23:31:51.359623 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-ca-bundle\") pod \"whisker-7cdb87984d-s8t98\" (UID: \"6252d405-10a5-459a-ab90-82f9637e1a5a\") " pod="calico-system/whisker-7cdb87984d-s8t98" Sep 11 23:31:51.359921 kubelet[2668]: I0911 23:31:51.359642 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb91f87b-09d0-4126-b7eb-549a67e8f409-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-s852n\" (UID: \"cb91f87b-09d0-4126-b7eb-549a67e8f409\") " pod="calico-system/goldmane-54d579b49d-s852n" Sep 11 23:31:51.360136 kubelet[2668]: I0911 23:31:51.359658 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkggw\" (UniqueName: \"kubernetes.io/projected/cb91f87b-09d0-4126-b7eb-549a67e8f409-kube-api-access-pkggw\") pod \"goldmane-54d579b49d-s852n\" (UID: \"cb91f87b-09d0-4126-b7eb-549a67e8f409\") " pod="calico-system/goldmane-54d579b49d-s852n" Sep 11 23:31:51.360136 kubelet[2668]: I0911 23:31:51.359682 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb91f87b-09d0-4126-b7eb-549a67e8f409-config\") pod \"goldmane-54d579b49d-s852n\" (UID: \"cb91f87b-09d0-4126-b7eb-549a67e8f409\") " pod="calico-system/goldmane-54d579b49d-s852n" Sep 11 23:31:51.389773 containerd[1515]: time="2025-09-11T23:31:51.389719172Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-ckb7g,Uid:341f9b5c-0ff7-44c6-9add-ef9f11e8780b,Namespace:calico-apiserver,Attempt:0,}" Sep 11 23:31:51.407342 kubelet[2668]: E0911 23:31:51.407031 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:51.407884 containerd[1515]: time="2025-09-11T23:31:51.407804699Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-d725f,Uid:32dd7b29-d990-4d4c-8903-e51a4239fdc5,Namespace:kube-system,Attempt:0,}" Sep 11 23:31:51.460604 kubelet[2668]: I0911 23:31:51.460552 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blqqc\" (UniqueName: \"kubernetes.io/projected/af6fbb68-721d-41a3-86cc-0d3211856c13-kube-api-access-blqqc\") pod \"coredns-674b8bbfcf-n5ml6\" (UID: \"af6fbb68-721d-41a3-86cc-0d3211856c13\") " pod="kube-system/coredns-674b8bbfcf-n5ml6" Sep 11 23:31:51.460721 kubelet[2668]: I0911 23:31:51.460625 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af6fbb68-721d-41a3-86cc-0d3211856c13-config-volume\") pod \"coredns-674b8bbfcf-n5ml6\" (UID: \"af6fbb68-721d-41a3-86cc-0d3211856c13\") " pod="kube-system/coredns-674b8bbfcf-n5ml6" Sep 11 23:31:51.471743 containerd[1515]: time="2025-09-11T23:31:51.471676641Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-cb484f97b-jkcgr,Uid:c0e77bb6-9a42-4b41-98bb-3d6bcd21104b,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:51.474331 containerd[1515]: time="2025-09-11T23:31:51.474221666Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-zjzkc,Uid:cd9cb3e9-4e82-450d-850d-add7a5cdeee3,Namespace:calico-apiserver,Attempt:0,}" Sep 11 23:31:51.496235 containerd[1515]: time="2025-09-11T23:31:51.496073706Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-s852n,Uid:cb91f87b-09d0-4126-b7eb-549a67e8f409,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:51.506249 containerd[1515]: time="2025-09-11T23:31:51.506205123Z" level=error msg="Failed to destroy network for sandbox \"4ca107bc983684ae2274395dd12c3e018b034c1a24cd25ca9cb750a300a0c2c7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.509131 containerd[1515]: time="2025-09-11T23:31:51.508948849Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-ckb7g,Uid:341f9b5c-0ff7-44c6-9add-ef9f11e8780b,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ca107bc983684ae2274395dd12c3e018b034c1a24cd25ca9cb750a300a0c2c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.520043 kubelet[2668]: E0911 23:31:51.519973 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ca107bc983684ae2274395dd12c3e018b034c1a24cd25ca9cb750a300a0c2c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.520172 kubelet[2668]: E0911 23:31:51.520068 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ca107bc983684ae2274395dd12c3e018b034c1a24cd25ca9cb750a300a0c2c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-56f49b666f-ckb7g" Sep 11 23:31:51.520172 kubelet[2668]: E0911 23:31:51.520094 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ca107bc983684ae2274395dd12c3e018b034c1a24cd25ca9cb750a300a0c2c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-56f49b666f-ckb7g" Sep 11 23:31:51.520862 kubelet[2668]: E0911 23:31:51.520824 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-56f49b666f-ckb7g_calico-apiserver(341f9b5c-0ff7-44c6-9add-ef9f11e8780b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-56f49b666f-ckb7g_calico-apiserver(341f9b5c-0ff7-44c6-9add-ef9f11e8780b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4ca107bc983684ae2274395dd12c3e018b034c1a24cd25ca9cb750a300a0c2c7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-56f49b666f-ckb7g" podUID="341f9b5c-0ff7-44c6-9add-ef9f11e8780b" Sep 11 23:31:51.529658 containerd[1515]: time="2025-09-11T23:31:51.529599483Z" level=error msg="Failed to destroy network for sandbox \"221b312747402300a6595569696c2d136e636127022a6ab7c3edd9393929ed29\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.531530 containerd[1515]: time="2025-09-11T23:31:51.531451276Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-d725f,Uid:32dd7b29-d990-4d4c-8903-e51a4239fdc5,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"221b312747402300a6595569696c2d136e636127022a6ab7c3edd9393929ed29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.532158 kubelet[2668]: E0911 23:31:51.531682 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"221b312747402300a6595569696c2d136e636127022a6ab7c3edd9393929ed29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.532158 kubelet[2668]: E0911 23:31:51.531733 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"221b312747402300a6595569696c2d136e636127022a6ab7c3edd9393929ed29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-d725f" Sep 11 23:31:51.532158 kubelet[2668]: E0911 23:31:51.531754 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"221b312747402300a6595569696c2d136e636127022a6ab7c3edd9393929ed29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-d725f" Sep 11 23:31:51.532266 kubelet[2668]: E0911 23:31:51.531799 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-d725f_kube-system(32dd7b29-d990-4d4c-8903-e51a4239fdc5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-d725f_kube-system(32dd7b29-d990-4d4c-8903-e51a4239fdc5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"221b312747402300a6595569696c2d136e636127022a6ab7c3edd9393929ed29\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-d725f" podUID="32dd7b29-d990-4d4c-8903-e51a4239fdc5" Sep 11 23:31:51.550581 containerd[1515]: time="2025-09-11T23:31:51.548598424Z" level=error msg="Failed to destroy network for sandbox \"59cc0bb188e74d3a2788b4e9db0d6df72c3ae854c4a6b1748d4698acff9ce3fc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.551801 containerd[1515]: time="2025-09-11T23:31:51.551761554Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-cb484f97b-jkcgr,Uid:c0e77bb6-9a42-4b41-98bb-3d6bcd21104b,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"59cc0bb188e74d3a2788b4e9db0d6df72c3ae854c4a6b1748d4698acff9ce3fc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.552057 kubelet[2668]: E0911 23:31:51.552002 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59cc0bb188e74d3a2788b4e9db0d6df72c3ae854c4a6b1748d4698acff9ce3fc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.552128 kubelet[2668]: E0911 23:31:51.552076 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59cc0bb188e74d3a2788b4e9db0d6df72c3ae854c4a6b1748d4698acff9ce3fc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-cb484f97b-jkcgr" Sep 11 23:31:51.552128 kubelet[2668]: E0911 23:31:51.552097 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59cc0bb188e74d3a2788b4e9db0d6df72c3ae854c4a6b1748d4698acff9ce3fc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-cb484f97b-jkcgr" Sep 11 23:31:51.552187 kubelet[2668]: E0911 23:31:51.552148 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-cb484f97b-jkcgr_calico-system(c0e77bb6-9a42-4b41-98bb-3d6bcd21104b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-cb484f97b-jkcgr_calico-system(c0e77bb6-9a42-4b41-98bb-3d6bcd21104b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"59cc0bb188e74d3a2788b4e9db0d6df72c3ae854c4a6b1748d4698acff9ce3fc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-cb484f97b-jkcgr" podUID="c0e77bb6-9a42-4b41-98bb-3d6bcd21104b" Sep 11 23:31:51.558647 containerd[1515]: time="2025-09-11T23:31:51.558604268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7cdb87984d-s8t98,Uid:6252d405-10a5-459a-ab90-82f9637e1a5a,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:51.560449 containerd[1515]: time="2025-09-11T23:31:51.560400015Z" level=error msg="Failed to destroy network for sandbox \"7410e1d0b22fd4797a9a572f9d37e2f277f8280295f79dd1aaf70e9dd666739c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.565724 containerd[1515]: time="2025-09-11T23:31:51.565645442Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-zjzkc,Uid:cd9cb3e9-4e82-450d-850d-add7a5cdeee3,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7410e1d0b22fd4797a9a572f9d37e2f277f8280295f79dd1aaf70e9dd666739c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.565977 kubelet[2668]: E0911 23:31:51.565928 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7410e1d0b22fd4797a9a572f9d37e2f277f8280295f79dd1aaf70e9dd666739c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.566066 kubelet[2668]: E0911 23:31:51.566002 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7410e1d0b22fd4797a9a572f9d37e2f277f8280295f79dd1aaf70e9dd666739c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-56f49b666f-zjzkc" Sep 11 23:31:51.566066 kubelet[2668]: E0911 23:31:51.566023 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7410e1d0b22fd4797a9a572f9d37e2f277f8280295f79dd1aaf70e9dd666739c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-56f49b666f-zjzkc" Sep 11 23:31:51.566251 kubelet[2668]: E0911 23:31:51.566071 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-56f49b666f-zjzkc_calico-apiserver(cd9cb3e9-4e82-450d-850d-add7a5cdeee3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-56f49b666f-zjzkc_calico-apiserver(cd9cb3e9-4e82-450d-850d-add7a5cdeee3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7410e1d0b22fd4797a9a572f9d37e2f277f8280295f79dd1aaf70e9dd666739c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-56f49b666f-zjzkc" podUID="cd9cb3e9-4e82-450d-850d-add7a5cdeee3" Sep 11 23:31:51.571067 containerd[1515]: time="2025-09-11T23:31:51.571013242Z" level=error msg="Failed to destroy network for sandbox \"0aefcb0beb6fba1be40937e84ebf0c9e5b675f8f4a813d8359c2cc4195546db9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.573235 containerd[1515]: time="2025-09-11T23:31:51.573014011Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-s852n,Uid:cb91f87b-09d0-4126-b7eb-549a67e8f409,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0aefcb0beb6fba1be40937e84ebf0c9e5b675f8f4a813d8359c2cc4195546db9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.573344 kubelet[2668]: E0911 23:31:51.573233 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0aefcb0beb6fba1be40937e84ebf0c9e5b675f8f4a813d8359c2cc4195546db9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.573344 kubelet[2668]: E0911 23:31:51.573287 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0aefcb0beb6fba1be40937e84ebf0c9e5b675f8f4a813d8359c2cc4195546db9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-s852n" Sep 11 23:31:51.573344 kubelet[2668]: E0911 23:31:51.573306 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0aefcb0beb6fba1be40937e84ebf0c9e5b675f8f4a813d8359c2cc4195546db9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-s852n" Sep 11 23:31:51.573456 kubelet[2668]: E0911 23:31:51.573350 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-s852n_calico-system(cb91f87b-09d0-4126-b7eb-549a67e8f409)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-s852n_calico-system(cb91f87b-09d0-4126-b7eb-549a67e8f409)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0aefcb0beb6fba1be40937e84ebf0c9e5b675f8f4a813d8359c2cc4195546db9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-s852n" podUID="cb91f87b-09d0-4126-b7eb-549a67e8f409" Sep 11 23:31:51.580510 kubelet[2668]: E0911 23:31:51.580440 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:31:51.581248 containerd[1515]: time="2025-09-11T23:31:51.581200945Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n5ml6,Uid:af6fbb68-721d-41a3-86cc-0d3211856c13,Namespace:kube-system,Attempt:0,}" Sep 11 23:31:51.621816 containerd[1515]: time="2025-09-11T23:31:51.621546313Z" level=error msg="Failed to destroy network for sandbox \"38f38621a5c1c8fc4dc71b3358f22aaa0a2f72a4bc0303afe10a6246a1e5c590\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.630853 containerd[1515]: time="2025-09-11T23:31:51.630788317Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7cdb87984d-s8t98,Uid:6252d405-10a5-459a-ab90-82f9637e1a5a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f38621a5c1c8fc4dc71b3358f22aaa0a2f72a4bc0303afe10a6246a1e5c590\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.631204 kubelet[2668]: E0911 23:31:51.631069 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f38621a5c1c8fc4dc71b3358f22aaa0a2f72a4bc0303afe10a6246a1e5c590\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.631263 kubelet[2668]: E0911 23:31:51.631229 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f38621a5c1c8fc4dc71b3358f22aaa0a2f72a4bc0303afe10a6246a1e5c590\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7cdb87984d-s8t98" Sep 11 23:31:51.631263 kubelet[2668]: E0911 23:31:51.631251 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f38621a5c1c8fc4dc71b3358f22aaa0a2f72a4bc0303afe10a6246a1e5c590\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7cdb87984d-s8t98" Sep 11 23:31:51.631616 kubelet[2668]: E0911 23:31:51.631305 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7cdb87984d-s8t98_calico-system(6252d405-10a5-459a-ab90-82f9637e1a5a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7cdb87984d-s8t98_calico-system(6252d405-10a5-459a-ab90-82f9637e1a5a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"38f38621a5c1c8fc4dc71b3358f22aaa0a2f72a4bc0303afe10a6246a1e5c590\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7cdb87984d-s8t98" podUID="6252d405-10a5-459a-ab90-82f9637e1a5a" Sep 11 23:31:51.636381 containerd[1515]: time="2025-09-11T23:31:51.636334296Z" level=error msg="Failed to destroy network for sandbox \"5654045e83024e90fe3cd6b2688c4be0b0cd17141a1a42acd2d4a4c578dc046f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.639132 containerd[1515]: time="2025-09-11T23:31:51.639086063Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n5ml6,Uid:af6fbb68-721d-41a3-86cc-0d3211856c13,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5654045e83024e90fe3cd6b2688c4be0b0cd17141a1a42acd2d4a4c578dc046f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.639538 kubelet[2668]: E0911 23:31:51.639491 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5654045e83024e90fe3cd6b2688c4be0b0cd17141a1a42acd2d4a4c578dc046f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.639613 kubelet[2668]: E0911 23:31:51.639561 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5654045e83024e90fe3cd6b2688c4be0b0cd17141a1a42acd2d4a4c578dc046f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-n5ml6" Sep 11 23:31:51.639613 kubelet[2668]: E0911 23:31:51.639584 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5654045e83024e90fe3cd6b2688c4be0b0cd17141a1a42acd2d4a4c578dc046f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-n5ml6" Sep 11 23:31:51.639668 kubelet[2668]: E0911 23:31:51.639631 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-n5ml6_kube-system(af6fbb68-721d-41a3-86cc-0d3211856c13)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-n5ml6_kube-system(af6fbb68-721d-41a3-86cc-0d3211856c13)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5654045e83024e90fe3cd6b2688c4be0b0cd17141a1a42acd2d4a4c578dc046f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-n5ml6" podUID="af6fbb68-721d-41a3-86cc-0d3211856c13" Sep 11 23:31:51.874724 systemd[1]: Created slice kubepods-besteffort-podd65db1fa_9a8f_403b_a056_75584e2ffe6b.slice - libcontainer container kubepods-besteffort-podd65db1fa_9a8f_403b_a056_75584e2ffe6b.slice. Sep 11 23:31:51.877413 containerd[1515]: time="2025-09-11T23:31:51.877342434Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-lgxh5,Uid:d65db1fa-9a8f-403b-a056-75584e2ffe6b,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:51.925848 containerd[1515]: time="2025-09-11T23:31:51.925799649Z" level=error msg="Failed to destroy network for sandbox \"355dd9963bd54819a3a09df0f1f4fe81d0137ef9f582c3c1415cc444183c6e85\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.927456 containerd[1515]: time="2025-09-11T23:31:51.927389535Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-lgxh5,Uid:d65db1fa-9a8f-403b-a056-75584e2ffe6b,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"355dd9963bd54819a3a09df0f1f4fe81d0137ef9f582c3c1415cc444183c6e85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.928951 kubelet[2668]: E0911 23:31:51.927702 2668 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"355dd9963bd54819a3a09df0f1f4fe81d0137ef9f582c3c1415cc444183c6e85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 23:31:51.928951 kubelet[2668]: E0911 23:31:51.927756 2668 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"355dd9963bd54819a3a09df0f1f4fe81d0137ef9f582c3c1415cc444183c6e85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:51.928951 kubelet[2668]: E0911 23:31:51.927777 2668 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"355dd9963bd54819a3a09df0f1f4fe81d0137ef9f582c3c1415cc444183c6e85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-lgxh5" Sep 11 23:31:51.929080 kubelet[2668]: E0911 23:31:51.927824 2668 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-lgxh5_calico-system(d65db1fa-9a8f-403b-a056-75584e2ffe6b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-lgxh5_calico-system(d65db1fa-9a8f-403b-a056-75584e2ffe6b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"355dd9963bd54819a3a09df0f1f4fe81d0137ef9f582c3c1415cc444183c6e85\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-lgxh5" podUID="d65db1fa-9a8f-403b-a056-75584e2ffe6b" Sep 11 23:31:51.956961 containerd[1515]: time="2025-09-11T23:31:51.956912094Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 11 23:31:52.378317 systemd[1]: run-netns-cni\x2da1386c46\x2daba5\x2d5289\x2d9c0e\x2dec78df491455.mount: Deactivated successfully. Sep 11 23:31:56.207225 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2283514343.mount: Deactivated successfully. Sep 11 23:31:56.480727 containerd[1515]: time="2025-09-11T23:31:56.480673333Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=151100457" Sep 11 23:31:56.485251 containerd[1515]: time="2025-09-11T23:31:56.485122006Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"151100319\" in 4.528164547s" Sep 11 23:31:56.485251 containerd[1515]: time="2025-09-11T23:31:56.485159609Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\"" Sep 11 23:31:56.495025 containerd[1515]: time="2025-09-11T23:31:56.494990199Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:56.495653 containerd[1515]: time="2025-09-11T23:31:56.495624415Z" level=info msg="ImageCreate event name:\"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:56.496107 containerd[1515]: time="2025-09-11T23:31:56.496086216Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:56.500585 containerd[1515]: time="2025-09-11T23:31:56.500541490Z" level=info msg="CreateContainer within sandbox \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 11 23:31:56.524385 containerd[1515]: time="2025-09-11T23:31:56.523229256Z" level=info msg="Container 0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:56.524533 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3270032059.mount: Deactivated successfully. Sep 11 23:31:56.532792 containerd[1515]: time="2025-09-11T23:31:56.532744297Z" level=info msg="CreateContainer within sandbox \"ba354accab2fad6696ab012f0d3f5bda37521ba93c5b4f93438c033e706cd36b\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134\"" Sep 11 23:31:56.533669 containerd[1515]: time="2025-09-11T23:31:56.533642017Z" level=info msg="StartContainer for \"0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134\"" Sep 11 23:31:56.535059 containerd[1515]: time="2025-09-11T23:31:56.535030740Z" level=info msg="connecting to shim 0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134" address="unix:///run/containerd/s/de97ef20ae042bfae0358d80ebda5abb21291fb4d0003c51ff50cc3a289a7ac2" protocol=ttrpc version=3 Sep 11 23:31:56.559677 systemd[1]: Started cri-containerd-0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134.scope - libcontainer container 0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134. Sep 11 23:31:56.598781 containerd[1515]: time="2025-09-11T23:31:56.598646725Z" level=info msg="StartContainer for \"0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134\" returns successfully" Sep 11 23:31:56.722172 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 11 23:31:56.722319 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 11 23:31:57.005397 kubelet[2668]: I0911 23:31:57.005105 2668 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-ca-bundle\") pod \"6252d405-10a5-459a-ab90-82f9637e1a5a\" (UID: \"6252d405-10a5-459a-ab90-82f9637e1a5a\") " Sep 11 23:31:57.005809 kubelet[2668]: I0911 23:31:57.005516 2668 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-backend-key-pair\") pod \"6252d405-10a5-459a-ab90-82f9637e1a5a\" (UID: \"6252d405-10a5-459a-ab90-82f9637e1a5a\") " Sep 11 23:31:57.005809 kubelet[2668]: I0911 23:31:57.005554 2668 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2r2b\" (UniqueName: \"kubernetes.io/projected/6252d405-10a5-459a-ab90-82f9637e1a5a-kube-api-access-r2r2b\") pod \"6252d405-10a5-459a-ab90-82f9637e1a5a\" (UID: \"6252d405-10a5-459a-ab90-82f9637e1a5a\") " Sep 11 23:31:57.037204 kubelet[2668]: I0911 23:31:57.036409 2668 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "6252d405-10a5-459a-ab90-82f9637e1a5a" (UID: "6252d405-10a5-459a-ab90-82f9637e1a5a"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 11 23:31:57.038735 kubelet[2668]: I0911 23:31:57.038691 2668 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6252d405-10a5-459a-ab90-82f9637e1a5a-kube-api-access-r2r2b" (OuterVolumeSpecName: "kube-api-access-r2r2b") pod "6252d405-10a5-459a-ab90-82f9637e1a5a" (UID: "6252d405-10a5-459a-ab90-82f9637e1a5a"). InnerVolumeSpecName "kube-api-access-r2r2b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 11 23:31:57.041322 kubelet[2668]: I0911 23:31:57.041279 2668 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "6252d405-10a5-459a-ab90-82f9637e1a5a" (UID: "6252d405-10a5-459a-ab90-82f9637e1a5a"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 11 23:31:57.073109 kubelet[2668]: I0911 23:31:57.073021 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-mxfjl" podStartSLOduration=1.7560329590000001 podStartE2EDuration="16.073003047s" podCreationTimestamp="2025-09-11 23:31:41 +0000 UTC" firstStartedPulling="2025-09-11 23:31:42.168882623 +0000 UTC m=+21.418402743" lastFinishedPulling="2025-09-11 23:31:56.485852711 +0000 UTC m=+35.735372831" observedRunningTime="2025-09-11 23:31:57.072076127 +0000 UTC m=+36.321596287" watchObservedRunningTime="2025-09-11 23:31:57.073003047 +0000 UTC m=+36.322523167" Sep 11 23:31:57.107033 kubelet[2668]: I0911 23:31:57.106972 2668 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Sep 11 23:31:57.107033 kubelet[2668]: I0911 23:31:57.107015 2668 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-r2r2b\" (UniqueName: \"kubernetes.io/projected/6252d405-10a5-459a-ab90-82f9637e1a5a-kube-api-access-r2r2b\") on node \"localhost\" DevicePath \"\"" Sep 11 23:31:57.107033 kubelet[2668]: I0911 23:31:57.107024 2668 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6252d405-10a5-459a-ab90-82f9637e1a5a-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Sep 11 23:31:57.208011 systemd[1]: var-lib-kubelet-pods-6252d405\x2d10a5\x2d459a\x2dab90\x2d82f9637e1a5a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dr2r2b.mount: Deactivated successfully. Sep 11 23:31:57.208109 systemd[1]: var-lib-kubelet-pods-6252d405\x2d10a5\x2d459a\x2dab90\x2d82f9637e1a5a-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 11 23:31:57.284930 systemd[1]: Removed slice kubepods-besteffort-pod6252d405_10a5_459a_ab90_82f9637e1a5a.slice - libcontainer container kubepods-besteffort-pod6252d405_10a5_459a_ab90_82f9637e1a5a.slice. Sep 11 23:31:57.381936 systemd[1]: Created slice kubepods-besteffort-pod1e8cbd7f_b87f_4568_95fc_8442afd80144.slice - libcontainer container kubepods-besteffort-pod1e8cbd7f_b87f_4568_95fc_8442afd80144.slice. Sep 11 23:31:57.509053 kubelet[2668]: I0911 23:31:57.509015 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8cbd7f-b87f-4568-95fc-8442afd80144-whisker-ca-bundle\") pod \"whisker-7dfd48b4cf-jmt9v\" (UID: \"1e8cbd7f-b87f-4568-95fc-8442afd80144\") " pod="calico-system/whisker-7dfd48b4cf-jmt9v" Sep 11 23:31:57.509053 kubelet[2668]: I0911 23:31:57.509061 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/1e8cbd7f-b87f-4568-95fc-8442afd80144-whisker-backend-key-pair\") pod \"whisker-7dfd48b4cf-jmt9v\" (UID: \"1e8cbd7f-b87f-4568-95fc-8442afd80144\") " pod="calico-system/whisker-7dfd48b4cf-jmt9v" Sep 11 23:31:57.509258 kubelet[2668]: I0911 23:31:57.509090 2668 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq4mp\" (UniqueName: \"kubernetes.io/projected/1e8cbd7f-b87f-4568-95fc-8442afd80144-kube-api-access-mq4mp\") pod \"whisker-7dfd48b4cf-jmt9v\" (UID: \"1e8cbd7f-b87f-4568-95fc-8442afd80144\") " pod="calico-system/whisker-7dfd48b4cf-jmt9v" Sep 11 23:31:57.685926 containerd[1515]: time="2025-09-11T23:31:57.685820805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7dfd48b4cf-jmt9v,Uid:1e8cbd7f-b87f-4568-95fc-8442afd80144,Namespace:calico-system,Attempt:0,}" Sep 11 23:31:57.851852 systemd-networkd[1433]: calic065b69c11b: Link UP Sep 11 23:31:57.852913 systemd-networkd[1433]: calic065b69c11b: Gained carrier Sep 11 23:31:57.864761 containerd[1515]: 2025-09-11 23:31:57.707 [INFO][3821] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 23:31:57.864761 containerd[1515]: 2025-09-11 23:31:57.738 [INFO][3821] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0 whisker-7dfd48b4cf- calico-system 1e8cbd7f-b87f-4568-95fc-8442afd80144 915 0 2025-09-11 23:31:57 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7dfd48b4cf projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-7dfd48b4cf-jmt9v eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calic065b69c11b [] [] }} ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-" Sep 11 23:31:57.864761 containerd[1515]: 2025-09-11 23:31:57.739 [INFO][3821] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.864761 containerd[1515]: 2025-09-11 23:31:57.807 [INFO][3836] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" HandleID="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Workload="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.807 [INFO][3836] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" HandleID="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Workload="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000514b90), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-7dfd48b4cf-jmt9v", "timestamp":"2025-09-11 23:31:57.807021047 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.807 [INFO][3836] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.807 [INFO][3836] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.807 [INFO][3836] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.817 [INFO][3836] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" host="localhost" Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.823 [INFO][3836] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.827 [INFO][3836] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.828 [INFO][3836] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.830 [INFO][3836] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:31:57.865141 containerd[1515]: 2025-09-11 23:31:57.830 [INFO][3836] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" host="localhost" Sep 11 23:31:57.865388 containerd[1515]: 2025-09-11 23:31:57.832 [INFO][3836] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259 Sep 11 23:31:57.865388 containerd[1515]: 2025-09-11 23:31:57.837 [INFO][3836] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" host="localhost" Sep 11 23:31:57.865388 containerd[1515]: 2025-09-11 23:31:57.842 [INFO][3836] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" host="localhost" Sep 11 23:31:57.865388 containerd[1515]: 2025-09-11 23:31:57.842 [INFO][3836] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" host="localhost" Sep 11 23:31:57.865388 containerd[1515]: 2025-09-11 23:31:57.842 [INFO][3836] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:31:57.865388 containerd[1515]: 2025-09-11 23:31:57.842 [INFO][3836] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" HandleID="k8s-pod-network.8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Workload="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.865695 containerd[1515]: 2025-09-11 23:31:57.845 [INFO][3821] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0", GenerateName:"whisker-7dfd48b4cf-", Namespace:"calico-system", SelfLink:"", UID:"1e8cbd7f-b87f-4568-95fc-8442afd80144", ResourceVersion:"915", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 57, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7dfd48b4cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-7dfd48b4cf-jmt9v", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calic065b69c11b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:31:57.865695 containerd[1515]: 2025-09-11 23:31:57.845 [INFO][3821] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.865777 containerd[1515]: 2025-09-11 23:31:57.846 [INFO][3821] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic065b69c11b ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.865777 containerd[1515]: 2025-09-11 23:31:57.852 [INFO][3821] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.865814 containerd[1515]: 2025-09-11 23:31:57.852 [INFO][3821] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0", GenerateName:"whisker-7dfd48b4cf-", Namespace:"calico-system", SelfLink:"", UID:"1e8cbd7f-b87f-4568-95fc-8442afd80144", ResourceVersion:"915", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 57, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7dfd48b4cf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259", Pod:"whisker-7dfd48b4cf-jmt9v", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calic065b69c11b", MAC:"ea:44:ec:1d:ed:fc", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:31:57.865861 containerd[1515]: 2025-09-11 23:31:57.860 [INFO][3821] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" Namespace="calico-system" Pod="whisker-7dfd48b4cf-jmt9v" WorkloadEndpoint="localhost-k8s-whisker--7dfd48b4cf--jmt9v-eth0" Sep 11 23:31:57.909114 containerd[1515]: time="2025-09-11T23:31:57.909053165Z" level=info msg="connecting to shim 8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259" address="unix:///run/containerd/s/94524c34e8276166cc9bf2033aaf82ceb3d3f1854e7d5e72fa871be58b4a2ab5" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:31:57.953684 systemd[1]: Started cri-containerd-8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259.scope - libcontainer container 8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259. Sep 11 23:31:57.965668 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:31:57.982187 kubelet[2668]: I0911 23:31:57.982136 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:31:58.000836 containerd[1515]: time="2025-09-11T23:31:58.000773676Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7dfd48b4cf-jmt9v,Uid:1e8cbd7f-b87f-4568-95fc-8442afd80144,Namespace:calico-system,Attempt:0,} returns sandbox id \"8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259\"" Sep 11 23:31:58.002899 containerd[1515]: time="2025-09-11T23:31:58.002817047Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 11 23:31:58.559760 systemd-networkd[1433]: vxlan.calico: Link UP Sep 11 23:31:58.559767 systemd-networkd[1433]: vxlan.calico: Gained carrier Sep 11 23:31:58.867313 kubelet[2668]: I0911 23:31:58.867213 2668 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6252d405-10a5-459a-ab90-82f9637e1a5a" path="/var/lib/kubelet/pods/6252d405-10a5-459a-ab90-82f9637e1a5a/volumes" Sep 11 23:31:59.329817 containerd[1515]: time="2025-09-11T23:31:59.329753505Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:59.330361 containerd[1515]: time="2025-09-11T23:31:59.330341432Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4605606" Sep 11 23:31:59.331335 containerd[1515]: time="2025-09-11T23:31:59.331296030Z" level=info msg="ImageCreate event name:\"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:59.333523 containerd[1515]: time="2025-09-11T23:31:59.333495208Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:31:59.334396 containerd[1515]: time="2025-09-11T23:31:59.334127459Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"5974839\" in 1.331221926s" Sep 11 23:31:59.334396 containerd[1515]: time="2025-09-11T23:31:59.334164382Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\"" Sep 11 23:31:59.337731 containerd[1515]: time="2025-09-11T23:31:59.337680388Z" level=info msg="CreateContainer within sandbox \"8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 11 23:31:59.360004 containerd[1515]: time="2025-09-11T23:31:59.359693013Z" level=info msg="Container 6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:31:59.368794 containerd[1515]: time="2025-09-11T23:31:59.368751947Z" level=info msg="CreateContainer within sandbox \"8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2\"" Sep 11 23:31:59.369532 containerd[1515]: time="2025-09-11T23:31:59.369498208Z" level=info msg="StartContainer for \"6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2\"" Sep 11 23:31:59.371113 containerd[1515]: time="2025-09-11T23:31:59.371063135Z" level=info msg="connecting to shim 6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2" address="unix:///run/containerd/s/94524c34e8276166cc9bf2033aaf82ceb3d3f1854e7d5e72fa871be58b4a2ab5" protocol=ttrpc version=3 Sep 11 23:31:59.395743 systemd[1]: Started cri-containerd-6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2.scope - libcontainer container 6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2. Sep 11 23:31:59.439944 containerd[1515]: time="2025-09-11T23:31:59.439895757Z" level=info msg="StartContainer for \"6fb3346e67cc2e527b123c36d30ce09b4f54b182f05812e91fc23ac06f41d1d2\" returns successfully" Sep 11 23:31:59.441297 containerd[1515]: time="2025-09-11T23:31:59.441270148Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 11 23:31:59.570908 systemd-networkd[1433]: calic065b69c11b: Gained IPv6LL Sep 11 23:32:00.297911 kubelet[2668]: I0911 23:32:00.297627 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:32:00.440196 containerd[1515]: time="2025-09-11T23:32:00.440114330Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134\" id:\"0b4b6a346cbd4385fb98d9d4b36a8833644f29e66a421f2a8af93f5098d962b5\" pid:4153 exited_at:{seconds:1757633520 nanos:439583648}" Sep 11 23:32:00.467335 systemd-networkd[1433]: vxlan.calico: Gained IPv6LL Sep 11 23:32:00.562600 containerd[1515]: time="2025-09-11T23:32:00.562317818Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134\" id:\"9bc3f54e0accb951ea3237994b1897480ca38bd20a901d87543b89787fef401f\" pid:4178 exited_at:{seconds:1757633520 nanos:559550679}" Sep 11 23:32:00.905171 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2011726794.mount: Deactivated successfully. Sep 11 23:32:00.975952 containerd[1515]: time="2025-09-11T23:32:00.975879509Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:00.978311 containerd[1515]: time="2025-09-11T23:32:00.978251016Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=30823700" Sep 11 23:32:00.979433 containerd[1515]: time="2025-09-11T23:32:00.979393906Z" level=info msg="ImageCreate event name:\"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:00.981873 containerd[1515]: time="2025-09-11T23:32:00.981816778Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:00.982578 containerd[1515]: time="2025-09-11T23:32:00.982530674Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"30823530\" in 1.541221882s" Sep 11 23:32:00.982578 containerd[1515]: time="2025-09-11T23:32:00.982570797Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\"" Sep 11 23:32:00.986873 containerd[1515]: time="2025-09-11T23:32:00.986827733Z" level=info msg="CreateContainer within sandbox \"8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 11 23:32:00.996375 containerd[1515]: time="2025-09-11T23:32:00.996315082Z" level=info msg="Container dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:01.004072 containerd[1515]: time="2025-09-11T23:32:01.004027566Z" level=info msg="CreateContainer within sandbox \"8458021f019e07921e56954945ca355a0696304157dc1733b130e01cd608f259\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934\"" Sep 11 23:32:01.004845 containerd[1515]: time="2025-09-11T23:32:01.004814987Z" level=info msg="StartContainer for \"dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934\"" Sep 11 23:32:01.006367 containerd[1515]: time="2025-09-11T23:32:01.006335864Z" level=info msg="connecting to shim dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934" address="unix:///run/containerd/s/94524c34e8276166cc9bf2033aaf82ceb3d3f1854e7d5e72fa871be58b4a2ab5" protocol=ttrpc version=3 Sep 11 23:32:01.031705 systemd[1]: Started cri-containerd-dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934.scope - libcontainer container dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934. Sep 11 23:32:01.071508 containerd[1515]: time="2025-09-11T23:32:01.070773062Z" level=info msg="StartContainer for \"dd95cab91504b53fcb72b226356f24497f762e5f31a6027b10695f9ee6a44934\" returns successfully" Sep 11 23:32:01.864961 containerd[1515]: time="2025-09-11T23:32:01.864919883Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-ckb7g,Uid:341f9b5c-0ff7-44c6-9add-ef9f11e8780b,Namespace:calico-apiserver,Attempt:0,}" Sep 11 23:32:01.976304 systemd-networkd[1433]: caliaceb4644477: Link UP Sep 11 23:32:01.978043 systemd-networkd[1433]: caliaceb4644477: Gained carrier Sep 11 23:32:01.994071 containerd[1515]: 2025-09-11 23:32:01.903 [INFO][4235] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0 calico-apiserver-56f49b666f- calico-apiserver 341f9b5c-0ff7-44c6-9add-ef9f11e8780b 844 0 2025-09-11 23:31:37 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:56f49b666f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-56f49b666f-ckb7g eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] caliaceb4644477 [] [] }} ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-" Sep 11 23:32:01.994071 containerd[1515]: 2025-09-11 23:32:01.904 [INFO][4235] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:01.994071 containerd[1515]: 2025-09-11 23:32:01.929 [INFO][4250] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" HandleID="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Workload="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.930 [INFO][4250] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" HandleID="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Workload="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c760), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-56f49b666f-ckb7g", "timestamp":"2025-09-11 23:32:01.929902482 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.930 [INFO][4250] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.930 [INFO][4250] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.931 [INFO][4250] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.940 [INFO][4250] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" host="localhost" Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.949 [INFO][4250] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.954 [INFO][4250] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.956 [INFO][4250] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.959 [INFO][4250] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:01.994266 containerd[1515]: 2025-09-11 23:32:01.959 [INFO][4250] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" host="localhost" Sep 11 23:32:01.994710 containerd[1515]: 2025-09-11 23:32:01.961 [INFO][4250] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21 Sep 11 23:32:01.994710 containerd[1515]: 2025-09-11 23:32:01.965 [INFO][4250] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" host="localhost" Sep 11 23:32:01.994710 containerd[1515]: 2025-09-11 23:32:01.971 [INFO][4250] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" host="localhost" Sep 11 23:32:01.994710 containerd[1515]: 2025-09-11 23:32:01.971 [INFO][4250] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" host="localhost" Sep 11 23:32:01.994710 containerd[1515]: 2025-09-11 23:32:01.971 [INFO][4250] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:01.994710 containerd[1515]: 2025-09-11 23:32:01.971 [INFO][4250] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" HandleID="k8s-pod-network.a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Workload="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:01.995137 containerd[1515]: 2025-09-11 23:32:01.973 [INFO][4235] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0", GenerateName:"calico-apiserver-56f49b666f-", Namespace:"calico-apiserver", SelfLink:"", UID:"341f9b5c-0ff7-44c6-9add-ef9f11e8780b", ResourceVersion:"844", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56f49b666f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-56f49b666f-ckb7g", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliaceb4644477", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:01.995211 containerd[1515]: 2025-09-11 23:32:01.973 [INFO][4235] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:01.995211 containerd[1515]: 2025-09-11 23:32:01.973 [INFO][4235] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliaceb4644477 ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:01.995211 containerd[1515]: 2025-09-11 23:32:01.977 [INFO][4235] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:01.995283 containerd[1515]: 2025-09-11 23:32:01.977 [INFO][4235] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0", GenerateName:"calico-apiserver-56f49b666f-", Namespace:"calico-apiserver", SelfLink:"", UID:"341f9b5c-0ff7-44c6-9add-ef9f11e8780b", ResourceVersion:"844", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56f49b666f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21", Pod:"calico-apiserver-56f49b666f-ckb7g", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliaceb4644477", MAC:"a6:1a:66:e8:6e:fc", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:01.995337 containerd[1515]: 2025-09-11 23:32:01.988 [INFO][4235] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-ckb7g" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--ckb7g-eth0" Sep 11 23:32:02.029006 containerd[1515]: time="2025-09-11T23:32:02.028489537Z" level=info msg="connecting to shim a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21" address="unix:///run/containerd/s/61c8941b18020e6f65731c62125d1afb89b3c2cf44aa5d48f31cd6ef6d0c08ae" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:02.065726 systemd[1]: Started cri-containerd-a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21.scope - libcontainer container a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21. Sep 11 23:32:02.079117 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:02.104200 containerd[1515]: time="2025-09-11T23:32:02.104156536Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-ckb7g,Uid:341f9b5c-0ff7-44c6-9add-ef9f11e8780b,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21\"" Sep 11 23:32:02.105926 containerd[1515]: time="2025-09-11T23:32:02.105895186Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 11 23:32:02.865400 containerd[1515]: time="2025-09-11T23:32:02.865280940Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-cb484f97b-jkcgr,Uid:c0e77bb6-9a42-4b41-98bb-3d6bcd21104b,Namespace:calico-system,Attempt:0,}" Sep 11 23:32:03.149721 systemd-networkd[1433]: calicc848d13d97: Link UP Sep 11 23:32:03.151620 systemd-networkd[1433]: calicc848d13d97: Gained carrier Sep 11 23:32:03.167443 kubelet[2668]: I0911 23:32:03.167320 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-7dfd48b4cf-jmt9v" podStartSLOduration=3.186612249 podStartE2EDuration="6.167302275s" podCreationTimestamp="2025-09-11 23:31:57 +0000 UTC" firstStartedPulling="2025-09-11 23:31:58.002580667 +0000 UTC m=+37.252100787" lastFinishedPulling="2025-09-11 23:32:00.983270733 +0000 UTC m=+40.232790813" observedRunningTime="2025-09-11 23:32:02.020305322 +0000 UTC m=+41.269825442" watchObservedRunningTime="2025-09-11 23:32:03.167302275 +0000 UTC m=+42.416822395" Sep 11 23:32:03.170807 containerd[1515]: 2025-09-11 23:32:02.919 [INFO][4317] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0 calico-kube-controllers-cb484f97b- calico-system c0e77bb6-9a42-4b41-98bb-3d6bcd21104b 849 0 2025-09-11 23:31:42 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:cb484f97b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-cb484f97b-jkcgr eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calicc848d13d97 [] [] }} ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-" Sep 11 23:32:03.170807 containerd[1515]: 2025-09-11 23:32:02.919 [INFO][4317] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.170807 containerd[1515]: 2025-09-11 23:32:02.970 [INFO][4332] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" HandleID="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Workload="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:02.970 [INFO][4332] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" HandleID="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Workload="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000511970), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-cb484f97b-jkcgr", "timestamp":"2025-09-11 23:32:02.970252779 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:02.970 [INFO][4332] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:02.970 [INFO][4332] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:02.970 [INFO][4332] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:03.108 [INFO][4332] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" host="localhost" Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:03.118 [INFO][4332] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:03.125 [INFO][4332] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:03.128 [INFO][4332] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:03.131 [INFO][4332] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:03.171007 containerd[1515]: 2025-09-11 23:32:03.131 [INFO][4332] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" host="localhost" Sep 11 23:32:03.171210 containerd[1515]: 2025-09-11 23:32:03.133 [INFO][4332] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca Sep 11 23:32:03.171210 containerd[1515]: 2025-09-11 23:32:03.137 [INFO][4332] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" host="localhost" Sep 11 23:32:03.171210 containerd[1515]: 2025-09-11 23:32:03.144 [INFO][4332] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" host="localhost" Sep 11 23:32:03.171210 containerd[1515]: 2025-09-11 23:32:03.144 [INFO][4332] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" host="localhost" Sep 11 23:32:03.171210 containerd[1515]: 2025-09-11 23:32:03.144 [INFO][4332] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:03.171210 containerd[1515]: 2025-09-11 23:32:03.144 [INFO][4332] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" HandleID="k8s-pod-network.c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Workload="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.171322 containerd[1515]: 2025-09-11 23:32:03.147 [INFO][4317] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0", GenerateName:"calico-kube-controllers-cb484f97b-", Namespace:"calico-system", SelfLink:"", UID:"c0e77bb6-9a42-4b41-98bb-3d6bcd21104b", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"cb484f97b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-cb484f97b-jkcgr", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calicc848d13d97", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:03.171369 containerd[1515]: 2025-09-11 23:32:03.147 [INFO][4317] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.171369 containerd[1515]: 2025-09-11 23:32:03.147 [INFO][4317] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicc848d13d97 ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.171369 containerd[1515]: 2025-09-11 23:32:03.151 [INFO][4317] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.171428 containerd[1515]: 2025-09-11 23:32:03.154 [INFO][4317] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0", GenerateName:"calico-kube-controllers-cb484f97b-", Namespace:"calico-system", SelfLink:"", UID:"c0e77bb6-9a42-4b41-98bb-3d6bcd21104b", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"cb484f97b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca", Pod:"calico-kube-controllers-cb484f97b-jkcgr", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calicc848d13d97", MAC:"2a:7a:8f:18:e1:d7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:03.171501 containerd[1515]: 2025-09-11 23:32:03.168 [INFO][4317] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" Namespace="calico-system" Pod="calico-kube-controllers-cb484f97b-jkcgr" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--cb484f97b--jkcgr-eth0" Sep 11 23:32:03.195527 containerd[1515]: time="2025-09-11T23:32:03.195281685Z" level=info msg="connecting to shim c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca" address="unix:///run/containerd/s/4a29ed1bba28f78346e58bff1d88fe808b5fece4e32638b3623957d2e1f26cf1" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:03.224667 systemd[1]: Started cri-containerd-c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca.scope - libcontainer container c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca. Sep 11 23:32:03.241636 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:03.289569 containerd[1515]: time="2025-09-11T23:32:03.289527712Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-cb484f97b-jkcgr,Uid:c0e77bb6-9a42-4b41-98bb-3d6bcd21104b,Namespace:calico-system,Attempt:0,} returns sandbox id \"c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca\"" Sep 11 23:32:03.346608 systemd-networkd[1433]: caliaceb4644477: Gained IPv6LL Sep 11 23:32:03.658173 systemd[1]: Started sshd@7-10.0.0.19:22-10.0.0.1:35276.service - OpenSSH per-connection server daemon (10.0.0.1:35276). Sep 11 23:32:03.745778 sshd[4401]: Accepted publickey for core from 10.0.0.1 port 35276 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:03.749261 sshd-session[4401]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:03.758872 systemd-logind[1485]: New session 8 of user core. Sep 11 23:32:03.772743 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 11 23:32:03.864715 containerd[1515]: time="2025-09-11T23:32:03.864670701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-s852n,Uid:cb91f87b-09d0-4126-b7eb-549a67e8f409,Namespace:calico-system,Attempt:0,}" Sep 11 23:32:03.864715 containerd[1515]: time="2025-09-11T23:32:03.864703223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-lgxh5,Uid:d65db1fa-9a8f-403b-a056-75584e2ffe6b,Namespace:calico-system,Attempt:0,}" Sep 11 23:32:04.106242 systemd-networkd[1433]: cali9315fc3eab2: Link UP Sep 11 23:32:04.109298 systemd-networkd[1433]: cali9315fc3eab2: Gained carrier Sep 11 23:32:04.122200 sshd[4404]: Connection closed by 10.0.0.1 port 35276 Sep 11 23:32:04.122684 sshd-session[4401]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:04.127906 systemd[1]: sshd@7-10.0.0.19:22-10.0.0.1:35276.service: Deactivated successfully. Sep 11 23:32:04.134365 systemd[1]: session-8.scope: Deactivated successfully. Sep 11 23:32:04.137374 systemd-logind[1485]: Session 8 logged out. Waiting for processes to exit. Sep 11 23:32:04.140902 systemd-logind[1485]: Removed session 8. Sep 11 23:32:04.141809 containerd[1515]: 2025-09-11 23:32:03.989 [INFO][4430] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--lgxh5-eth0 csi-node-driver- calico-system d65db1fa-9a8f-403b-a056-75584e2ffe6b 724 0 2025-09-11 23:31:42 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-lgxh5 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali9315fc3eab2 [] [] }} ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-" Sep 11 23:32:04.141809 containerd[1515]: 2025-09-11 23:32:03.989 [INFO][4430] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.141809 containerd[1515]: 2025-09-11 23:32:04.029 [INFO][4458] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" HandleID="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Workload="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.029 [INFO][4458] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" HandleID="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Workload="localhost-k8s-csi--node--driver--lgxh5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000123750), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-lgxh5", "timestamp":"2025-09-11 23:32:04.029395247 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.029 [INFO][4458] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.029 [INFO][4458] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.030 [INFO][4458] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.045 [INFO][4458] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" host="localhost" Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.055 [INFO][4458] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.065 [INFO][4458] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.069 [INFO][4458] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.074 [INFO][4458] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:04.142195 containerd[1515]: 2025-09-11 23:32:04.074 [INFO][4458] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" host="localhost" Sep 11 23:32:04.142418 containerd[1515]: 2025-09-11 23:32:04.076 [INFO][4458] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464 Sep 11 23:32:04.142418 containerd[1515]: 2025-09-11 23:32:04.083 [INFO][4458] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" host="localhost" Sep 11 23:32:04.142418 containerd[1515]: 2025-09-11 23:32:04.095 [INFO][4458] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" host="localhost" Sep 11 23:32:04.142418 containerd[1515]: 2025-09-11 23:32:04.095 [INFO][4458] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" host="localhost" Sep 11 23:32:04.142418 containerd[1515]: 2025-09-11 23:32:04.095 [INFO][4458] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:04.142418 containerd[1515]: 2025-09-11 23:32:04.095 [INFO][4458] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" HandleID="k8s-pod-network.05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Workload="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.143751 containerd[1515]: 2025-09-11 23:32:04.103 [INFO][4430] cni-plugin/k8s.go 418: Populated endpoint ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--lgxh5-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d65db1fa-9a8f-403b-a056-75584e2ffe6b", ResourceVersion:"724", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-lgxh5", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali9315fc3eab2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:04.143826 containerd[1515]: 2025-09-11 23:32:04.103 [INFO][4430] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.143826 containerd[1515]: 2025-09-11 23:32:04.103 [INFO][4430] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9315fc3eab2 ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.143826 containerd[1515]: 2025-09-11 23:32:04.106 [INFO][4430] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.143885 containerd[1515]: 2025-09-11 23:32:04.109 [INFO][4430] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--lgxh5-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d65db1fa-9a8f-403b-a056-75584e2ffe6b", ResourceVersion:"724", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464", Pod:"csi-node-driver-lgxh5", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali9315fc3eab2", MAC:"e6:26:33:e9:1f:d4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:04.143931 containerd[1515]: 2025-09-11 23:32:04.127 [INFO][4430] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" Namespace="calico-system" Pod="csi-node-driver-lgxh5" WorkloadEndpoint="localhost-k8s-csi--node--driver--lgxh5-eth0" Sep 11 23:32:04.225666 systemd-networkd[1433]: cali6b880faf9ef: Link UP Sep 11 23:32:04.231695 systemd-networkd[1433]: cali6b880faf9ef: Gained carrier Sep 11 23:32:04.260238 containerd[1515]: time="2025-09-11T23:32:04.259276192Z" level=info msg="connecting to shim 05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464" address="unix:///run/containerd/s/ee9fdbcaaa58e16e816071da17e823d432c404e5bbc64e9418325c2646d5ae54" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:04.266525 containerd[1515]: 2025-09-11 23:32:03.980 [INFO][4420] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--54d579b49d--s852n-eth0 goldmane-54d579b49d- calico-system cb91f87b-09d0-4126-b7eb-549a67e8f409 851 0 2025-09-11 23:31:41 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-54d579b49d-s852n eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali6b880faf9ef [] [] }} ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-" Sep 11 23:32:04.266525 containerd[1515]: 2025-09-11 23:32:03.980 [INFO][4420] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.266525 containerd[1515]: 2025-09-11 23:32:04.067 [INFO][4456] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" HandleID="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Workload="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.068 [INFO][4456] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" HandleID="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Workload="localhost-k8s-goldmane--54d579b49d--s852n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004d790), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-54d579b49d-s852n", "timestamp":"2025-09-11 23:32:04.067641626 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.068 [INFO][4456] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.096 [INFO][4456] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.096 [INFO][4456] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.148 [INFO][4456] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" host="localhost" Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.156 [INFO][4456] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.173 [INFO][4456] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.177 [INFO][4456] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.182 [INFO][4456] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:04.266746 containerd[1515]: 2025-09-11 23:32:04.183 [INFO][4456] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" host="localhost" Sep 11 23:32:04.266998 containerd[1515]: 2025-09-11 23:32:04.186 [INFO][4456] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f Sep 11 23:32:04.266998 containerd[1515]: 2025-09-11 23:32:04.194 [INFO][4456] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" host="localhost" Sep 11 23:32:04.266998 containerd[1515]: 2025-09-11 23:32:04.206 [INFO][4456] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" host="localhost" Sep 11 23:32:04.266998 containerd[1515]: 2025-09-11 23:32:04.206 [INFO][4456] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" host="localhost" Sep 11 23:32:04.266998 containerd[1515]: 2025-09-11 23:32:04.207 [INFO][4456] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:04.266998 containerd[1515]: 2025-09-11 23:32:04.207 [INFO][4456] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" HandleID="k8s-pod-network.33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Workload="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.267128 containerd[1515]: 2025-09-11 23:32:04.217 [INFO][4420] cni-plugin/k8s.go 418: Populated endpoint ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--s852n-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"cb91f87b-09d0-4126-b7eb-549a67e8f409", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-54d579b49d-s852n", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali6b880faf9ef", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:04.267128 containerd[1515]: 2025-09-11 23:32:04.217 [INFO][4420] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.267203 containerd[1515]: 2025-09-11 23:32:04.217 [INFO][4420] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6b880faf9ef ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.267203 containerd[1515]: 2025-09-11 23:32:04.230 [INFO][4420] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.267239 containerd[1515]: 2025-09-11 23:32:04.232 [INFO][4420] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--s852n-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"cb91f87b-09d0-4126-b7eb-549a67e8f409", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f", Pod:"goldmane-54d579b49d-s852n", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali6b880faf9ef", MAC:"a2:76:2d:37:4f:c3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:04.271443 containerd[1515]: 2025-09-11 23:32:04.247 [INFO][4420] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" Namespace="calico-system" Pod="goldmane-54d579b49d-s852n" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--s852n-eth0" Sep 11 23:32:04.310946 containerd[1515]: time="2025-09-11T23:32:04.310884929Z" level=info msg="connecting to shim 33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f" address="unix:///run/containerd/s/9f0bd42a089419fd6709e62394a6c009514d65f89b562beb4dced3fd2d92fc42" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:04.329991 systemd[1]: Started cri-containerd-05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464.scope - libcontainer container 05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464. Sep 11 23:32:04.388905 systemd[1]: Started cri-containerd-33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f.scope - libcontainer container 33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f. Sep 11 23:32:04.402924 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:04.426045 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:04.427345 containerd[1515]: time="2025-09-11T23:32:04.427283746Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-lgxh5,Uid:d65db1fa-9a8f-403b-a056-75584e2ffe6b,Namespace:calico-system,Attempt:0,} returns sandbox id \"05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464\"" Sep 11 23:32:04.462206 containerd[1515]: time="2025-09-11T23:32:04.462141523Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-s852n,Uid:cb91f87b-09d0-4126-b7eb-549a67e8f409,Namespace:calico-system,Attempt:0,} returns sandbox id \"33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f\"" Sep 11 23:32:04.818950 systemd-networkd[1433]: calicc848d13d97: Gained IPv6LL Sep 11 23:32:04.864900 kubelet[2668]: E0911 23:32:04.864856 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:04.865279 containerd[1515]: time="2025-09-11T23:32:04.865237075Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-d725f,Uid:32dd7b29-d990-4d4c-8903-e51a4239fdc5,Namespace:kube-system,Attempt:0,}" Sep 11 23:32:04.949303 containerd[1515]: time="2025-09-11T23:32:04.949231531Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:04.950724 containerd[1515]: time="2025-09-11T23:32:04.950682915Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=44530807" Sep 11 23:32:04.952001 containerd[1515]: time="2025-09-11T23:32:04.951947886Z" level=info msg="ImageCreate event name:\"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:04.955999 containerd[1515]: time="2025-09-11T23:32:04.955958533Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:04.957397 containerd[1515]: time="2025-09-11T23:32:04.957364074Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"45900064\" in 2.85123031s" Sep 11 23:32:04.957488 containerd[1515]: time="2025-09-11T23:32:04.957411437Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\"" Sep 11 23:32:04.960167 containerd[1515]: time="2025-09-11T23:32:04.960026944Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 11 23:32:04.963930 containerd[1515]: time="2025-09-11T23:32:04.963893181Z" level=info msg="CreateContainer within sandbox \"a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 11 23:32:04.977433 containerd[1515]: time="2025-09-11T23:32:04.977381227Z" level=info msg="Container 288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:04.992567 containerd[1515]: time="2025-09-11T23:32:04.992176727Z" level=info msg="CreateContainer within sandbox \"a9056770f37356d933979e31423f28a7d48cf5e0a2ea08a97e865173388a2a21\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263\"" Sep 11 23:32:04.994963 containerd[1515]: time="2025-09-11T23:32:04.994675266Z" level=info msg="StartContainer for \"288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263\"" Sep 11 23:32:04.995830 containerd[1515]: time="2025-09-11T23:32:04.995797147Z" level=info msg="connecting to shim 288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263" address="unix:///run/containerd/s/61c8941b18020e6f65731c62125d1afb89b3c2cf44aa5d48f31cd6ef6d0c08ae" protocol=ttrpc version=3 Sep 11 23:32:05.023957 systemd[1]: Started cri-containerd-288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263.scope - libcontainer container 288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263. Sep 11 23:32:05.071822 systemd-networkd[1433]: cali2e2eb2c547c: Link UP Sep 11 23:32:05.074819 systemd-networkd[1433]: cali2e2eb2c547c: Gained carrier Sep 11 23:32:05.093211 containerd[1515]: time="2025-09-11T23:32:05.093163618Z" level=info msg="StartContainer for \"288d259597c57a8c11e9fb14d2eeab3da603af1c3d7912d5e160de7d765d4263\" returns successfully" Sep 11 23:32:05.093591 containerd[1515]: 2025-09-11 23:32:04.945 [INFO][4596] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--d725f-eth0 coredns-674b8bbfcf- kube-system 32dd7b29-d990-4d4c-8903-e51a4239fdc5 847 0 2025-09-11 23:31:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-d725f eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2e2eb2c547c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-" Sep 11 23:32:05.093591 containerd[1515]: 2025-09-11 23:32:04.946 [INFO][4596] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.093591 containerd[1515]: 2025-09-11 23:32:05.002 [INFO][4614] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" HandleID="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Workload="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.003 [INFO][4614] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" HandleID="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Workload="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001a05b0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-d725f", "timestamp":"2025-09-11 23:32:05.002970178 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.003 [INFO][4614] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.003 [INFO][4614] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.003 [INFO][4614] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.019 [INFO][4614] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" host="localhost" Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.033 [INFO][4614] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.041 [INFO][4614] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.043 [INFO][4614] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.046 [INFO][4614] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:05.093734 containerd[1515]: 2025-09-11 23:32:05.046 [INFO][4614] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" host="localhost" Sep 11 23:32:05.094621 containerd[1515]: 2025-09-11 23:32:05.048 [INFO][4614] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45 Sep 11 23:32:05.094621 containerd[1515]: 2025-09-11 23:32:05.053 [INFO][4614] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" host="localhost" Sep 11 23:32:05.094621 containerd[1515]: 2025-09-11 23:32:05.061 [INFO][4614] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" host="localhost" Sep 11 23:32:05.094621 containerd[1515]: 2025-09-11 23:32:05.061 [INFO][4614] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" host="localhost" Sep 11 23:32:05.094621 containerd[1515]: 2025-09-11 23:32:05.061 [INFO][4614] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:05.094621 containerd[1515]: 2025-09-11 23:32:05.061 [INFO][4614] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" HandleID="k8s-pod-network.aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Workload="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.095307 containerd[1515]: 2025-09-11 23:32:05.065 [INFO][4596] cni-plugin/k8s.go 418: Populated endpoint ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--d725f-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"32dd7b29-d990-4d4c-8903-e51a4239fdc5", ResourceVersion:"847", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-d725f", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2e2eb2c547c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:05.095397 containerd[1515]: 2025-09-11 23:32:05.065 [INFO][4596] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.095397 containerd[1515]: 2025-09-11 23:32:05.065 [INFO][4596] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2e2eb2c547c ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.095397 containerd[1515]: 2025-09-11 23:32:05.072 [INFO][4596] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.095458 containerd[1515]: 2025-09-11 23:32:05.073 [INFO][4596] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--d725f-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"32dd7b29-d990-4d4c-8903-e51a4239fdc5", ResourceVersion:"847", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45", Pod:"coredns-674b8bbfcf-d725f", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2e2eb2c547c", MAC:"f2:61:73:21:5a:bf", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:05.095458 containerd[1515]: 2025-09-11 23:32:05.087 [INFO][4596] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" Namespace="kube-system" Pod="coredns-674b8bbfcf-d725f" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--d725f-eth0" Sep 11 23:32:05.132128 containerd[1515]: time="2025-09-11T23:32:05.132075385Z" level=info msg="connecting to shim aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45" address="unix:///run/containerd/s/a3d0c273ecc97fd1cb60795d50f74335007e54da070263f30c89d269439fc65f" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:05.157666 systemd[1]: Started cri-containerd-aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45.scope - libcontainer container aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45. Sep 11 23:32:05.173223 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:05.203620 containerd[1515]: time="2025-09-11T23:32:05.203525351Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-d725f,Uid:32dd7b29-d990-4d4c-8903-e51a4239fdc5,Namespace:kube-system,Attempt:0,} returns sandbox id \"aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45\"" Sep 11 23:32:05.205488 kubelet[2668]: E0911 23:32:05.204903 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:05.211167 containerd[1515]: time="2025-09-11T23:32:05.211117083Z" level=info msg="CreateContainer within sandbox \"aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 11 23:32:05.235808 containerd[1515]: time="2025-09-11T23:32:05.235757050Z" level=info msg="Container e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:05.243803 containerd[1515]: time="2025-09-11T23:32:05.243744049Z" level=info msg="CreateContainer within sandbox \"aed914358fd74def4cc6341ede6da8ae719f9e7b7a8f887905554c282fc19f45\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13\"" Sep 11 23:32:05.244566 containerd[1515]: time="2025-09-11T23:32:05.244529144Z" level=info msg="StartContainer for \"e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13\"" Sep 11 23:32:05.245359 containerd[1515]: time="2025-09-11T23:32:05.245318120Z" level=info msg="connecting to shim e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13" address="unix:///run/containerd/s/a3d0c273ecc97fd1cb60795d50f74335007e54da070263f30c89d269439fc65f" protocol=ttrpc version=3 Sep 11 23:32:05.268669 systemd[1]: Started cri-containerd-e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13.scope - libcontainer container e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13. Sep 11 23:32:05.328443 containerd[1515]: time="2025-09-11T23:32:05.327636808Z" level=info msg="StartContainer for \"e91047c551df8a82c5c3f535584106c0b42cd1b267f619990d047e64dec1ea13\" returns successfully" Sep 11 23:32:05.650658 systemd-networkd[1433]: cali6b880faf9ef: Gained IPv6LL Sep 11 23:32:05.865028 kubelet[2668]: E0911 23:32:05.864549 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:05.865432 containerd[1515]: time="2025-09-11T23:32:05.864738484Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-zjzkc,Uid:cd9cb3e9-4e82-450d-850d-add7a5cdeee3,Namespace:calico-apiserver,Attempt:0,}" Sep 11 23:32:05.865432 containerd[1515]: time="2025-09-11T23:32:05.864904295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n5ml6,Uid:af6fbb68-721d-41a3-86cc-0d3211856c13,Namespace:kube-system,Attempt:0,}" Sep 11 23:32:05.907593 systemd-networkd[1433]: cali9315fc3eab2: Gained IPv6LL Sep 11 23:32:06.042046 kubelet[2668]: E0911 23:32:06.040789 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:06.064173 kubelet[2668]: I0911 23:32:06.064087 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-56f49b666f-ckb7g" podStartSLOduration=26.211238933 podStartE2EDuration="29.064069199s" podCreationTimestamp="2025-09-11 23:31:37 +0000 UTC" firstStartedPulling="2025-09-11 23:32:02.105668209 +0000 UTC m=+41.355188329" lastFinishedPulling="2025-09-11 23:32:04.958498475 +0000 UTC m=+44.208018595" observedRunningTime="2025-09-11 23:32:06.061184201 +0000 UTC m=+45.310704281" watchObservedRunningTime="2025-09-11 23:32:06.064069199 +0000 UTC m=+45.313589319" Sep 11 23:32:06.131154 systemd-networkd[1433]: cali963bfc87209: Link UP Sep 11 23:32:06.131438 systemd-networkd[1433]: cali963bfc87209: Gained carrier Sep 11 23:32:06.165307 kubelet[2668]: I0911 23:32:06.164890 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-d725f" podStartSLOduration=38.164871675 podStartE2EDuration="38.164871675s" podCreationTimestamp="2025-09-11 23:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:32:06.090686265 +0000 UTC m=+45.340206385" watchObservedRunningTime="2025-09-11 23:32:06.164871675 +0000 UTC m=+45.414391795" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.935 [INFO][4750] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0 calico-apiserver-56f49b666f- calico-apiserver cd9cb3e9-4e82-450d-850d-add7a5cdeee3 850 0 2025-09-11 23:31:37 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:56f49b666f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-56f49b666f-zjzkc eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali963bfc87209 [] [] }} ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.937 [INFO][4750] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.995 [INFO][4778] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" HandleID="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Workload="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.995 [INFO][4778] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" HandleID="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Workload="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000137ae0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-56f49b666f-zjzkc", "timestamp":"2025-09-11 23:32:05.99541032 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.995 [INFO][4778] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.995 [INFO][4778] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:05.995 [INFO][4778] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.017 [INFO][4778] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.033 [INFO][4778] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.052 [INFO][4778] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.062 [INFO][4778] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.070 [INFO][4778] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.070 [INFO][4778] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.079 [INFO][4778] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.092 [INFO][4778] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.116 [INFO][4778] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.117 [INFO][4778] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" host="localhost" Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.117 [INFO][4778] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:06.168491 containerd[1515]: 2025-09-11 23:32:06.117 [INFO][4778] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" HandleID="k8s-pod-network.29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Workload="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.169181 containerd[1515]: 2025-09-11 23:32:06.126 [INFO][4750] cni-plugin/k8s.go 418: Populated endpoint ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0", GenerateName:"calico-apiserver-56f49b666f-", Namespace:"calico-apiserver", SelfLink:"", UID:"cd9cb3e9-4e82-450d-850d-add7a5cdeee3", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56f49b666f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-56f49b666f-zjzkc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali963bfc87209", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:06.169181 containerd[1515]: 2025-09-11 23:32:06.126 [INFO][4750] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.169181 containerd[1515]: 2025-09-11 23:32:06.126 [INFO][4750] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali963bfc87209 ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.169181 containerd[1515]: 2025-09-11 23:32:06.133 [INFO][4750] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.169181 containerd[1515]: 2025-09-11 23:32:06.136 [INFO][4750] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0", GenerateName:"calico-apiserver-56f49b666f-", Namespace:"calico-apiserver", SelfLink:"", UID:"cd9cb3e9-4e82-450d-850d-add7a5cdeee3", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"56f49b666f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe", Pod:"calico-apiserver-56f49b666f-zjzkc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali963bfc87209", MAC:"9e:ab:5a:be:6a:31", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:06.169181 containerd[1515]: 2025-09-11 23:32:06.165 [INFO][4750] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" Namespace="calico-apiserver" Pod="calico-apiserver-56f49b666f-zjzkc" WorkloadEndpoint="localhost-k8s-calico--apiserver--56f49b666f--zjzkc-eth0" Sep 11 23:32:06.321160 systemd-networkd[1433]: calie45dae97c8c: Link UP Sep 11 23:32:06.321882 systemd-networkd[1433]: calie45dae97c8c: Gained carrier Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:05.944 [INFO][4762] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0 coredns-674b8bbfcf- kube-system af6fbb68-721d-41a3-86cc-0d3211856c13 853 0 2025-09-11 23:31:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-n5ml6 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calie45dae97c8c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:05.945 [INFO][4762] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.017 [INFO][4784] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" HandleID="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Workload="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.018 [INFO][4784] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" HandleID="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Workload="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c650), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-n5ml6", "timestamp":"2025-09-11 23:32:06.016915683 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.018 [INFO][4784] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.117 [INFO][4784] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.117 [INFO][4784] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.144 [INFO][4784] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.167 [INFO][4784] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.177 [INFO][4784] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.219 [INFO][4784] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.223 [INFO][4784] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.223 [INFO][4784] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.227 [INFO][4784] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.292 [INFO][4784] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.307 [INFO][4784] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.308 [INFO][4784] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" host="localhost" Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.309 [INFO][4784] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 23:32:06.342520 containerd[1515]: 2025-09-11 23:32:06.309 [INFO][4784] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" HandleID="k8s-pod-network.945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Workload="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.343303 containerd[1515]: 2025-09-11 23:32:06.316 [INFO][4762] cni-plugin/k8s.go 418: Populated endpoint ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"af6fbb68-721d-41a3-86cc-0d3211856c13", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-n5ml6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calie45dae97c8c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:06.343303 containerd[1515]: 2025-09-11 23:32:06.316 [INFO][4762] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.343303 containerd[1515]: 2025-09-11 23:32:06.316 [INFO][4762] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie45dae97c8c ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.343303 containerd[1515]: 2025-09-11 23:32:06.322 [INFO][4762] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.343303 containerd[1515]: 2025-09-11 23:32:06.322 [INFO][4762] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"af6fbb68-721d-41a3-86cc-0d3211856c13", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 23, 31, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe", Pod:"coredns-674b8bbfcf-n5ml6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calie45dae97c8c", MAC:"06:9a:83:b4:b2:41", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 23:32:06.343303 containerd[1515]: 2025-09-11 23:32:06.339 [INFO][4762] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" Namespace="kube-system" Pod="coredns-674b8bbfcf-n5ml6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--n5ml6-eth0" Sep 11 23:32:06.461236 containerd[1515]: time="2025-09-11T23:32:06.461059598Z" level=info msg="connecting to shim 29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe" address="unix:///run/containerd/s/bcc5704c1d064c9d672a2b4f941a60dcbd33cdbd5c1f4a16b81f2d56cd479e73" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:06.495508 containerd[1515]: time="2025-09-11T23:32:06.495134896Z" level=info msg="connecting to shim 945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe" address="unix:///run/containerd/s/958dc46055839741e8bc49582498e903a468af4cad830b9c450d0f112d54b93c" namespace=k8s.io protocol=ttrpc version=3 Sep 11 23:32:06.514695 systemd[1]: Started cri-containerd-29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe.scope - libcontainer container 29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe. Sep 11 23:32:06.540056 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:06.553762 systemd[1]: Started cri-containerd-945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe.scope - libcontainer container 945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe. Sep 11 23:32:06.578734 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 23:32:06.599620 containerd[1515]: time="2025-09-11T23:32:06.598494588Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-56f49b666f-zjzkc,Uid:cd9cb3e9-4e82-450d-850d-add7a5cdeee3,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe\"" Sep 11 23:32:06.611673 containerd[1515]: time="2025-09-11T23:32:06.611557685Z" level=info msg="CreateContainer within sandbox \"29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 11 23:32:06.622717 containerd[1515]: time="2025-09-11T23:32:06.622533438Z" level=info msg="Container 36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:06.639124 containerd[1515]: time="2025-09-11T23:32:06.639079173Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-n5ml6,Uid:af6fbb68-721d-41a3-86cc-0d3211856c13,Namespace:kube-system,Attempt:0,} returns sandbox id \"945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe\"" Sep 11 23:32:06.639952 containerd[1515]: time="2025-09-11T23:32:06.639797462Z" level=info msg="CreateContainer within sandbox \"29f069db71e7c0233ef2ab21d8a2bd2ae0bc713e87dcfef098c7ab5516e8c9fe\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3\"" Sep 11 23:32:06.640871 containerd[1515]: time="2025-09-11T23:32:06.640655081Z" level=info msg="StartContainer for \"36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3\"" Sep 11 23:32:06.641152 kubelet[2668]: E0911 23:32:06.641128 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:06.645564 containerd[1515]: time="2025-09-11T23:32:06.645457051Z" level=info msg="connecting to shim 36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3" address="unix:///run/containerd/s/bcc5704c1d064c9d672a2b4f941a60dcbd33cdbd5c1f4a16b81f2d56cd479e73" protocol=ttrpc version=3 Sep 11 23:32:06.664786 containerd[1515]: time="2025-09-11T23:32:06.664723693Z" level=info msg="CreateContainer within sandbox \"945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 11 23:32:06.688700 systemd[1]: Started cri-containerd-36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3.scope - libcontainer container 36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3. Sep 11 23:32:06.768935 containerd[1515]: time="2025-09-11T23:32:06.768887760Z" level=info msg="Container 9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:06.770935 containerd[1515]: time="2025-09-11T23:32:06.770896978Z" level=info msg="StartContainer for \"36d315f4b1f46163bf0a070f211807b798a59013e3331219b61b744cd9e46fa3\" returns successfully" Sep 11 23:32:06.780008 containerd[1515]: time="2025-09-11T23:32:06.779914716Z" level=info msg="CreateContainer within sandbox \"945a09cb1fd84239938c4ba8b4bb846e2a3b869d0043f5b36636bc815d1a2afe\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba\"" Sep 11 23:32:06.781040 containerd[1515]: time="2025-09-11T23:32:06.780980469Z" level=info msg="StartContainer for \"9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba\"" Sep 11 23:32:06.783002 containerd[1515]: time="2025-09-11T23:32:06.782961765Z" level=info msg="connecting to shim 9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba" address="unix:///run/containerd/s/958dc46055839741e8bc49582498e903a468af4cad830b9c450d0f112d54b93c" protocol=ttrpc version=3 Sep 11 23:32:06.818689 systemd[1]: Started cri-containerd-9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba.scope - libcontainer container 9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba. Sep 11 23:32:06.909105 containerd[1515]: time="2025-09-11T23:32:06.909037736Z" level=info msg="StartContainer for \"9ebe356e68154f1f037ece8c5519376beef3229189f3f4ad5ed4b49c0aa467ba\" returns successfully" Sep 11 23:32:06.930605 systemd-networkd[1433]: cali2e2eb2c547c: Gained IPv6LL Sep 11 23:32:07.047258 kubelet[2668]: E0911 23:32:07.047137 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:07.050189 kubelet[2668]: E0911 23:32:07.050117 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:07.050601 kubelet[2668]: I0911 23:32:07.050575 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:32:07.102151 kubelet[2668]: I0911 23:32:07.102071 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-n5ml6" podStartSLOduration=39.1020384 podStartE2EDuration="39.1020384s" podCreationTimestamp="2025-09-11 23:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:32:07.07690003 +0000 UTC m=+46.326420230" watchObservedRunningTime="2025-09-11 23:32:07.1020384 +0000 UTC m=+46.351558520" Sep 11 23:32:07.121491 kubelet[2668]: I0911 23:32:07.121412 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-56f49b666f-zjzkc" podStartSLOduration=30.121393222 podStartE2EDuration="30.121393222s" podCreationTimestamp="2025-09-11 23:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 23:32:07.121016036 +0000 UTC m=+46.370536156" watchObservedRunningTime="2025-09-11 23:32:07.121393222 +0000 UTC m=+46.370913342" Sep 11 23:32:07.315613 systemd-networkd[1433]: cali963bfc87209: Gained IPv6LL Sep 11 23:32:07.541185 containerd[1515]: time="2025-09-11T23:32:07.541124728Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:07.543021 containerd[1515]: time="2025-09-11T23:32:07.542971212Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=48134957" Sep 11 23:32:07.544140 containerd[1515]: time="2025-09-11T23:32:07.544114529Z" level=info msg="ImageCreate event name:\"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:07.546280 containerd[1515]: time="2025-09-11T23:32:07.546240072Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:07.546926 containerd[1515]: time="2025-09-11T23:32:07.546894396Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"49504166\" in 2.586719401s" Sep 11 23:32:07.546991 containerd[1515]: time="2025-09-11T23:32:07.546929678Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\"" Sep 11 23:32:07.552664 containerd[1515]: time="2025-09-11T23:32:07.552563537Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 11 23:32:07.572500 containerd[1515]: time="2025-09-11T23:32:07.572365709Z" level=info msg="CreateContainer within sandbox \"c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 11 23:32:07.582862 containerd[1515]: time="2025-09-11T23:32:07.582815452Z" level=info msg="Container 4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:07.595478 containerd[1515]: time="2025-09-11T23:32:07.594342387Z" level=info msg="CreateContainer within sandbox \"c87eb6c828335787f449ecfd4a5ab618313279bf933c4ad4f21272d5783b64ca\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a\"" Sep 11 23:32:07.596657 containerd[1515]: time="2025-09-11T23:32:07.596613300Z" level=info msg="StartContainer for \"4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a\"" Sep 11 23:32:07.598507 containerd[1515]: time="2025-09-11T23:32:07.597829421Z" level=info msg="connecting to shim 4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a" address="unix:///run/containerd/s/4a29ed1bba28f78346e58bff1d88fe808b5fece4e32638b3623957d2e1f26cf1" protocol=ttrpc version=3 Sep 11 23:32:07.634626 systemd-networkd[1433]: calie45dae97c8c: Gained IPv6LL Sep 11 23:32:07.638704 systemd[1]: Started cri-containerd-4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a.scope - libcontainer container 4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a. Sep 11 23:32:07.739407 containerd[1515]: time="2025-09-11T23:32:07.739368620Z" level=info msg="StartContainer for \"4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a\" returns successfully" Sep 11 23:32:08.056141 kubelet[2668]: E0911 23:32:08.056090 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:08.056551 kubelet[2668]: E0911 23:32:08.056255 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:08.056551 kubelet[2668]: I0911 23:32:08.056362 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:32:08.071571 kubelet[2668]: I0911 23:32:08.071477 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-cb484f97b-jkcgr" podStartSLOduration=21.81118399 podStartE2EDuration="26.0714347s" podCreationTimestamp="2025-09-11 23:31:42 +0000 UTC" firstStartedPulling="2025-09-11 23:32:03.290843968 +0000 UTC m=+42.540364088" lastFinishedPulling="2025-09-11 23:32:07.551094678 +0000 UTC m=+46.800614798" observedRunningTime="2025-09-11 23:32:08.069297039 +0000 UTC m=+47.318817199" watchObservedRunningTime="2025-09-11 23:32:08.0714347 +0000 UTC m=+47.320954820" Sep 11 23:32:08.134683 containerd[1515]: time="2025-09-11T23:32:08.134625908Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a\" id:\"ea0cc6cd071d78c3ed258acee2b7fd56aea76b7751a95abb1dad6bc11563d9a9\" pid:5054 exited_at:{seconds:1757633528 nanos:122316536}" Sep 11 23:32:09.058692 kubelet[2668]: I0911 23:32:09.058653 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:32:09.060202 kubelet[2668]: E0911 23:32:09.059832 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 23:32:09.143733 systemd[1]: Started sshd@8-10.0.0.19:22-10.0.0.1:35280.service - OpenSSH per-connection server daemon (10.0.0.1:35280). Sep 11 23:32:09.258492 sshd[5075]: Accepted publickey for core from 10.0.0.1 port 35280 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:09.260861 sshd-session[5075]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:09.269880 systemd-logind[1485]: New session 9 of user core. Sep 11 23:32:09.275698 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 11 23:32:09.342376 containerd[1515]: time="2025-09-11T23:32:09.342139435Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:09.343072 containerd[1515]: time="2025-09-11T23:32:09.342881803Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8227489" Sep 11 23:32:09.344128 containerd[1515]: time="2025-09-11T23:32:09.344061480Z" level=info msg="ImageCreate event name:\"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:09.346641 containerd[1515]: time="2025-09-11T23:32:09.346551681Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:09.347005 containerd[1515]: time="2025-09-11T23:32:09.346969708Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"9596730\" in 1.794350127s" Sep 11 23:32:09.347054 containerd[1515]: time="2025-09-11T23:32:09.347009791Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\"" Sep 11 23:32:09.351641 containerd[1515]: time="2025-09-11T23:32:09.351594368Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 11 23:32:09.356095 containerd[1515]: time="2025-09-11T23:32:09.356043256Z" level=info msg="CreateContainer within sandbox \"05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 11 23:32:09.377495 containerd[1515]: time="2025-09-11T23:32:09.376625029Z" level=info msg="Container 044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:09.387033 containerd[1515]: time="2025-09-11T23:32:09.386976619Z" level=info msg="CreateContainer within sandbox \"05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1\"" Sep 11 23:32:09.387762 containerd[1515]: time="2025-09-11T23:32:09.387716787Z" level=info msg="StartContainer for \"044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1\"" Sep 11 23:32:09.389355 containerd[1515]: time="2025-09-11T23:32:09.389264607Z" level=info msg="connecting to shim 044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1" address="unix:///run/containerd/s/ee9fdbcaaa58e16e816071da17e823d432c404e5bbc64e9418325c2646d5ae54" protocol=ttrpc version=3 Sep 11 23:32:09.434986 systemd[1]: Started cri-containerd-044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1.scope - libcontainer container 044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1. Sep 11 23:32:09.567437 containerd[1515]: time="2025-09-11T23:32:09.567383904Z" level=info msg="StartContainer for \"044c4712d9f2198ba4d371a8665930b0ce29963fe07dd3733bcfee080824c4d1\" returns successfully" Sep 11 23:32:09.585106 sshd[5082]: Connection closed by 10.0.0.1 port 35280 Sep 11 23:32:09.585932 sshd-session[5075]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:09.591304 systemd-logind[1485]: Session 9 logged out. Waiting for processes to exit. Sep 11 23:32:09.591672 systemd[1]: session-9.scope: Deactivated successfully. Sep 11 23:32:09.592874 systemd[1]: sshd@8-10.0.0.19:22-10.0.0.1:35280.service: Deactivated successfully. Sep 11 23:32:09.598769 systemd-logind[1485]: Removed session 9. Sep 11 23:32:11.113897 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4175267813.mount: Deactivated successfully. Sep 11 23:32:11.975911 containerd[1515]: time="2025-09-11T23:32:11.975867281Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:11.977560 containerd[1515]: time="2025-09-11T23:32:11.977517345Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=61845332" Sep 11 23:32:11.978656 containerd[1515]: time="2025-09-11T23:32:11.978601772Z" level=info msg="ImageCreate event name:\"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:11.982255 containerd[1515]: time="2025-09-11T23:32:11.982200838Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:11.984218 containerd[1515]: time="2025-09-11T23:32:11.984080915Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"61845178\" in 2.632435705s" Sep 11 23:32:11.984218 containerd[1515]: time="2025-09-11T23:32:11.984121758Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\"" Sep 11 23:32:11.986758 containerd[1515]: time="2025-09-11T23:32:11.986711160Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 11 23:32:11.990423 containerd[1515]: time="2025-09-11T23:32:11.990380910Z" level=info msg="CreateContainer within sandbox \"33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 11 23:32:12.000977 containerd[1515]: time="2025-09-11T23:32:12.000546746Z" level=info msg="Container 64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:12.016978 containerd[1515]: time="2025-09-11T23:32:12.016929515Z" level=info msg="CreateContainer within sandbox \"33d35e46e9e36c64fd13a983102691cdca90ba9aa4809a237bc9d3faa629786f\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3\"" Sep 11 23:32:12.018104 containerd[1515]: time="2025-09-11T23:32:12.017693642Z" level=info msg="StartContainer for \"64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3\"" Sep 11 23:32:12.023482 containerd[1515]: time="2025-09-11T23:32:12.023427555Z" level=info msg="connecting to shim 64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3" address="unix:///run/containerd/s/9f0bd42a089419fd6709e62394a6c009514d65f89b562beb4dced3fd2d92fc42" protocol=ttrpc version=3 Sep 11 23:32:12.050711 systemd[1]: Started cri-containerd-64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3.scope - libcontainer container 64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3. Sep 11 23:32:12.100435 containerd[1515]: time="2025-09-11T23:32:12.100386736Z" level=info msg="StartContainer for \"64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3\" returns successfully" Sep 11 23:32:13.098659 kubelet[2668]: I0911 23:32:13.098344 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-s852n" podStartSLOduration=24.578940715999998 podStartE2EDuration="32.098328076s" podCreationTimestamp="2025-09-11 23:31:41 +0000 UTC" firstStartedPulling="2025-09-11 23:32:04.4657291 +0000 UTC m=+43.715249220" lastFinishedPulling="2025-09-11 23:32:11.98511646 +0000 UTC m=+51.234636580" observedRunningTime="2025-09-11 23:32:13.097879289 +0000 UTC m=+52.347399409" watchObservedRunningTime="2025-09-11 23:32:13.098328076 +0000 UTC m=+52.347848196" Sep 11 23:32:13.216380 containerd[1515]: time="2025-09-11T23:32:13.216328115Z" level=info msg="TaskExit event in podsandbox handler container_id:\"64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3\" id:\"c020732d2b27af01129d2687a62442e308cf8b43977cf300afe0885a26c3764c\" pid:5187 exit_status:1 exited_at:{seconds:1757633533 nanos:215943731}" Sep 11 23:32:13.699346 containerd[1515]: time="2025-09-11T23:32:13.699294096Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:13.699736 containerd[1515]: time="2025-09-11T23:32:13.699695160Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=13761208" Sep 11 23:32:13.701899 containerd[1515]: time="2025-09-11T23:32:13.700761785Z" level=info msg="ImageCreate event name:\"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:13.703972 containerd[1515]: time="2025-09-11T23:32:13.703923017Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 23:32:13.706182 containerd[1515]: time="2025-09-11T23:32:13.706137111Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"15130401\" in 1.719371388s" Sep 11 23:32:13.706182 containerd[1515]: time="2025-09-11T23:32:13.706182554Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\"" Sep 11 23:32:13.719960 containerd[1515]: time="2025-09-11T23:32:13.719900986Z" level=info msg="CreateContainer within sandbox \"05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 11 23:32:13.744624 containerd[1515]: time="2025-09-11T23:32:13.743409172Z" level=info msg="Container 2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c: CDI devices from CRI Config.CDIDevices: []" Sep 11 23:32:13.881865 containerd[1515]: time="2025-09-11T23:32:13.881643599Z" level=info msg="CreateContainer within sandbox \"05dd5fbda3b7b336065df426f4d7e71f2f218d7725ade412d412c0e2fda59464\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c\"" Sep 11 23:32:13.882626 containerd[1515]: time="2025-09-11T23:32:13.882564975Z" level=info msg="StartContainer for \"2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c\"" Sep 11 23:32:13.884515 containerd[1515]: time="2025-09-11T23:32:13.884435088Z" level=info msg="connecting to shim 2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c" address="unix:///run/containerd/s/ee9fdbcaaa58e16e816071da17e823d432c404e5bbc64e9418325c2646d5ae54" protocol=ttrpc version=3 Sep 11 23:32:13.911716 systemd[1]: Started cri-containerd-2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c.scope - libcontainer container 2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c. Sep 11 23:32:14.038888 containerd[1515]: time="2025-09-11T23:32:14.038845264Z" level=info msg="StartContainer for \"2b611e1dd191129c215f74a12401091a283a9c4da1e57b3f3ef109083f60ef0c\" returns successfully" Sep 11 23:32:14.103654 kubelet[2668]: I0911 23:32:14.103562 2668 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-lgxh5" podStartSLOduration=22.826342899 podStartE2EDuration="32.103540332s" podCreationTimestamp="2025-09-11 23:31:42 +0000 UTC" firstStartedPulling="2025-09-11 23:32:04.430178393 +0000 UTC m=+43.679698513" lastFinishedPulling="2025-09-11 23:32:13.707375826 +0000 UTC m=+52.956895946" observedRunningTime="2025-09-11 23:32:14.103120427 +0000 UTC m=+53.352640547" watchObservedRunningTime="2025-09-11 23:32:14.103540332 +0000 UTC m=+53.353060492" Sep 11 23:32:14.173827 containerd[1515]: time="2025-09-11T23:32:14.173779253Z" level=info msg="TaskExit event in podsandbox handler container_id:\"64d7563bfed3ed4f66568a054b8b003786b38a39dacb4b59775c8561554354f3\" id:\"6e8058e8830368d5f84e1d610fc9eb23c461c41dd9ddb40a88782a4ce82aa0f2\" pid:5249 exit_status:1 exited_at:{seconds:1757633534 nanos:173412391}" Sep 11 23:32:14.599804 systemd[1]: Started sshd@9-10.0.0.19:22-10.0.0.1:50416.service - OpenSSH per-connection server daemon (10.0.0.1:50416). Sep 11 23:32:14.670123 sshd[5261]: Accepted publickey for core from 10.0.0.1 port 50416 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:14.671775 sshd-session[5261]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:14.676697 systemd-logind[1485]: New session 10 of user core. Sep 11 23:32:14.694378 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 11 23:32:14.924571 sshd[5264]: Connection closed by 10.0.0.1 port 50416 Sep 11 23:32:14.924263 sshd-session[5261]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:14.940221 systemd[1]: sshd@9-10.0.0.19:22-10.0.0.1:50416.service: Deactivated successfully. Sep 11 23:32:14.942955 systemd[1]: session-10.scope: Deactivated successfully. Sep 11 23:32:14.944608 systemd-logind[1485]: Session 10 logged out. Waiting for processes to exit. Sep 11 23:32:14.947993 systemd[1]: Started sshd@10-10.0.0.19:22-10.0.0.1:50432.service - OpenSSH per-connection server daemon (10.0.0.1:50432). Sep 11 23:32:14.949598 systemd-logind[1485]: Removed session 10. Sep 11 23:32:14.959527 kubelet[2668]: I0911 23:32:14.959484 2668 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 11 23:32:14.964639 kubelet[2668]: I0911 23:32:14.964590 2668 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 11 23:32:15.008402 sshd[5279]: Accepted publickey for core from 10.0.0.1 port 50432 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:15.009916 sshd-session[5279]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:15.015562 systemd-logind[1485]: New session 11 of user core. Sep 11 23:32:15.024753 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 11 23:32:15.296810 sshd[5284]: Connection closed by 10.0.0.1 port 50432 Sep 11 23:32:15.297640 sshd-session[5279]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:15.309936 systemd[1]: sshd@10-10.0.0.19:22-10.0.0.1:50432.service: Deactivated successfully. Sep 11 23:32:15.315429 systemd[1]: session-11.scope: Deactivated successfully. Sep 11 23:32:15.316574 systemd-logind[1485]: Session 11 logged out. Waiting for processes to exit. Sep 11 23:32:15.322557 systemd[1]: Started sshd@11-10.0.0.19:22-10.0.0.1:50442.service - OpenSSH per-connection server daemon (10.0.0.1:50442). Sep 11 23:32:15.324947 systemd-logind[1485]: Removed session 11. Sep 11 23:32:15.402310 sshd[5299]: Accepted publickey for core from 10.0.0.1 port 50442 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:15.403872 sshd-session[5299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:15.410323 systemd-logind[1485]: New session 12 of user core. Sep 11 23:32:15.416684 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 11 23:32:15.616401 sshd[5302]: Connection closed by 10.0.0.1 port 50442 Sep 11 23:32:15.616879 sshd-session[5299]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:15.621438 systemd[1]: sshd@11-10.0.0.19:22-10.0.0.1:50442.service: Deactivated successfully. Sep 11 23:32:15.623373 systemd[1]: session-12.scope: Deactivated successfully. Sep 11 23:32:15.624280 systemd-logind[1485]: Session 12 logged out. Waiting for processes to exit. Sep 11 23:32:15.625861 systemd-logind[1485]: Removed session 12. Sep 11 23:32:20.631724 systemd[1]: Started sshd@12-10.0.0.19:22-10.0.0.1:60294.service - OpenSSH per-connection server daemon (10.0.0.1:60294). Sep 11 23:32:20.705756 sshd[5326]: Accepted publickey for core from 10.0.0.1 port 60294 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:20.708857 sshd-session[5326]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:20.713856 systemd-logind[1485]: New session 13 of user core. Sep 11 23:32:20.723709 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 11 23:32:20.969576 sshd[5329]: Connection closed by 10.0.0.1 port 60294 Sep 11 23:32:20.969540 sshd-session[5326]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:20.985126 systemd[1]: sshd@12-10.0.0.19:22-10.0.0.1:60294.service: Deactivated successfully. Sep 11 23:32:20.987322 systemd[1]: session-13.scope: Deactivated successfully. Sep 11 23:32:20.988930 systemd-logind[1485]: Session 13 logged out. Waiting for processes to exit. Sep 11 23:32:20.992803 systemd[1]: Started sshd@13-10.0.0.19:22-10.0.0.1:60296.service - OpenSSH per-connection server daemon (10.0.0.1:60296). Sep 11 23:32:20.993981 systemd-logind[1485]: Removed session 13. Sep 11 23:32:21.047442 sshd[5344]: Accepted publickey for core from 10.0.0.1 port 60296 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:21.048954 sshd-session[5344]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:21.053591 systemd-logind[1485]: New session 14 of user core. Sep 11 23:32:21.064686 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 11 23:32:21.320558 sshd[5347]: Connection closed by 10.0.0.1 port 60296 Sep 11 23:32:21.321045 sshd-session[5344]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:21.331977 systemd[1]: sshd@13-10.0.0.19:22-10.0.0.1:60296.service: Deactivated successfully. Sep 11 23:32:21.333987 systemd[1]: session-14.scope: Deactivated successfully. Sep 11 23:32:21.334879 systemd-logind[1485]: Session 14 logged out. Waiting for processes to exit. Sep 11 23:32:21.337578 systemd[1]: Started sshd@14-10.0.0.19:22-10.0.0.1:60306.service - OpenSSH per-connection server daemon (10.0.0.1:60306). Sep 11 23:32:21.338110 systemd-logind[1485]: Removed session 14. Sep 11 23:32:21.400405 sshd[5358]: Accepted publickey for core from 10.0.0.1 port 60306 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:21.401954 sshd-session[5358]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:21.406994 systemd-logind[1485]: New session 15 of user core. Sep 11 23:32:21.415674 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 11 23:32:22.164481 sshd[5361]: Connection closed by 10.0.0.1 port 60306 Sep 11 23:32:22.165198 sshd-session[5358]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:22.175567 systemd[1]: sshd@14-10.0.0.19:22-10.0.0.1:60306.service: Deactivated successfully. Sep 11 23:32:22.179180 systemd[1]: session-15.scope: Deactivated successfully. Sep 11 23:32:22.182283 systemd-logind[1485]: Session 15 logged out. Waiting for processes to exit. Sep 11 23:32:22.185031 systemd[1]: Started sshd@15-10.0.0.19:22-10.0.0.1:60312.service - OpenSSH per-connection server daemon (10.0.0.1:60312). Sep 11 23:32:22.187836 systemd-logind[1485]: Removed session 15. Sep 11 23:32:22.251919 sshd[5381]: Accepted publickey for core from 10.0.0.1 port 60312 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:22.253438 sshd-session[5381]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:22.259905 systemd-logind[1485]: New session 16 of user core. Sep 11 23:32:22.270682 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 11 23:32:22.635810 sshd[5385]: Connection closed by 10.0.0.1 port 60312 Sep 11 23:32:22.636360 sshd-session[5381]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:22.649500 systemd[1]: sshd@15-10.0.0.19:22-10.0.0.1:60312.service: Deactivated successfully. Sep 11 23:32:22.651734 systemd[1]: session-16.scope: Deactivated successfully. Sep 11 23:32:22.653751 systemd-logind[1485]: Session 16 logged out. Waiting for processes to exit. Sep 11 23:32:22.657826 systemd[1]: Started sshd@16-10.0.0.19:22-10.0.0.1:60326.service - OpenSSH per-connection server daemon (10.0.0.1:60326). Sep 11 23:32:22.659701 systemd-logind[1485]: Removed session 16. Sep 11 23:32:22.717022 sshd[5397]: Accepted publickey for core from 10.0.0.1 port 60326 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:22.718535 sshd-session[5397]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:22.723679 systemd-logind[1485]: New session 17 of user core. Sep 11 23:32:22.736684 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 11 23:32:22.875612 sshd[5400]: Connection closed by 10.0.0.1 port 60326 Sep 11 23:32:22.875971 sshd-session[5397]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:22.880262 systemd-logind[1485]: Session 17 logged out. Waiting for processes to exit. Sep 11 23:32:22.881492 systemd[1]: sshd@16-10.0.0.19:22-10.0.0.1:60326.service: Deactivated successfully. Sep 11 23:32:22.884731 systemd[1]: session-17.scope: Deactivated successfully. Sep 11 23:32:22.887420 systemd-logind[1485]: Removed session 17. Sep 11 23:32:27.896173 systemd[1]: Started sshd@17-10.0.0.19:22-10.0.0.1:60340.service - OpenSSH per-connection server daemon (10.0.0.1:60340). Sep 11 23:32:27.954823 sshd[5417]: Accepted publickey for core from 10.0.0.1 port 60340 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:27.956190 sshd-session[5417]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:27.961413 systemd-logind[1485]: New session 18 of user core. Sep 11 23:32:27.975652 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 11 23:32:28.108597 sshd[5420]: Connection closed by 10.0.0.1 port 60340 Sep 11 23:32:28.109111 sshd-session[5417]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:28.114407 systemd[1]: sshd@17-10.0.0.19:22-10.0.0.1:60340.service: Deactivated successfully. Sep 11 23:32:28.117276 systemd[1]: session-18.scope: Deactivated successfully. Sep 11 23:32:28.119642 systemd-logind[1485]: Session 18 logged out. Waiting for processes to exit. Sep 11 23:32:28.121978 systemd-logind[1485]: Removed session 18. Sep 11 23:32:29.884824 kubelet[2668]: I0911 23:32:29.884406 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:32:30.528339 containerd[1515]: time="2025-09-11T23:32:30.528220038Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0bff34f37ab00d5728ef5161e14aaefbdfef7c72850d5c6e255da7c38ad16134\" id:\"5e11d5943ce44314b86f436e473e94b53ece6a6b17b1d61ee08320d2534ac270\" pid:5449 exited_at:{seconds:1757633550 nanos:527871654}" Sep 11 23:32:33.122713 systemd[1]: Started sshd@18-10.0.0.19:22-10.0.0.1:52302.service - OpenSSH per-connection server daemon (10.0.0.1:52302). Sep 11 23:32:33.175252 sshd[5462]: Accepted publickey for core from 10.0.0.1 port 52302 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:33.176825 sshd-session[5462]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:33.181380 systemd-logind[1485]: New session 19 of user core. Sep 11 23:32:33.188647 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 11 23:32:33.315393 sshd[5465]: Connection closed by 10.0.0.1 port 52302 Sep 11 23:32:33.316985 sshd-session[5462]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:33.321254 systemd[1]: sshd@18-10.0.0.19:22-10.0.0.1:52302.service: Deactivated successfully. Sep 11 23:32:33.323491 systemd[1]: session-19.scope: Deactivated successfully. Sep 11 23:32:33.324447 systemd-logind[1485]: Session 19 logged out. Waiting for processes to exit. Sep 11 23:32:33.326712 systemd-logind[1485]: Removed session 19. Sep 11 23:32:37.275725 kubelet[2668]: I0911 23:32:37.275679 2668 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 23:32:38.113061 containerd[1515]: time="2025-09-11T23:32:38.112910702Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4eeefff504bd746f5715662be24700d8f902b3578c5ae30042a8087810f1b02a\" id:\"b5f3420718665c3b5a7a36cb3d50b4def7008ee80f1c156608ce9c1739e6369a\" pid:5495 exited_at:{seconds:1757633558 nanos:112534512}" Sep 11 23:32:38.329624 systemd[1]: Started sshd@19-10.0.0.19:22-10.0.0.1:52326.service - OpenSSH per-connection server daemon (10.0.0.1:52326). Sep 11 23:32:38.402257 sshd[5506]: Accepted publickey for core from 10.0.0.1 port 52326 ssh2: RSA SHA256:pULdEgqoZ1CjXpNcHD/2mxhbP7BalAGKKlfd6deKmwI Sep 11 23:32:38.403970 sshd-session[5506]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 23:32:38.408743 systemd-logind[1485]: New session 20 of user core. Sep 11 23:32:38.416710 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 11 23:32:38.582248 sshd[5509]: Connection closed by 10.0.0.1 port 52326 Sep 11 23:32:38.582582 sshd-session[5506]: pam_unix(sshd:session): session closed for user core Sep 11 23:32:38.588562 systemd[1]: sshd@19-10.0.0.19:22-10.0.0.1:52326.service: Deactivated successfully. Sep 11 23:32:38.591397 systemd[1]: session-20.scope: Deactivated successfully. Sep 11 23:32:38.593390 systemd-logind[1485]: Session 20 logged out. Waiting for processes to exit. Sep 11 23:32:38.595067 systemd-logind[1485]: Removed session 20. Sep 11 23:32:38.865237 kubelet[2668]: E0911 23:32:38.865180 2668 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"