Sep 12 22:02:18.763180 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Sep 12 22:02:18.763201 kernel: Linux version 6.12.47-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Fri Sep 12 20:38:46 -00 2025 Sep 12 22:02:18.763211 kernel: KASLR enabled Sep 12 22:02:18.763216 kernel: efi: EFI v2.7 by EDK II Sep 12 22:02:18.763221 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb832018 ACPI 2.0=0xdbfd0018 RNG=0xdbfd0a18 MEMRESERVE=0xdb838218 Sep 12 22:02:18.763227 kernel: random: crng init done Sep 12 22:02:18.763233 kernel: secureboot: Secure boot disabled Sep 12 22:02:18.763239 kernel: ACPI: Early table checksum verification disabled Sep 12 22:02:18.763245 kernel: ACPI: RSDP 0x00000000DBFD0018 000024 (v02 BOCHS ) Sep 12 22:02:18.763252 kernel: ACPI: XSDT 0x00000000DBFD0F18 000064 (v01 BOCHS BXPC 00000001 01000013) Sep 12 22:02:18.763258 kernel: ACPI: FACP 0x00000000DBFD0B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763263 kernel: ACPI: DSDT 0x00000000DBF0E018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763269 kernel: ACPI: APIC 0x00000000DBFD0C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763275 kernel: ACPI: PPTT 0x00000000DBFD0098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763282 kernel: ACPI: GTDT 0x00000000DBFD0818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763289 kernel: ACPI: MCFG 0x00000000DBFD0A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763295 kernel: ACPI: SPCR 0x00000000DBFD0918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763301 kernel: ACPI: DBG2 0x00000000DBFD0998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763307 kernel: ACPI: IORT 0x00000000DBFD0198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 22:02:18.763313 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Sep 12 22:02:18.763319 kernel: ACPI: Use ACPI SPCR as default console: No Sep 12 22:02:18.763325 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Sep 12 22:02:18.763331 kernel: NODE_DATA(0) allocated [mem 0xdc965a00-0xdc96cfff] Sep 12 22:02:18.763336 kernel: Zone ranges: Sep 12 22:02:18.763342 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Sep 12 22:02:18.763350 kernel: DMA32 empty Sep 12 22:02:18.763356 kernel: Normal empty Sep 12 22:02:18.763361 kernel: Device empty Sep 12 22:02:18.763367 kernel: Movable zone start for each node Sep 12 22:02:18.763373 kernel: Early memory node ranges Sep 12 22:02:18.763379 kernel: node 0: [mem 0x0000000040000000-0x00000000db81ffff] Sep 12 22:02:18.763385 kernel: node 0: [mem 0x00000000db820000-0x00000000db82ffff] Sep 12 22:02:18.763391 kernel: node 0: [mem 0x00000000db830000-0x00000000dc09ffff] Sep 12 22:02:18.763397 kernel: node 0: [mem 0x00000000dc0a0000-0x00000000dc2dffff] Sep 12 22:02:18.763403 kernel: node 0: [mem 0x00000000dc2e0000-0x00000000dc36ffff] Sep 12 22:02:18.763409 kernel: node 0: [mem 0x00000000dc370000-0x00000000dc45ffff] Sep 12 22:02:18.763415 kernel: node 0: [mem 0x00000000dc460000-0x00000000dc52ffff] Sep 12 22:02:18.763422 kernel: node 0: [mem 0x00000000dc530000-0x00000000dc5cffff] Sep 12 22:02:18.763428 kernel: node 0: [mem 0x00000000dc5d0000-0x00000000dce1ffff] Sep 12 22:02:18.763434 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Sep 12 22:02:18.763442 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Sep 12 22:02:18.763449 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Sep 12 22:02:18.763455 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Sep 12 22:02:18.763463 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Sep 12 22:02:18.763469 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Sep 12 22:02:18.763476 kernel: cma: Reserved 16 MiB at 0x00000000d8000000 on node -1 Sep 12 22:02:18.763482 kernel: psci: probing for conduit method from ACPI. Sep 12 22:02:18.763488 kernel: psci: PSCIv1.1 detected in firmware. Sep 12 22:02:18.763495 kernel: psci: Using standard PSCI v0.2 function IDs Sep 12 22:02:18.763501 kernel: psci: Trusted OS migration not required Sep 12 22:02:18.763507 kernel: psci: SMC Calling Convention v1.1 Sep 12 22:02:18.763513 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Sep 12 22:02:18.763520 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Sep 12 22:02:18.763528 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Sep 12 22:02:18.763534 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Sep 12 22:02:18.763541 kernel: Detected PIPT I-cache on CPU0 Sep 12 22:02:18.763547 kernel: CPU features: detected: GIC system register CPU interface Sep 12 22:02:18.763553 kernel: CPU features: detected: Spectre-v4 Sep 12 22:02:18.763559 kernel: CPU features: detected: Spectre-BHB Sep 12 22:02:18.763566 kernel: CPU features: kernel page table isolation forced ON by KASLR Sep 12 22:02:18.763572 kernel: CPU features: detected: Kernel page table isolation (KPTI) Sep 12 22:02:18.763578 kernel: CPU features: detected: ARM erratum 1418040 Sep 12 22:02:18.763585 kernel: CPU features: detected: SSBS not fully self-synchronizing Sep 12 22:02:18.763591 kernel: alternatives: applying boot alternatives Sep 12 22:02:18.763598 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=319fa5fb212e5dd8bf766d2f9f0bbb61d6aa6c81f2813f4b5b49defba0af2b2f Sep 12 22:02:18.763606 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 12 22:02:18.763613 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 12 22:02:18.763619 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 12 22:02:18.763626 kernel: Fallback order for Node 0: 0 Sep 12 22:02:18.763633 kernel: Built 1 zonelists, mobility grouping on. Total pages: 643072 Sep 12 22:02:18.763639 kernel: Policy zone: DMA Sep 12 22:02:18.763645 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 12 22:02:18.763652 kernel: software IO TLB: SWIOTLB bounce buffer size adjusted to 2MB Sep 12 22:02:18.763658 kernel: software IO TLB: area num 4. Sep 12 22:02:18.763665 kernel: software IO TLB: SWIOTLB bounce buffer size roundup to 4MB Sep 12 22:02:18.763672 kernel: software IO TLB: mapped [mem 0x00000000d7c00000-0x00000000d8000000] (4MB) Sep 12 22:02:18.763679 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 12 22:02:18.763686 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 12 22:02:18.763693 kernel: rcu: RCU event tracing is enabled. Sep 12 22:02:18.763700 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 12 22:02:18.763706 kernel: Trampoline variant of Tasks RCU enabled. Sep 12 22:02:18.763712 kernel: Tracing variant of Tasks RCU enabled. Sep 12 22:02:18.763719 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 12 22:02:18.763725 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 12 22:02:18.763732 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 12 22:02:18.763739 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 12 22:02:18.763745 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Sep 12 22:02:18.763753 kernel: GICv3: 256 SPIs implemented Sep 12 22:02:18.763759 kernel: GICv3: 0 Extended SPIs implemented Sep 12 22:02:18.763765 kernel: Root IRQ handler: gic_handle_irq Sep 12 22:02:18.763772 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Sep 12 22:02:18.763778 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Sep 12 22:02:18.763784 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Sep 12 22:02:18.763791 kernel: ITS [mem 0x08080000-0x0809ffff] Sep 12 22:02:18.763797 kernel: ITS@0x0000000008080000: allocated 8192 Devices @40110000 (indirect, esz 8, psz 64K, shr 1) Sep 12 22:02:18.763804 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @40120000 (flat, esz 8, psz 64K, shr 1) Sep 12 22:02:18.763816 kernel: GICv3: using LPI property table @0x0000000040130000 Sep 12 22:02:18.763824 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040140000 Sep 12 22:02:18.763830 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 12 22:02:18.763839 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 12 22:02:18.763845 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Sep 12 22:02:18.763852 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Sep 12 22:02:18.763859 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Sep 12 22:02:18.763865 kernel: arm-pv: using stolen time PV Sep 12 22:02:18.763872 kernel: Console: colour dummy device 80x25 Sep 12 22:02:18.763878 kernel: ACPI: Core revision 20240827 Sep 12 22:02:18.763885 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Sep 12 22:02:18.763891 kernel: pid_max: default: 32768 minimum: 301 Sep 12 22:02:18.763898 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 12 22:02:18.763906 kernel: landlock: Up and running. Sep 12 22:02:18.763912 kernel: SELinux: Initializing. Sep 12 22:02:18.763919 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 12 22:02:18.763926 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 12 22:02:18.763932 kernel: rcu: Hierarchical SRCU implementation. Sep 12 22:02:18.763939 kernel: rcu: Max phase no-delay instances is 400. Sep 12 22:02:18.763945 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 12 22:02:18.763952 kernel: Remapping and enabling EFI services. Sep 12 22:02:18.763959 kernel: smp: Bringing up secondary CPUs ... Sep 12 22:02:18.763981 kernel: Detected PIPT I-cache on CPU1 Sep 12 22:02:18.763988 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Sep 12 22:02:18.763995 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040150000 Sep 12 22:02:18.764004 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 12 22:02:18.764011 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Sep 12 22:02:18.764018 kernel: Detected PIPT I-cache on CPU2 Sep 12 22:02:18.764025 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Sep 12 22:02:18.764032 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040160000 Sep 12 22:02:18.764040 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 12 22:02:18.764047 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Sep 12 22:02:18.764054 kernel: Detected PIPT I-cache on CPU3 Sep 12 22:02:18.764061 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Sep 12 22:02:18.764068 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040170000 Sep 12 22:02:18.764075 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 12 22:02:18.764081 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Sep 12 22:02:18.764088 kernel: smp: Brought up 1 node, 4 CPUs Sep 12 22:02:18.764095 kernel: SMP: Total of 4 processors activated. Sep 12 22:02:18.764103 kernel: CPU: All CPU(s) started at EL1 Sep 12 22:02:18.764110 kernel: CPU features: detected: 32-bit EL0 Support Sep 12 22:02:18.764117 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Sep 12 22:02:18.764124 kernel: CPU features: detected: Common not Private translations Sep 12 22:02:18.764131 kernel: CPU features: detected: CRC32 instructions Sep 12 22:02:18.764138 kernel: CPU features: detected: Enhanced Virtualization Traps Sep 12 22:02:18.764145 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Sep 12 22:02:18.764152 kernel: CPU features: detected: LSE atomic instructions Sep 12 22:02:18.764159 kernel: CPU features: detected: Privileged Access Never Sep 12 22:02:18.764167 kernel: CPU features: detected: RAS Extension Support Sep 12 22:02:18.764174 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Sep 12 22:02:18.764181 kernel: alternatives: applying system-wide alternatives Sep 12 22:02:18.764188 kernel: CPU features: detected: Hardware dirty bit management on CPU0-3 Sep 12 22:02:18.764195 kernel: Memory: 2424480K/2572288K available (11136K kernel code, 2440K rwdata, 9068K rodata, 38976K init, 1038K bss, 125472K reserved, 16384K cma-reserved) Sep 12 22:02:18.764203 kernel: devtmpfs: initialized Sep 12 22:02:18.764210 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 12 22:02:18.764216 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 12 22:02:18.764223 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Sep 12 22:02:18.764231 kernel: 0 pages in range for non-PLT usage Sep 12 22:02:18.764238 kernel: 508560 pages in range for PLT usage Sep 12 22:02:18.764245 kernel: pinctrl core: initialized pinctrl subsystem Sep 12 22:02:18.764252 kernel: SMBIOS 3.0.0 present. Sep 12 22:02:18.764259 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Sep 12 22:02:18.764265 kernel: DMI: Memory slots populated: 1/1 Sep 12 22:02:18.764272 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 12 22:02:18.764279 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Sep 12 22:02:18.764286 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Sep 12 22:02:18.764295 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Sep 12 22:02:18.764302 kernel: audit: initializing netlink subsys (disabled) Sep 12 22:02:18.764308 kernel: audit: type=2000 audit(0.019:1): state=initialized audit_enabled=0 res=1 Sep 12 22:02:18.764316 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 12 22:02:18.764323 kernel: cpuidle: using governor menu Sep 12 22:02:18.764331 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Sep 12 22:02:18.764338 kernel: ASID allocator initialised with 32768 entries Sep 12 22:02:18.764345 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 12 22:02:18.764352 kernel: Serial: AMBA PL011 UART driver Sep 12 22:02:18.764360 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 12 22:02:18.764367 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Sep 12 22:02:18.764374 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Sep 12 22:02:18.764381 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Sep 12 22:02:18.764387 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 12 22:02:18.764394 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Sep 12 22:02:18.764401 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Sep 12 22:02:18.764408 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Sep 12 22:02:18.764415 kernel: ACPI: Added _OSI(Module Device) Sep 12 22:02:18.764423 kernel: ACPI: Added _OSI(Processor Device) Sep 12 22:02:18.764430 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 12 22:02:18.764437 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 12 22:02:18.764444 kernel: ACPI: Interpreter enabled Sep 12 22:02:18.764451 kernel: ACPI: Using GIC for interrupt routing Sep 12 22:02:18.764458 kernel: ACPI: MCFG table detected, 1 entries Sep 12 22:02:18.764464 kernel: ACPI: CPU0 has been hot-added Sep 12 22:02:18.764471 kernel: ACPI: CPU1 has been hot-added Sep 12 22:02:18.764478 kernel: ACPI: CPU2 has been hot-added Sep 12 22:02:18.764485 kernel: ACPI: CPU3 has been hot-added Sep 12 22:02:18.764493 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Sep 12 22:02:18.764500 kernel: printk: legacy console [ttyAMA0] enabled Sep 12 22:02:18.764506 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 12 22:02:18.764640 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 12 22:02:18.764706 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 12 22:02:18.764766 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 12 22:02:18.764838 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Sep 12 22:02:18.764902 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Sep 12 22:02:18.764912 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Sep 12 22:02:18.764919 kernel: PCI host bridge to bus 0000:00 Sep 12 22:02:18.765017 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Sep 12 22:02:18.765079 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Sep 12 22:02:18.765134 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Sep 12 22:02:18.765188 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 12 22:02:18.765269 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Sep 12 22:02:18.765342 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 12 22:02:18.765404 kernel: pci 0000:00:01.0: BAR 0 [io 0x0000-0x001f] Sep 12 22:02:18.765464 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff] Sep 12 22:02:18.765523 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Sep 12 22:02:18.765583 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Sep 12 22:02:18.765643 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff]: assigned Sep 12 22:02:18.765705 kernel: pci 0000:00:01.0: BAR 0 [io 0x1000-0x101f]: assigned Sep 12 22:02:18.765759 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Sep 12 22:02:18.765818 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Sep 12 22:02:18.765879 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Sep 12 22:02:18.765888 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Sep 12 22:02:18.765895 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Sep 12 22:02:18.765902 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Sep 12 22:02:18.765911 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Sep 12 22:02:18.765918 kernel: iommu: Default domain type: Translated Sep 12 22:02:18.765925 kernel: iommu: DMA domain TLB invalidation policy: strict mode Sep 12 22:02:18.765933 kernel: efivars: Registered efivars operations Sep 12 22:02:18.765939 kernel: vgaarb: loaded Sep 12 22:02:18.765946 kernel: clocksource: Switched to clocksource arch_sys_counter Sep 12 22:02:18.765953 kernel: VFS: Disk quotas dquot_6.6.0 Sep 12 22:02:18.765960 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 12 22:02:18.765967 kernel: pnp: PnP ACPI init Sep 12 22:02:18.766153 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Sep 12 22:02:18.766166 kernel: pnp: PnP ACPI: found 1 devices Sep 12 22:02:18.766173 kernel: NET: Registered PF_INET protocol family Sep 12 22:02:18.766181 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 12 22:02:18.766188 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 12 22:02:18.766196 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 12 22:02:18.766203 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 12 22:02:18.766210 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 12 22:02:18.766220 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 12 22:02:18.766227 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 12 22:02:18.766234 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 12 22:02:18.766241 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 12 22:02:18.766248 kernel: PCI: CLS 0 bytes, default 64 Sep 12 22:02:18.766255 kernel: kvm [1]: HYP mode not available Sep 12 22:02:18.766262 kernel: Initialise system trusted keyrings Sep 12 22:02:18.766269 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 12 22:02:18.766276 kernel: Key type asymmetric registered Sep 12 22:02:18.766284 kernel: Asymmetric key parser 'x509' registered Sep 12 22:02:18.766291 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 12 22:02:18.766298 kernel: io scheduler mq-deadline registered Sep 12 22:02:18.766305 kernel: io scheduler kyber registered Sep 12 22:02:18.766312 kernel: io scheduler bfq registered Sep 12 22:02:18.766320 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Sep 12 22:02:18.766327 kernel: ACPI: button: Power Button [PWRB] Sep 12 22:02:18.766334 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Sep 12 22:02:18.766401 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Sep 12 22:02:18.766412 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 12 22:02:18.766419 kernel: thunder_xcv, ver 1.0 Sep 12 22:02:18.766427 kernel: thunder_bgx, ver 1.0 Sep 12 22:02:18.766434 kernel: nicpf, ver 1.0 Sep 12 22:02:18.766441 kernel: nicvf, ver 1.0 Sep 12 22:02:18.766511 kernel: rtc-efi rtc-efi.0: registered as rtc0 Sep 12 22:02:18.766571 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-09-12T22:02:18 UTC (1757714538) Sep 12 22:02:18.766581 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 12 22:02:18.766588 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Sep 12 22:02:18.766597 kernel: watchdog: NMI not fully supported Sep 12 22:02:18.766604 kernel: watchdog: Hard watchdog permanently disabled Sep 12 22:02:18.766611 kernel: NET: Registered PF_INET6 protocol family Sep 12 22:02:18.766618 kernel: Segment Routing with IPv6 Sep 12 22:02:18.766624 kernel: In-situ OAM (IOAM) with IPv6 Sep 12 22:02:18.766631 kernel: NET: Registered PF_PACKET protocol family Sep 12 22:02:18.766638 kernel: Key type dns_resolver registered Sep 12 22:02:18.766645 kernel: registered taskstats version 1 Sep 12 22:02:18.766652 kernel: Loading compiled-in X.509 certificates Sep 12 22:02:18.766660 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.47-flatcar: 2d7730e6d35b3fbd1c590cd72a2500b2380c020e' Sep 12 22:02:18.766667 kernel: Demotion targets for Node 0: null Sep 12 22:02:18.766674 kernel: Key type .fscrypt registered Sep 12 22:02:18.766681 kernel: Key type fscrypt-provisioning registered Sep 12 22:02:18.766687 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 12 22:02:18.766694 kernel: ima: Allocated hash algorithm: sha1 Sep 12 22:02:18.766701 kernel: ima: No architecture policies found Sep 12 22:02:18.766708 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Sep 12 22:02:18.766716 kernel: clk: Disabling unused clocks Sep 12 22:02:18.766724 kernel: PM: genpd: Disabling unused power domains Sep 12 22:02:18.766730 kernel: Warning: unable to open an initial console. Sep 12 22:02:18.766738 kernel: Freeing unused kernel memory: 38976K Sep 12 22:02:18.766744 kernel: Run /init as init process Sep 12 22:02:18.766751 kernel: with arguments: Sep 12 22:02:18.766758 kernel: /init Sep 12 22:02:18.766765 kernel: with environment: Sep 12 22:02:18.766782 kernel: HOME=/ Sep 12 22:02:18.766790 kernel: TERM=linux Sep 12 22:02:18.766799 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 12 22:02:18.766807 systemd[1]: Successfully made /usr/ read-only. Sep 12 22:02:18.766861 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 22:02:18.766869 systemd[1]: Detected virtualization kvm. Sep 12 22:02:18.766877 systemd[1]: Detected architecture arm64. Sep 12 22:02:18.766884 systemd[1]: Running in initrd. Sep 12 22:02:18.766892 systemd[1]: No hostname configured, using default hostname. Sep 12 22:02:18.766903 systemd[1]: Hostname set to . Sep 12 22:02:18.766910 systemd[1]: Initializing machine ID from VM UUID. Sep 12 22:02:18.766918 systemd[1]: Queued start job for default target initrd.target. Sep 12 22:02:18.766926 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 22:02:18.766933 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 22:02:18.766942 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 12 22:02:18.766950 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 22:02:18.766957 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 12 22:02:18.766967 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 12 22:02:18.766991 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 12 22:02:18.767006 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 12 22:02:18.767014 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 22:02:18.767022 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 22:02:18.767029 systemd[1]: Reached target paths.target - Path Units. Sep 12 22:02:18.767037 systemd[1]: Reached target slices.target - Slice Units. Sep 12 22:02:18.767046 systemd[1]: Reached target swap.target - Swaps. Sep 12 22:02:18.767053 systemd[1]: Reached target timers.target - Timer Units. Sep 12 22:02:18.767061 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 22:02:18.767068 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 22:02:18.767075 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 12 22:02:18.767083 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 12 22:02:18.767090 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 22:02:18.767098 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 22:02:18.767106 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 22:02:18.767114 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 22:02:18.767121 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 12 22:02:18.767129 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 22:02:18.767136 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 12 22:02:18.767144 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 12 22:02:18.767152 systemd[1]: Starting systemd-fsck-usr.service... Sep 12 22:02:18.767159 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 22:02:18.767167 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 22:02:18.767176 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:02:18.767183 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 12 22:02:18.767191 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 22:02:18.767199 systemd[1]: Finished systemd-fsck-usr.service. Sep 12 22:02:18.767228 systemd-journald[244]: Collecting audit messages is disabled. Sep 12 22:02:18.767246 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 12 22:02:18.767255 systemd-journald[244]: Journal started Sep 12 22:02:18.767275 systemd-journald[244]: Runtime Journal (/run/log/journal/e1aa75063f484c41a890bad49dc22132) is 6M, max 48.5M, 42.4M free. Sep 12 22:02:18.760954 systemd-modules-load[245]: Inserted module 'overlay' Sep 12 22:02:18.771306 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 22:02:18.772997 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 12 22:02:18.774839 systemd-modules-load[245]: Inserted module 'br_netfilter' Sep 12 22:02:18.775613 kernel: Bridge firewalling registered Sep 12 22:02:18.777076 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:02:18.778419 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 22:02:18.782172 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 12 22:02:18.783666 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 22:02:18.785435 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 22:02:18.793998 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 22:02:18.796648 systemd-tmpfiles[265]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 12 22:02:18.797228 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 22:02:18.802002 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 22:02:18.803399 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 22:02:18.806793 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 22:02:18.809416 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 22:02:18.812848 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 22:02:18.814835 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 12 22:02:18.831754 dracut-cmdline[288]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=319fa5fb212e5dd8bf766d2f9f0bbb61d6aa6c81f2813f4b5b49defba0af2b2f Sep 12 22:02:18.839657 systemd-resolved[279]: Positive Trust Anchors: Sep 12 22:02:18.839679 systemd-resolved[279]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 22:02:18.839710 systemd-resolved[279]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 22:02:18.844583 systemd-resolved[279]: Defaulting to hostname 'linux'. Sep 12 22:02:18.845544 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 22:02:18.847578 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 22:02:18.902011 kernel: SCSI subsystem initialized Sep 12 22:02:18.906990 kernel: Loading iSCSI transport class v2.0-870. Sep 12 22:02:18.914001 kernel: iscsi: registered transport (tcp) Sep 12 22:02:18.927006 kernel: iscsi: registered transport (qla4xxx) Sep 12 22:02:18.927023 kernel: QLogic iSCSI HBA Driver Sep 12 22:02:18.943407 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 22:02:18.963275 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 22:02:18.968707 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 22:02:19.006286 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 12 22:02:19.008402 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 12 22:02:19.071010 kernel: raid6: neonx8 gen() 15760 MB/s Sep 12 22:02:19.087996 kernel: raid6: neonx4 gen() 15796 MB/s Sep 12 22:02:19.104993 kernel: raid6: neonx2 gen() 13193 MB/s Sep 12 22:02:19.121991 kernel: raid6: neonx1 gen() 10445 MB/s Sep 12 22:02:19.138991 kernel: raid6: int64x8 gen() 6895 MB/s Sep 12 22:02:19.155991 kernel: raid6: int64x4 gen() 7335 MB/s Sep 12 22:02:19.172989 kernel: raid6: int64x2 gen() 6098 MB/s Sep 12 22:02:19.189991 kernel: raid6: int64x1 gen() 5049 MB/s Sep 12 22:02:19.190013 kernel: raid6: using algorithm neonx4 gen() 15796 MB/s Sep 12 22:02:19.206995 kernel: raid6: .... xor() 12349 MB/s, rmw enabled Sep 12 22:02:19.207009 kernel: raid6: using neon recovery algorithm Sep 12 22:02:19.212005 kernel: xor: measuring software checksum speed Sep 12 22:02:19.212025 kernel: 8regs : 21624 MB/sec Sep 12 22:02:19.213072 kernel: 32regs : 21676 MB/sec Sep 12 22:02:19.213085 kernel: arm64_neon : 28061 MB/sec Sep 12 22:02:19.213094 kernel: xor: using function: arm64_neon (28061 MB/sec) Sep 12 22:02:19.265002 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 12 22:02:19.270932 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 12 22:02:19.273131 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 22:02:19.304660 systemd-udevd[498]: Using default interface naming scheme 'v255'. Sep 12 22:02:19.308816 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 22:02:19.310624 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 12 22:02:19.341723 dracut-pre-trigger[506]: rd.md=0: removing MD RAID activation Sep 12 22:02:19.362397 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 22:02:19.364337 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 22:02:19.415618 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 22:02:19.418828 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 12 22:02:19.465041 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Sep 12 22:02:19.471044 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 12 22:02:19.469697 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 22:02:19.475995 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 12 22:02:19.476028 kernel: GPT:9289727 != 19775487 Sep 12 22:02:19.476038 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 12 22:02:19.476053 kernel: GPT:9289727 != 19775487 Sep 12 22:02:19.476062 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 12 22:02:19.476071 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 22:02:19.469825 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:02:19.474270 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:02:19.479818 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:02:19.505908 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 12 22:02:19.509038 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 12 22:02:19.510028 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:02:19.529848 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 12 22:02:19.537924 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 12 22:02:19.544658 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 12 22:02:19.545606 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 12 22:02:19.547892 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 22:02:19.550028 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 22:02:19.551623 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 22:02:19.553802 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 12 22:02:19.555383 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 12 22:02:19.578995 disk-uuid[589]: Primary Header is updated. Sep 12 22:02:19.578995 disk-uuid[589]: Secondary Entries is updated. Sep 12 22:02:19.578995 disk-uuid[589]: Secondary Header is updated. Sep 12 22:02:19.583089 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 22:02:19.584552 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 12 22:02:20.589002 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 22:02:20.589560 disk-uuid[592]: The operation has completed successfully. Sep 12 22:02:20.613775 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 12 22:02:20.613876 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 12 22:02:20.646643 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 12 22:02:20.670742 sh[608]: Success Sep 12 22:02:20.682399 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 12 22:02:20.682443 kernel: device-mapper: uevent: version 1.0.3 Sep 12 22:02:20.683402 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 12 22:02:20.690030 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Sep 12 22:02:20.711788 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 12 22:02:20.714162 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 12 22:02:20.737151 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 12 22:02:20.742707 kernel: BTRFS: device fsid 254e43f1-b609-42b8-bcc5-437252095415 devid 1 transid 38 /dev/mapper/usr (253:0) scanned by mount (621) Sep 12 22:02:20.742738 kernel: BTRFS info (device dm-0): first mount of filesystem 254e43f1-b609-42b8-bcc5-437252095415 Sep 12 22:02:20.742748 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Sep 12 22:02:20.747209 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 12 22:02:20.747249 kernel: BTRFS info (device dm-0): enabling free space tree Sep 12 22:02:20.748297 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 12 22:02:20.749353 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 12 22:02:20.750488 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 12 22:02:20.751239 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 12 22:02:20.753942 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 12 22:02:20.778136 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (652) Sep 12 22:02:20.778178 kernel: BTRFS info (device vda6): first mount of filesystem 5dadbedd-e975-4944-978a-462cb6ec6aa0 Sep 12 22:02:20.780008 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 12 22:02:20.783041 kernel: BTRFS info (device vda6): turning on async discard Sep 12 22:02:20.783079 kernel: BTRFS info (device vda6): enabling free space tree Sep 12 22:02:20.786996 kernel: BTRFS info (device vda6): last unmount of filesystem 5dadbedd-e975-4944-978a-462cb6ec6aa0 Sep 12 22:02:20.788033 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 12 22:02:20.790282 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 12 22:02:20.851021 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 22:02:20.853541 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 22:02:20.893439 systemd-networkd[794]: lo: Link UP Sep 12 22:02:20.893452 systemd-networkd[794]: lo: Gained carrier Sep 12 22:02:20.894175 systemd-networkd[794]: Enumeration completed Sep 12 22:02:20.894439 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 22:02:20.894643 systemd-networkd[794]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:02:20.894647 systemd-networkd[794]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 12 22:02:20.895376 systemd-networkd[794]: eth0: Link UP Sep 12 22:02:20.895464 systemd-networkd[794]: eth0: Gained carrier Sep 12 22:02:20.895473 systemd-networkd[794]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:02:20.896195 systemd[1]: Reached target network.target - Network. Sep 12 22:02:20.907419 ignition[699]: Ignition 2.22.0 Sep 12 22:02:20.907435 ignition[699]: Stage: fetch-offline Sep 12 22:02:20.907463 ignition[699]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:02:20.907470 ignition[699]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 12 22:02:20.907540 ignition[699]: parsed url from cmdline: "" Sep 12 22:02:20.907543 ignition[699]: no config URL provided Sep 12 22:02:20.907548 ignition[699]: reading system config file "/usr/lib/ignition/user.ign" Sep 12 22:02:20.907554 ignition[699]: no config at "/usr/lib/ignition/user.ign" Sep 12 22:02:20.907572 ignition[699]: op(1): [started] loading QEMU firmware config module Sep 12 22:02:20.907576 ignition[699]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 12 22:02:20.912190 ignition[699]: op(1): [finished] loading QEMU firmware config module Sep 12 22:02:20.918038 systemd-networkd[794]: eth0: DHCPv4 address 10.0.0.29/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 12 22:02:20.955222 ignition[699]: parsing config with SHA512: 4ebf57ffafb06cd17fd8ebb3f915d8b2d282d254b0d41c3cc1417f91562a3f466ee7410ca15b994c7cf2cd538c01604f1720cf48d959803539abd972d00f4971 Sep 12 22:02:20.960177 unknown[699]: fetched base config from "system" Sep 12 22:02:20.960187 unknown[699]: fetched user config from "qemu" Sep 12 22:02:20.960629 ignition[699]: fetch-offline: fetch-offline passed Sep 12 22:02:20.960683 ignition[699]: Ignition finished successfully Sep 12 22:02:20.962842 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 22:02:20.964221 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 12 22:02:20.964932 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 12 22:02:20.990955 ignition[810]: Ignition 2.22.0 Sep 12 22:02:20.990984 ignition[810]: Stage: kargs Sep 12 22:02:20.991107 ignition[810]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:02:20.991115 ignition[810]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 12 22:02:20.994540 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 12 22:02:20.991789 ignition[810]: kargs: kargs passed Sep 12 22:02:20.991836 ignition[810]: Ignition finished successfully Sep 12 22:02:20.996590 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 12 22:02:21.027724 ignition[818]: Ignition 2.22.0 Sep 12 22:02:21.027742 ignition[818]: Stage: disks Sep 12 22:02:21.027891 ignition[818]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:02:21.027901 ignition[818]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 12 22:02:21.030574 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 12 22:02:21.028628 ignition[818]: disks: disks passed Sep 12 22:02:21.031597 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 12 22:02:21.028670 ignition[818]: Ignition finished successfully Sep 12 22:02:21.032843 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 12 22:02:21.034335 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 22:02:21.035424 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 22:02:21.036994 systemd[1]: Reached target basic.target - Basic System. Sep 12 22:02:21.038959 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 12 22:02:21.064743 systemd-fsck[828]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 12 22:02:21.068825 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 12 22:02:21.071379 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 12 22:02:21.135993 kernel: EXT4-fs (vda9): mounted filesystem a7b592ec-3c41-4dc2-88a7-056c1f18b418 r/w with ordered data mode. Quota mode: none. Sep 12 22:02:21.136942 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 12 22:02:21.138009 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 12 22:02:21.139888 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 22:02:21.141369 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 12 22:02:21.142166 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 12 22:02:21.142208 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 12 22:02:21.142231 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 22:02:21.154293 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 12 22:02:21.156506 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 12 22:02:21.158794 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (836) Sep 12 22:02:21.161022 kernel: BTRFS info (device vda6): first mount of filesystem 5dadbedd-e975-4944-978a-462cb6ec6aa0 Sep 12 22:02:21.161046 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 12 22:02:21.163305 kernel: BTRFS info (device vda6): turning on async discard Sep 12 22:02:21.163344 kernel: BTRFS info (device vda6): enabling free space tree Sep 12 22:02:21.164329 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 22:02:21.189731 initrd-setup-root[860]: cut: /sysroot/etc/passwd: No such file or directory Sep 12 22:02:21.193996 initrd-setup-root[867]: cut: /sysroot/etc/group: No such file or directory Sep 12 22:02:21.197277 initrd-setup-root[874]: cut: /sysroot/etc/shadow: No such file or directory Sep 12 22:02:21.201167 initrd-setup-root[881]: cut: /sysroot/etc/gshadow: No such file or directory Sep 12 22:02:21.265351 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 12 22:02:21.267375 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 12 22:02:21.268763 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 12 22:02:21.285991 kernel: BTRFS info (device vda6): last unmount of filesystem 5dadbedd-e975-4944-978a-462cb6ec6aa0 Sep 12 22:02:21.297009 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 12 22:02:21.309994 ignition[950]: INFO : Ignition 2.22.0 Sep 12 22:02:21.309994 ignition[950]: INFO : Stage: mount Sep 12 22:02:21.309994 ignition[950]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 22:02:21.309994 ignition[950]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 12 22:02:21.312742 ignition[950]: INFO : mount: mount passed Sep 12 22:02:21.312742 ignition[950]: INFO : Ignition finished successfully Sep 12 22:02:21.315066 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 12 22:02:21.317401 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 12 22:02:21.873136 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 12 22:02:21.874607 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 22:02:21.908616 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (962) Sep 12 22:02:21.908666 kernel: BTRFS info (device vda6): first mount of filesystem 5dadbedd-e975-4944-978a-462cb6ec6aa0 Sep 12 22:02:21.908677 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 12 22:02:21.913109 kernel: BTRFS info (device vda6): turning on async discard Sep 12 22:02:21.913151 kernel: BTRFS info (device vda6): enabling free space tree Sep 12 22:02:21.914594 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 22:02:21.962185 ignition[979]: INFO : Ignition 2.22.0 Sep 12 22:02:21.962185 ignition[979]: INFO : Stage: files Sep 12 22:02:21.964594 ignition[979]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 22:02:21.964594 ignition[979]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 12 22:02:21.964594 ignition[979]: DEBUG : files: compiled without relabeling support, skipping Sep 12 22:02:21.967326 ignition[979]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 12 22:02:21.967326 ignition[979]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 12 22:02:21.970824 ignition[979]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 12 22:02:21.973476 ignition[979]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 12 22:02:21.976549 ignition[979]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 12 22:02:21.974111 unknown[979]: wrote ssh authorized keys file for user: core Sep 12 22:02:21.982282 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 12 22:02:21.982282 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Sep 12 22:02:22.044143 systemd-networkd[794]: eth0: Gained IPv6LL Sep 12 22:02:22.693681 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 12 22:02:23.867662 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 12 22:02:23.867662 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 22:02:23.870637 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 22:02:23.881320 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 22:02:23.881320 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 22:02:23.881320 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 12 22:02:23.881320 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 12 22:02:23.881320 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 12 22:02:23.881320 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Sep 12 22:02:24.338930 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 12 22:02:24.789045 ignition[979]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 12 22:02:24.789045 ignition[979]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 12 22:02:24.793609 ignition[979]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 22:02:24.795797 ignition[979]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 22:02:24.795797 ignition[979]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 12 22:02:24.798927 ignition[979]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 12 22:02:24.798927 ignition[979]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 12 22:02:24.798927 ignition[979]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 12 22:02:24.798927 ignition[979]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 12 22:02:24.798927 ignition[979]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 12 22:02:24.815049 ignition[979]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 12 22:02:24.818563 ignition[979]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 12 22:02:24.819738 ignition[979]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 12 22:02:24.819738 ignition[979]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 12 22:02:24.819738 ignition[979]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 12 22:02:24.819738 ignition[979]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 12 22:02:24.819738 ignition[979]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 12 22:02:24.819738 ignition[979]: INFO : files: files passed Sep 12 22:02:24.819738 ignition[979]: INFO : Ignition finished successfully Sep 12 22:02:24.822963 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 12 22:02:24.827115 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 12 22:02:24.829821 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 12 22:02:24.838509 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 12 22:02:24.841256 initrd-setup-root-after-ignition[1008]: grep: /sysroot/oem/oem-release: No such file or directory Sep 12 22:02:24.839486 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 12 22:02:24.843517 initrd-setup-root-after-ignition[1010]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 22:02:24.845032 initrd-setup-root-after-ignition[1010]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 12 22:02:24.846152 initrd-setup-root-after-ignition[1014]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 22:02:24.845523 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 22:02:24.847778 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 12 22:02:24.853482 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 12 22:02:24.894555 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 12 22:02:24.894666 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 12 22:02:24.896427 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 12 22:02:24.897751 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 12 22:02:24.899175 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 12 22:02:24.899947 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 12 22:02:24.931362 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 22:02:24.933511 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 12 22:02:24.952128 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 12 22:02:24.953164 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 22:02:24.954880 systemd[1]: Stopped target timers.target - Timer Units. Sep 12 22:02:24.956372 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 12 22:02:24.956501 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 22:02:24.958530 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 12 22:02:24.960033 systemd[1]: Stopped target basic.target - Basic System. Sep 12 22:02:24.961338 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 12 22:02:24.962656 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 22:02:24.964103 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 12 22:02:24.965580 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 12 22:02:24.967002 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 12 22:02:24.968512 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 22:02:24.969957 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 12 22:02:24.971508 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 12 22:02:24.972760 systemd[1]: Stopped target swap.target - Swaps. Sep 12 22:02:24.973914 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 12 22:02:24.974047 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 12 22:02:24.975788 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 12 22:02:24.977333 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 22:02:24.978833 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 12 22:02:24.979598 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 22:02:24.980623 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 12 22:02:24.980735 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 12 22:02:24.982892 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 12 22:02:24.983021 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 22:02:24.984694 systemd[1]: Stopped target paths.target - Path Units. Sep 12 22:02:24.985861 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 12 22:02:24.989056 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 22:02:24.990041 systemd[1]: Stopped target slices.target - Slice Units. Sep 12 22:02:24.991704 systemd[1]: Stopped target sockets.target - Socket Units. Sep 12 22:02:24.992893 systemd[1]: iscsid.socket: Deactivated successfully. Sep 12 22:02:24.992993 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 22:02:24.994176 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 12 22:02:24.994256 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 22:02:24.995423 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 12 22:02:24.995540 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 22:02:24.996943 systemd[1]: ignition-files.service: Deactivated successfully. Sep 12 22:02:24.997060 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 12 22:02:24.999007 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 12 22:02:25.000913 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 12 22:02:25.001651 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 12 22:02:25.001779 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 22:02:25.003172 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 12 22:02:25.003271 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 22:02:25.008069 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 12 22:02:25.015095 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 12 22:02:25.023754 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 12 22:02:25.029748 ignition[1035]: INFO : Ignition 2.22.0 Sep 12 22:02:25.029748 ignition[1035]: INFO : Stage: umount Sep 12 22:02:25.032177 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 22:02:25.032177 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 12 22:02:25.032177 ignition[1035]: INFO : umount: umount passed Sep 12 22:02:25.032177 ignition[1035]: INFO : Ignition finished successfully Sep 12 22:02:25.033256 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 12 22:02:25.033381 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 12 22:02:25.035823 systemd[1]: Stopped target network.target - Network. Sep 12 22:02:25.036781 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 12 22:02:25.036852 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 12 22:02:25.038160 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 12 22:02:25.038201 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 12 22:02:25.040280 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 12 22:02:25.040327 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 12 22:02:25.041753 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 12 22:02:25.041805 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 12 22:02:25.043171 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 12 22:02:25.044596 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 12 22:02:25.051577 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 12 22:02:25.051748 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 12 22:02:25.055325 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 12 22:02:25.055532 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 12 22:02:25.056999 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 12 22:02:25.059560 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 12 22:02:25.060145 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 12 22:02:25.061872 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 12 22:02:25.061912 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 12 22:02:25.064368 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 12 22:02:25.065753 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 12 22:02:25.065817 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 22:02:25.067444 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 12 22:02:25.067482 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 12 22:02:25.069745 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 12 22:02:25.069787 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 12 22:02:25.071258 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 12 22:02:25.071299 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 22:02:25.073652 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 22:02:25.077363 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 12 22:02:25.077425 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 12 22:02:25.088598 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 12 22:02:25.095237 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 12 22:02:25.096998 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 12 22:02:25.097841 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 22:02:25.100101 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 12 22:02:25.100181 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 12 22:02:25.101613 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 12 22:02:25.101677 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 12 22:02:25.103074 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 12 22:02:25.103103 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 22:02:25.104384 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 12 22:02:25.104426 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 12 22:02:25.106568 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 12 22:02:25.106613 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 12 22:02:25.108695 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 12 22:02:25.108743 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 22:02:25.110949 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 12 22:02:25.111008 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 12 22:02:25.113167 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 12 22:02:25.114624 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 12 22:02:25.114676 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 22:02:25.117136 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 12 22:02:25.117176 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 22:02:25.119681 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 12 22:02:25.119722 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 22:02:25.122811 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 12 22:02:25.122854 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 22:02:25.124755 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 22:02:25.124805 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:02:25.128334 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Sep 12 22:02:25.128382 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Sep 12 22:02:25.128411 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 12 22:02:25.128442 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 22:02:25.128753 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 12 22:02:25.128863 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 12 22:02:25.130481 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 12 22:02:25.132034 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 12 22:02:25.148648 systemd[1]: Switching root. Sep 12 22:02:25.175183 systemd-journald[244]: Journal stopped Sep 12 22:02:25.900386 systemd-journald[244]: Received SIGTERM from PID 1 (systemd). Sep 12 22:02:25.900435 kernel: SELinux: policy capability network_peer_controls=1 Sep 12 22:02:25.900451 kernel: SELinux: policy capability open_perms=1 Sep 12 22:02:25.900464 kernel: SELinux: policy capability extended_socket_class=1 Sep 12 22:02:25.900473 kernel: SELinux: policy capability always_check_network=0 Sep 12 22:02:25.900482 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 12 22:02:25.900491 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 12 22:02:25.900500 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 12 22:02:25.900511 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 12 22:02:25.900520 kernel: SELinux: policy capability userspace_initial_context=0 Sep 12 22:02:25.900530 systemd[1]: Successfully loaded SELinux policy in 63.604ms. Sep 12 22:02:25.900550 kernel: audit: type=1403 audit(1757714545.354:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 12 22:02:25.900564 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 5.203ms. Sep 12 22:02:25.900575 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 22:02:25.900586 systemd[1]: Detected virtualization kvm. Sep 12 22:02:25.900596 systemd[1]: Detected architecture arm64. Sep 12 22:02:25.900607 systemd[1]: Detected first boot. Sep 12 22:02:25.900618 systemd[1]: Initializing machine ID from VM UUID. Sep 12 22:02:25.900628 zram_generator::config[1080]: No configuration found. Sep 12 22:02:25.900642 kernel: NET: Registered PF_VSOCK protocol family Sep 12 22:02:25.900652 systemd[1]: Populated /etc with preset unit settings. Sep 12 22:02:25.900664 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 12 22:02:25.900674 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 12 22:02:25.900685 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 12 22:02:25.900695 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 12 22:02:25.900709 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 12 22:02:25.900720 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 12 22:02:25.900730 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 12 22:02:25.900740 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 12 22:02:25.900749 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 12 22:02:25.900759 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 12 22:02:25.900769 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 12 22:02:25.900779 systemd[1]: Created slice user.slice - User and Session Slice. Sep 12 22:02:25.900799 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 22:02:25.900809 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 22:02:25.900819 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 12 22:02:25.900829 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 12 22:02:25.900840 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 12 22:02:25.900850 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 22:02:25.900860 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Sep 12 22:02:25.900870 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 22:02:25.900882 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 22:02:25.900892 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 12 22:02:25.900902 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 12 22:02:25.900912 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 12 22:02:25.900922 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 12 22:02:25.900932 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 22:02:25.900942 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 22:02:25.900952 systemd[1]: Reached target slices.target - Slice Units. Sep 12 22:02:25.900962 systemd[1]: Reached target swap.target - Swaps. Sep 12 22:02:25.900981 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 12 22:02:25.900992 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 12 22:02:25.901002 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 12 22:02:25.901012 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 22:02:25.901022 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 22:02:25.901032 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 22:02:25.901042 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 12 22:02:25.901051 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 12 22:02:25.901061 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 12 22:02:25.901073 systemd[1]: Mounting media.mount - External Media Directory... Sep 12 22:02:25.901083 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 12 22:02:25.901093 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 12 22:02:25.901104 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 12 22:02:25.901114 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 12 22:02:25.901124 systemd[1]: Reached target machines.target - Containers. Sep 12 22:02:25.901134 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 12 22:02:25.901143 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:02:25.901154 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 22:02:25.901165 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 12 22:02:25.901175 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:02:25.901185 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 22:02:25.901195 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:02:25.901205 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 12 22:02:25.901215 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 22:02:25.901225 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 12 22:02:25.901235 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 12 22:02:25.901247 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 12 22:02:25.901257 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 12 22:02:25.901267 systemd[1]: Stopped systemd-fsck-usr.service. Sep 12 22:02:25.901277 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:02:25.901288 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 22:02:25.901298 kernel: loop: module loaded Sep 12 22:02:25.901308 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 22:02:25.901317 kernel: ACPI: bus type drm_connector registered Sep 12 22:02:25.901327 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 22:02:25.901338 kernel: fuse: init (API version 7.41) Sep 12 22:02:25.901348 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 12 22:02:25.901358 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 12 22:02:25.901368 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 22:02:25.901378 systemd[1]: verity-setup.service: Deactivated successfully. Sep 12 22:02:25.901387 systemd[1]: Stopped verity-setup.service. Sep 12 22:02:25.901398 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 12 22:02:25.901409 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 12 22:02:25.901419 systemd[1]: Mounted media.mount - External Media Directory. Sep 12 22:02:25.901430 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 12 22:02:25.901460 systemd-journald[1152]: Collecting audit messages is disabled. Sep 12 22:02:25.901484 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 12 22:02:25.901495 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 12 22:02:25.901505 systemd-journald[1152]: Journal started Sep 12 22:02:25.901526 systemd-journald[1152]: Runtime Journal (/run/log/journal/e1aa75063f484c41a890bad49dc22132) is 6M, max 48.5M, 42.4M free. Sep 12 22:02:25.704142 systemd[1]: Queued start job for default target multi-user.target. Sep 12 22:02:25.722865 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 12 22:02:25.723246 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 12 22:02:25.903752 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 22:02:25.905502 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 12 22:02:25.906851 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 22:02:25.908262 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 12 22:02:25.908441 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 12 22:02:25.909708 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:02:25.909865 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:02:25.912272 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 22:02:25.912439 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 22:02:25.913603 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:02:25.913758 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:02:25.915114 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 12 22:02:25.915271 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 12 22:02:25.916566 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 22:02:25.916753 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 22:02:25.918072 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 22:02:25.919291 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 22:02:25.920646 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 12 22:02:25.922198 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 12 22:02:25.933730 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 22:02:25.935892 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 12 22:02:25.937936 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 12 22:02:25.939027 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 12 22:02:25.939056 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 22:02:25.940678 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 12 22:02:25.949713 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 12 22:02:25.950781 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:02:25.951782 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 12 22:02:25.953575 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 12 22:02:25.954743 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 22:02:25.957310 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 12 22:02:25.958324 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 22:02:25.959173 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 22:02:25.961203 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 12 22:02:25.964198 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 12 22:02:25.967290 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 22:02:25.967489 systemd-journald[1152]: Time spent on flushing to /var/log/journal/e1aa75063f484c41a890bad49dc22132 is 21.391ms for 892 entries. Sep 12 22:02:25.967489 systemd-journald[1152]: System Journal (/var/log/journal/e1aa75063f484c41a890bad49dc22132) is 8M, max 195.6M, 187.6M free. Sep 12 22:02:26.002384 systemd-journald[1152]: Received client request to flush runtime journal. Sep 12 22:02:26.002428 kernel: loop0: detected capacity change from 0 to 203944 Sep 12 22:02:25.971137 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 12 22:02:25.972293 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 12 22:02:25.983432 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 12 22:02:25.985360 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 12 22:02:25.988307 systemd-tmpfiles[1199]: ACLs are not supported, ignoring. Sep 12 22:02:25.988317 systemd-tmpfiles[1199]: ACLs are not supported, ignoring. Sep 12 22:02:25.988750 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 12 22:02:25.994014 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 22:02:26.004164 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 22:02:26.005750 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 12 22:02:26.011996 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 12 22:02:26.013241 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 12 22:02:26.026202 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 12 22:02:26.042004 kernel: loop1: detected capacity change from 0 to 119368 Sep 12 22:02:26.045665 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 12 22:02:26.049129 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 22:02:26.065001 kernel: loop2: detected capacity change from 0 to 100632 Sep 12 22:02:26.069643 systemd-tmpfiles[1219]: ACLs are not supported, ignoring. Sep 12 22:02:26.069682 systemd-tmpfiles[1219]: ACLs are not supported, ignoring. Sep 12 22:02:26.074038 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 22:02:26.100161 kernel: loop3: detected capacity change from 0 to 203944 Sep 12 22:02:26.106996 kernel: loop4: detected capacity change from 0 to 119368 Sep 12 22:02:26.112994 kernel: loop5: detected capacity change from 0 to 100632 Sep 12 22:02:26.117860 (sd-merge)[1224]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 12 22:02:26.118264 (sd-merge)[1224]: Merged extensions into '/usr'. Sep 12 22:02:26.122572 systemd[1]: Reload requested from client PID 1198 ('systemd-sysext') (unit systemd-sysext.service)... Sep 12 22:02:26.122700 systemd[1]: Reloading... Sep 12 22:02:26.192005 zram_generator::config[1250]: No configuration found. Sep 12 22:02:26.235884 ldconfig[1193]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 12 22:02:26.332484 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 12 22:02:26.333019 systemd[1]: Reloading finished in 209 ms. Sep 12 22:02:26.352434 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 12 22:02:26.353703 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 12 22:02:26.367145 systemd[1]: Starting ensure-sysext.service... Sep 12 22:02:26.368813 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 22:02:26.377236 systemd[1]: Reload requested from client PID 1284 ('systemctl') (unit ensure-sysext.service)... Sep 12 22:02:26.377249 systemd[1]: Reloading... Sep 12 22:02:26.387337 systemd-tmpfiles[1285]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 12 22:02:26.387366 systemd-tmpfiles[1285]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 12 22:02:26.387596 systemd-tmpfiles[1285]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 12 22:02:26.387780 systemd-tmpfiles[1285]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 12 22:02:26.388423 systemd-tmpfiles[1285]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 12 22:02:26.388621 systemd-tmpfiles[1285]: ACLs are not supported, ignoring. Sep 12 22:02:26.388662 systemd-tmpfiles[1285]: ACLs are not supported, ignoring. Sep 12 22:02:26.391817 systemd-tmpfiles[1285]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 22:02:26.391828 systemd-tmpfiles[1285]: Skipping /boot Sep 12 22:02:26.397688 systemd-tmpfiles[1285]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 22:02:26.397802 systemd-tmpfiles[1285]: Skipping /boot Sep 12 22:02:26.422997 zram_generator::config[1312]: No configuration found. Sep 12 22:02:26.551524 systemd[1]: Reloading finished in 173 ms. Sep 12 22:02:26.563596 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 12 22:02:26.569369 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 22:02:26.576096 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 22:02:26.578175 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 12 22:02:26.588859 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 12 22:02:26.591798 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 22:02:26.595186 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 22:02:26.597651 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 12 22:02:26.609351 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 12 22:02:26.614052 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:02:26.615397 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:02:26.618284 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:02:26.621323 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 22:02:26.622381 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:02:26.622498 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:02:26.625012 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 12 22:02:26.626865 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:02:26.627426 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:02:26.629182 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:02:26.629334 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:02:26.630908 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 22:02:26.631208 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 22:02:26.636582 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 12 22:02:26.641454 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:02:26.643214 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:02:26.643616 augenrules[1382]: No rules Sep 12 22:02:26.644877 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:02:26.647794 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 22:02:26.648694 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:02:26.648820 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:02:26.651723 systemd-udevd[1353]: Using default interface naming scheme 'v255'. Sep 12 22:02:26.661120 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 12 22:02:26.662385 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 22:02:26.663664 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 12 22:02:26.666202 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 22:02:26.666422 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 22:02:26.667766 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 12 22:02:26.669411 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:02:26.669586 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:02:26.670959 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:02:26.671138 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:02:26.672674 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 22:02:26.672870 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 22:02:26.676141 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 12 22:02:26.683257 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 22:02:26.688253 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 22:02:26.690142 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:02:26.691484 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:02:26.694967 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 22:02:26.696666 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:02:26.701205 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 22:02:26.702119 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:02:26.702168 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:02:26.721672 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 22:02:26.722649 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 22:02:26.723453 systemd[1]: Finished ensure-sysext.service. Sep 12 22:02:26.734771 systemd-resolved[1352]: Positive Trust Anchors: Sep 12 22:02:26.734798 systemd-resolved[1352]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 22:02:26.734830 systemd-resolved[1352]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 22:02:26.737133 augenrules[1417]: /sbin/augenrules: No change Sep 12 22:02:26.743486 systemd-resolved[1352]: Defaulting to hostname 'linux'. Sep 12 22:02:26.746647 augenrules[1452]: No rules Sep 12 22:02:26.756177 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 22:02:26.757893 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 22:02:26.758148 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 22:02:26.759375 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:02:26.759537 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:02:26.760907 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 22:02:26.761144 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 22:02:26.762371 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:02:26.762528 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:02:26.763845 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 22:02:26.764027 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 22:02:26.775715 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Sep 12 22:02:26.777753 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 22:02:26.780137 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 22:02:26.780218 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 22:02:26.782702 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 12 22:02:26.831959 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 12 22:02:26.832191 systemd-networkd[1435]: lo: Link UP Sep 12 22:02:26.832208 systemd-networkd[1435]: lo: Gained carrier Sep 12 22:02:26.833190 systemd-networkd[1435]: Enumeration completed Sep 12 22:02:26.833595 systemd-networkd[1435]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:02:26.833604 systemd-networkd[1435]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 12 22:02:26.833760 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 22:02:26.834224 systemd-networkd[1435]: eth0: Link UP Sep 12 22:02:26.834323 systemd-networkd[1435]: eth0: Gained carrier Sep 12 22:02:26.834336 systemd-networkd[1435]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:02:26.834781 systemd[1]: Reached target network.target - Network. Sep 12 22:02:26.837392 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 12 22:02:26.839664 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 12 22:02:26.842089 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 12 22:02:26.849346 systemd-networkd[1435]: eth0: DHCPv4 address 10.0.0.29/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 12 22:02:26.852660 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 12 22:02:26.853933 systemd-timesyncd[1470]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 12 22:02:26.854002 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 22:02:26.854048 systemd-timesyncd[1470]: Initial clock synchronization to Fri 2025-09-12 22:02:26.484733 UTC. Sep 12 22:02:26.854955 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 12 22:02:26.856037 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 12 22:02:26.857080 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 12 22:02:26.858064 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 12 22:02:26.858092 systemd[1]: Reached target paths.target - Path Units. Sep 12 22:02:26.858845 systemd[1]: Reached target time-set.target - System Time Set. Sep 12 22:02:26.859800 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 12 22:02:26.860734 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 12 22:02:26.861781 systemd[1]: Reached target timers.target - Timer Units. Sep 12 22:02:26.863272 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 12 22:02:26.865646 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 12 22:02:26.868473 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 12 22:02:26.869656 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 12 22:02:26.870727 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 12 22:02:26.874029 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 12 22:02:26.875252 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 12 22:02:26.878026 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 12 22:02:26.879532 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 12 22:02:26.880871 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 12 22:02:26.883863 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 22:02:26.884805 systemd[1]: Reached target basic.target - Basic System. Sep 12 22:02:26.885641 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 12 22:02:26.885669 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 12 22:02:26.887258 systemd[1]: Starting containerd.service - containerd container runtime... Sep 12 22:02:26.889111 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 12 22:02:26.893006 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 12 22:02:26.902152 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 12 22:02:26.904504 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 12 22:02:26.907109 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 12 22:02:26.911157 jq[1497]: false Sep 12 22:02:26.911603 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 12 22:02:26.914485 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 12 22:02:26.918118 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 12 22:02:26.919907 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 12 22:02:26.924267 extend-filesystems[1498]: Found /dev/vda6 Sep 12 22:02:26.925261 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 12 22:02:26.926812 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 12 22:02:26.927268 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 12 22:02:26.928132 systemd[1]: Starting update-engine.service - Update Engine... Sep 12 22:02:26.930209 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 12 22:02:26.933040 extend-filesystems[1498]: Found /dev/vda9 Sep 12 22:02:26.935009 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 12 22:02:26.936255 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 12 22:02:26.936428 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 12 22:02:26.936659 systemd[1]: motdgen.service: Deactivated successfully. Sep 12 22:02:26.936834 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 12 22:02:26.941335 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 12 22:02:26.943006 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 12 22:02:26.949262 jq[1516]: true Sep 12 22:02:26.949659 extend-filesystems[1498]: Checking size of /dev/vda9 Sep 12 22:02:26.960609 update_engine[1515]: I20250912 22:02:26.960388 1515 main.cc:92] Flatcar Update Engine starting Sep 12 22:02:26.967332 jq[1531]: true Sep 12 22:02:26.968637 extend-filesystems[1498]: Resized partition /dev/vda9 Sep 12 22:02:26.971359 (ntainerd)[1535]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 12 22:02:26.973632 tar[1522]: linux-arm64/helm Sep 12 22:02:26.976169 extend-filesystems[1541]: resize2fs 1.47.3 (8-Jul-2025) Sep 12 22:02:26.976379 dbus-daemon[1493]: [system] SELinux support is enabled Sep 12 22:02:26.976517 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 12 22:02:26.979898 update_engine[1515]: I20250912 22:02:26.979391 1515 update_check_scheduler.cc:74] Next update check in 6m55s Sep 12 22:02:26.982400 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 12 22:02:26.982529 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 12 22:02:26.985982 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 12 22:02:26.984133 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 12 22:02:26.984237 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 12 22:02:26.986398 systemd[1]: Started update-engine.service - Update Engine. Sep 12 22:02:26.989678 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:02:26.993109 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 12 22:02:27.015991 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 12 22:02:27.026211 extend-filesystems[1541]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 12 22:02:27.026211 extend-filesystems[1541]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 12 22:02:27.026211 extend-filesystems[1541]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 12 22:02:27.031102 extend-filesystems[1498]: Resized filesystem in /dev/vda9 Sep 12 22:02:27.036097 bash[1557]: Updated "/home/core/.ssh/authorized_keys" Sep 12 22:02:27.028703 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 12 22:02:27.028916 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 12 22:02:27.034557 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 12 22:02:27.043652 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 12 22:02:27.104064 locksmithd[1547]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 12 22:02:27.115764 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:02:27.126865 systemd-logind[1513]: Watching system buttons on /dev/input/event0 (Power Button) Sep 12 22:02:27.128134 systemd-logind[1513]: New seat seat0. Sep 12 22:02:27.130722 systemd[1]: Started systemd-logind.service - User Login Management. Sep 12 22:02:27.150172 containerd[1535]: time="2025-09-12T22:02:27Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 12 22:02:27.151280 containerd[1535]: time="2025-09-12T22:02:27.151247939Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 12 22:02:27.160108 containerd[1535]: time="2025-09-12T22:02:27.160064463Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="9.806µs" Sep 12 22:02:27.160169 containerd[1535]: time="2025-09-12T22:02:27.160107730Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 12 22:02:27.160169 containerd[1535]: time="2025-09-12T22:02:27.160126464Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 12 22:02:27.160284 containerd[1535]: time="2025-09-12T22:02:27.160264544Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 12 22:02:27.160307 containerd[1535]: time="2025-09-12T22:02:27.160284003Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 12 22:02:27.160327 containerd[1535]: time="2025-09-12T22:02:27.160305980Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160382 containerd[1535]: time="2025-09-12T22:02:27.160362906Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160382 containerd[1535]: time="2025-09-12T22:02:27.160378512Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160623 containerd[1535]: time="2025-09-12T22:02:27.160601333Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160647 containerd[1535]: time="2025-09-12T22:02:27.160622204Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160647 containerd[1535]: time="2025-09-12T22:02:27.160633497Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160647 containerd[1535]: time="2025-09-12T22:02:27.160641548Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160723 containerd[1535]: time="2025-09-12T22:02:27.160708394Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160908 containerd[1535]: time="2025-09-12T22:02:27.160888979Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160942 containerd[1535]: time="2025-09-12T22:02:27.160922593Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 12 22:02:27.160976 containerd[1535]: time="2025-09-12T22:02:27.160942357Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 12 22:02:27.160995 containerd[1535]: time="2025-09-12T22:02:27.160983755Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 12 22:02:27.162269 containerd[1535]: time="2025-09-12T22:02:27.162242277Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 12 22:02:27.162360 containerd[1535]: time="2025-09-12T22:02:27.162337090Z" level=info msg="metadata content store policy set" policy=shared Sep 12 22:02:27.165828 containerd[1535]: time="2025-09-12T22:02:27.165789146Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 12 22:02:27.165888 containerd[1535]: time="2025-09-12T22:02:27.165873506Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 12 22:02:27.165930 containerd[1535]: time="2025-09-12T22:02:27.165891286Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 12 22:02:27.165930 containerd[1535]: time="2025-09-12T22:02:27.165903915Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 12 22:02:27.165930 containerd[1535]: time="2025-09-12T22:02:27.165916391Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 12 22:02:27.165986 containerd[1535]: time="2025-09-12T22:02:27.165931500Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 12 22:02:27.165986 containerd[1535]: time="2025-09-12T22:02:27.165944053Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 12 22:02:27.165986 containerd[1535]: time="2025-09-12T22:02:27.165955270Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 12 22:02:27.165986 containerd[1535]: time="2025-09-12T22:02:27.165966183Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 12 22:02:27.166051 containerd[1535]: time="2025-09-12T22:02:27.165986061Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 12 22:02:27.166051 containerd[1535]: time="2025-09-12T22:02:27.165996286Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 12 22:02:27.166051 containerd[1535]: time="2025-09-12T22:02:27.166009640Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166149743Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166177405Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166199230Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166212774Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166222847Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166232615Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166244214Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166254477Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166269930Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166279926Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166289579Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166463334Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166477108Z" level=info msg="Start snapshots syncer" Sep 12 22:02:27.167976 containerd[1535]: time="2025-09-12T22:02:27.166503129Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 12 22:02:27.168215 containerd[1535]: time="2025-09-12T22:02:27.166705538Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 12 22:02:27.168215 containerd[1535]: time="2025-09-12T22:02:27.166752430Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.166836141Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.166966094Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167010315Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167021762Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167031567Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167053010Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167063159Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167073881Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167097422Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167109898Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167120696Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167169496Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167182125Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 12 22:02:27.168310 containerd[1535]: time="2025-09-12T22:02:27.167190938Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167200401Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167207879Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167216616Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167227338Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167333292Z" level=info msg="runtime interface created" Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167339511Z" level=info msg="created NRI interface" Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167351339Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167362595Z" level=info msg="Connect containerd service" Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.167389837Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 12 22:02:27.168525 containerd[1535]: time="2025-09-12T22:02:27.168219733Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 12 22:02:27.230803 containerd[1535]: time="2025-09-12T22:02:27.230738511Z" level=info msg="Start subscribing containerd event" Sep 12 22:02:27.230904 containerd[1535]: time="2025-09-12T22:02:27.230819665Z" level=info msg="Start recovering state" Sep 12 22:02:27.230923 containerd[1535]: time="2025-09-12T22:02:27.230901811Z" level=info msg="Start event monitor" Sep 12 22:02:27.230923 containerd[1535]: time="2025-09-12T22:02:27.230915203Z" level=info msg="Start cni network conf syncer for default" Sep 12 22:02:27.230923 containerd[1535]: time="2025-09-12T22:02:27.230923063Z" level=info msg="Start streaming server" Sep 12 22:02:27.231003 containerd[1535]: time="2025-09-12T22:02:27.230934319Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 12 22:02:27.231003 containerd[1535]: time="2025-09-12T22:02:27.230940729Z" level=info msg="runtime interface starting up..." Sep 12 22:02:27.231003 containerd[1535]: time="2025-09-12T22:02:27.230945689Z" level=info msg="starting plugins..." Sep 12 22:02:27.231003 containerd[1535]: time="2025-09-12T22:02:27.230965338Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 12 22:02:27.231188 containerd[1535]: time="2025-09-12T22:02:27.231163703Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 12 22:02:27.231368 containerd[1535]: time="2025-09-12T22:02:27.231347226Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 12 22:02:27.231433 containerd[1535]: time="2025-09-12T22:02:27.231419948Z" level=info msg="containerd successfully booted in 0.081590s" Sep 12 22:02:27.231599 systemd[1]: Started containerd.service - containerd container runtime. Sep 12 22:02:27.265508 tar[1522]: linux-arm64/LICENSE Sep 12 22:02:27.265588 tar[1522]: linux-arm64/README.md Sep 12 22:02:27.281940 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 12 22:02:27.410899 sshd_keygen[1521]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 12 22:02:27.431054 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 12 22:02:27.433550 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 12 22:02:27.451387 systemd[1]: issuegen.service: Deactivated successfully. Sep 12 22:02:27.451625 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 12 22:02:27.454023 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 12 22:02:27.477857 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 12 22:02:27.480374 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 12 22:02:27.482348 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Sep 12 22:02:27.483351 systemd[1]: Reached target getty.target - Login Prompts. Sep 12 22:02:28.444101 systemd-networkd[1435]: eth0: Gained IPv6LL Sep 12 22:02:28.448021 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 12 22:02:28.449347 systemd[1]: Reached target network-online.target - Network is Online. Sep 12 22:02:28.451334 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 12 22:02:28.453327 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:02:28.455142 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 12 22:02:28.477698 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 12 22:02:28.477934 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 12 22:02:28.479681 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 12 22:02:28.483007 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 12 22:02:28.981339 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:28.982515 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 12 22:02:28.984407 (kubelet)[1636]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 22:02:28.985792 systemd[1]: Startup finished in 1.976s (kernel) + 6.742s (initrd) + 3.694s (userspace) = 12.413s. Sep 12 22:02:29.330481 kubelet[1636]: E0912 22:02:29.330356 1636 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 22:02:29.332734 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 22:02:29.332863 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 22:02:29.334097 systemd[1]: kubelet.service: Consumed 768ms CPU time, 256.3M memory peak. Sep 12 22:02:31.668036 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 12 22:02:31.671523 systemd[1]: Started sshd@0-10.0.0.29:22-10.0.0.1:39460.service - OpenSSH per-connection server daemon (10.0.0.1:39460). Sep 12 22:02:31.755739 sshd[1649]: Accepted publickey for core from 10.0.0.1 port 39460 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:31.757522 sshd-session[1649]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:31.763920 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 12 22:02:31.764883 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 12 22:02:31.775954 systemd-logind[1513]: New session 1 of user core. Sep 12 22:02:31.792999 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 12 22:02:31.795755 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 12 22:02:31.816119 (systemd)[1655]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 12 22:02:31.818853 systemd-logind[1513]: New session c1 of user core. Sep 12 22:02:31.934392 systemd[1655]: Queued start job for default target default.target. Sep 12 22:02:31.952073 systemd[1655]: Created slice app.slice - User Application Slice. Sep 12 22:02:31.952102 systemd[1655]: Reached target paths.target - Paths. Sep 12 22:02:31.952141 systemd[1655]: Reached target timers.target - Timers. Sep 12 22:02:31.953385 systemd[1655]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 12 22:02:31.966047 systemd[1655]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 12 22:02:31.966169 systemd[1655]: Reached target sockets.target - Sockets. Sep 12 22:02:31.966213 systemd[1655]: Reached target basic.target - Basic System. Sep 12 22:02:31.966240 systemd[1655]: Reached target default.target - Main User Target. Sep 12 22:02:31.966265 systemd[1655]: Startup finished in 141ms. Sep 12 22:02:31.966530 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 12 22:02:31.977198 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 12 22:02:32.040794 systemd[1]: Started sshd@1-10.0.0.29:22-10.0.0.1:39468.service - OpenSSH per-connection server daemon (10.0.0.1:39468). Sep 12 22:02:32.107158 sshd[1666]: Accepted publickey for core from 10.0.0.1 port 39468 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:32.108511 sshd-session[1666]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:32.112166 systemd-logind[1513]: New session 2 of user core. Sep 12 22:02:32.124137 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 12 22:02:32.178308 sshd[1669]: Connection closed by 10.0.0.1 port 39468 Sep 12 22:02:32.178643 sshd-session[1666]: pam_unix(sshd:session): session closed for user core Sep 12 22:02:32.181151 systemd[1]: sshd@1-10.0.0.29:22-10.0.0.1:39468.service: Deactivated successfully. Sep 12 22:02:32.182492 systemd[1]: session-2.scope: Deactivated successfully. Sep 12 22:02:32.193227 systemd-logind[1513]: Session 2 logged out. Waiting for processes to exit. Sep 12 22:02:32.194343 systemd[1]: Started sshd@2-10.0.0.29:22-10.0.0.1:39470.service - OpenSSH per-connection server daemon (10.0.0.1:39470). Sep 12 22:02:32.195407 systemd-logind[1513]: Removed session 2. Sep 12 22:02:32.237166 sshd[1675]: Accepted publickey for core from 10.0.0.1 port 39470 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:32.240815 sshd-session[1675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:32.245358 systemd-logind[1513]: New session 3 of user core. Sep 12 22:02:32.254153 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 12 22:02:32.302458 sshd[1678]: Connection closed by 10.0.0.1 port 39470 Sep 12 22:02:32.302934 sshd-session[1675]: pam_unix(sshd:session): session closed for user core Sep 12 22:02:32.319839 systemd[1]: sshd@2-10.0.0.29:22-10.0.0.1:39470.service: Deactivated successfully. Sep 12 22:02:32.323348 systemd[1]: session-3.scope: Deactivated successfully. Sep 12 22:02:32.324762 systemd-logind[1513]: Session 3 logged out. Waiting for processes to exit. Sep 12 22:02:32.326068 systemd[1]: Started sshd@3-10.0.0.29:22-10.0.0.1:39478.service - OpenSSH per-connection server daemon (10.0.0.1:39478). Sep 12 22:02:32.329578 systemd-logind[1513]: Removed session 3. Sep 12 22:02:32.386662 sshd[1684]: Accepted publickey for core from 10.0.0.1 port 39478 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:32.388619 sshd-session[1684]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:32.394753 systemd-logind[1513]: New session 4 of user core. Sep 12 22:02:32.402561 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 12 22:02:32.457455 sshd[1687]: Connection closed by 10.0.0.1 port 39478 Sep 12 22:02:32.458259 sshd-session[1684]: pam_unix(sshd:session): session closed for user core Sep 12 22:02:32.472969 systemd[1]: sshd@3-10.0.0.29:22-10.0.0.1:39478.service: Deactivated successfully. Sep 12 22:02:32.475571 systemd[1]: session-4.scope: Deactivated successfully. Sep 12 22:02:32.476746 systemd-logind[1513]: Session 4 logged out. Waiting for processes to exit. Sep 12 22:02:32.480638 systemd[1]: Started sshd@4-10.0.0.29:22-10.0.0.1:39480.service - OpenSSH per-connection server daemon (10.0.0.1:39480). Sep 12 22:02:32.481798 systemd-logind[1513]: Removed session 4. Sep 12 22:02:32.534409 sshd[1693]: Accepted publickey for core from 10.0.0.1 port 39480 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:32.535719 sshd-session[1693]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:32.541048 systemd-logind[1513]: New session 5 of user core. Sep 12 22:02:32.553147 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 12 22:02:32.611265 sudo[1697]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 12 22:02:32.611513 sudo[1697]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:02:32.628781 sudo[1697]: pam_unix(sudo:session): session closed for user root Sep 12 22:02:32.630776 sshd[1696]: Connection closed by 10.0.0.1 port 39480 Sep 12 22:02:32.630588 sshd-session[1693]: pam_unix(sshd:session): session closed for user core Sep 12 22:02:32.643856 systemd[1]: sshd@4-10.0.0.29:22-10.0.0.1:39480.service: Deactivated successfully. Sep 12 22:02:32.647120 systemd[1]: session-5.scope: Deactivated successfully. Sep 12 22:02:32.647884 systemd-logind[1513]: Session 5 logged out. Waiting for processes to exit. Sep 12 22:02:32.650186 systemd[1]: Started sshd@5-10.0.0.29:22-10.0.0.1:39482.service - OpenSSH per-connection server daemon (10.0.0.1:39482). Sep 12 22:02:32.651874 systemd-logind[1513]: Removed session 5. Sep 12 22:02:32.712408 sshd[1703]: Accepted publickey for core from 10.0.0.1 port 39482 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:32.713261 sshd-session[1703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:32.716733 systemd-logind[1513]: New session 6 of user core. Sep 12 22:02:32.733152 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 12 22:02:32.782669 sudo[1708]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 12 22:02:32.782923 sudo[1708]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:02:32.821814 sudo[1708]: pam_unix(sudo:session): session closed for user root Sep 12 22:02:32.826606 sudo[1707]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 12 22:02:32.826873 sudo[1707]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:02:32.834903 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 22:02:32.874916 augenrules[1730]: No rules Sep 12 22:02:32.876238 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 22:02:32.877052 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 22:02:32.878435 sudo[1707]: pam_unix(sudo:session): session closed for user root Sep 12 22:02:32.880179 sshd[1706]: Connection closed by 10.0.0.1 port 39482 Sep 12 22:02:32.880109 sshd-session[1703]: pam_unix(sshd:session): session closed for user core Sep 12 22:02:32.901936 systemd[1]: sshd@5-10.0.0.29:22-10.0.0.1:39482.service: Deactivated successfully. Sep 12 22:02:32.903402 systemd[1]: session-6.scope: Deactivated successfully. Sep 12 22:02:32.904029 systemd-logind[1513]: Session 6 logged out. Waiting for processes to exit. Sep 12 22:02:32.906223 systemd[1]: Started sshd@6-10.0.0.29:22-10.0.0.1:39486.service - OpenSSH per-connection server daemon (10.0.0.1:39486). Sep 12 22:02:32.907845 systemd-logind[1513]: Removed session 6. Sep 12 22:02:32.960210 sshd[1739]: Accepted publickey for core from 10.0.0.1 port 39486 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:02:32.961301 sshd-session[1739]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:02:32.964856 systemd-logind[1513]: New session 7 of user core. Sep 12 22:02:32.971116 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 12 22:02:33.019911 sudo[1744]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 12 22:02:33.020532 sudo[1744]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:02:33.283884 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 12 22:02:33.301296 (dockerd)[1764]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 12 22:02:33.491752 dockerd[1764]: time="2025-09-12T22:02:33.491694140Z" level=info msg="Starting up" Sep 12 22:02:33.492544 dockerd[1764]: time="2025-09-12T22:02:33.492523718Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 12 22:02:33.502283 dockerd[1764]: time="2025-09-12T22:02:33.502253106Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 12 22:02:33.634032 dockerd[1764]: time="2025-09-12T22:02:33.633729336Z" level=info msg="Loading containers: start." Sep 12 22:02:33.641986 kernel: Initializing XFRM netlink socket Sep 12 22:02:33.838555 systemd-networkd[1435]: docker0: Link UP Sep 12 22:02:33.843359 dockerd[1764]: time="2025-09-12T22:02:33.843311732Z" level=info msg="Loading containers: done." Sep 12 22:02:33.857922 dockerd[1764]: time="2025-09-12T22:02:33.857864173Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 12 22:02:33.858088 dockerd[1764]: time="2025-09-12T22:02:33.857958616Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 12 22:02:33.858088 dockerd[1764]: time="2025-09-12T22:02:33.858056428Z" level=info msg="Initializing buildkit" Sep 12 22:02:33.881652 dockerd[1764]: time="2025-09-12T22:02:33.881608331Z" level=info msg="Completed buildkit initialization" Sep 12 22:02:33.886543 dockerd[1764]: time="2025-09-12T22:02:33.886270742Z" level=info msg="Daemon has completed initialization" Sep 12 22:02:33.886543 dockerd[1764]: time="2025-09-12T22:02:33.886327032Z" level=info msg="API listen on /run/docker.sock" Sep 12 22:02:33.886535 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 12 22:02:34.512469 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2488416636-merged.mount: Deactivated successfully. Sep 12 22:02:34.598633 containerd[1535]: time="2025-09-12T22:02:34.598566441Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\"" Sep 12 22:02:35.147244 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1975468675.mount: Deactivated successfully. Sep 12 22:02:36.258956 containerd[1535]: time="2025-09-12T22:02:36.258909386Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:36.260232 containerd[1535]: time="2025-09-12T22:02:36.260177506Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.13: active requests=0, bytes read=25687327" Sep 12 22:02:36.261449 containerd[1535]: time="2025-09-12T22:02:36.261129790Z" level=info msg="ImageCreate event name:\"sha256:0b1c07d8fd4a3526d5c44502e682df3627a3b01c1e608e5e24c3519c8fb337b6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:36.264256 containerd[1535]: time="2025-09-12T22:02:36.264220550Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:36.265298 containerd[1535]: time="2025-09-12T22:02:36.265255195Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.13\" with image id \"sha256:0b1c07d8fd4a3526d5c44502e682df3627a3b01c1e608e5e24c3519c8fb337b6\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.13\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e\", size \"25683924\" in 1.666641122s" Sep 12 22:02:36.265394 containerd[1535]: time="2025-09-12T22:02:36.265379171Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\" returns image reference \"sha256:0b1c07d8fd4a3526d5c44502e682df3627a3b01c1e608e5e24c3519c8fb337b6\"" Sep 12 22:02:36.266727 containerd[1535]: time="2025-09-12T22:02:36.266631592Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\"" Sep 12 22:02:37.428050 containerd[1535]: time="2025-09-12T22:02:37.427998959Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:37.428545 containerd[1535]: time="2025-09-12T22:02:37.428511383Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.13: active requests=0, bytes read=22459769" Sep 12 22:02:37.429515 containerd[1535]: time="2025-09-12T22:02:37.429480554Z" level=info msg="ImageCreate event name:\"sha256:c359cb88f3d2147f2cb4c5ada4fbdeadc4b1c009d66c8f33f3856efaf04ee6ef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:37.432483 containerd[1535]: time="2025-09-12T22:02:37.432451883Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:37.433451 containerd[1535]: time="2025-09-12T22:02:37.433420106Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.13\" with image id \"sha256:c359cb88f3d2147f2cb4c5ada4fbdeadc4b1c009d66c8f33f3856efaf04ee6ef\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.13\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09\", size \"24028542\" in 1.166682394s" Sep 12 22:02:37.433490 containerd[1535]: time="2025-09-12T22:02:37.433457052Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\" returns image reference \"sha256:c359cb88f3d2147f2cb4c5ada4fbdeadc4b1c009d66c8f33f3856efaf04ee6ef\"" Sep 12 22:02:37.434347 containerd[1535]: time="2025-09-12T22:02:37.434325027Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\"" Sep 12 22:02:38.762160 containerd[1535]: time="2025-09-12T22:02:38.762106040Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:38.763304 containerd[1535]: time="2025-09-12T22:02:38.763258627Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.13: active requests=0, bytes read=17127508" Sep 12 22:02:38.764178 containerd[1535]: time="2025-09-12T22:02:38.764150571Z" level=info msg="ImageCreate event name:\"sha256:5e3cbe2ba7db787c6aebfcf4484156dd4ebd7ede811ef72e8929593e59a5fa27\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:38.767427 containerd[1535]: time="2025-09-12T22:02:38.767396090Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:38.768728 containerd[1535]: time="2025-09-12T22:02:38.768277983Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.13\" with image id \"sha256:5e3cbe2ba7db787c6aebfcf4484156dd4ebd7ede811ef72e8929593e59a5fa27\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.13\", repo digest \"registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241\", size \"18696299\" in 1.333922202s" Sep 12 22:02:38.768728 containerd[1535]: time="2025-09-12T22:02:38.768309129Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\" returns image reference \"sha256:5e3cbe2ba7db787c6aebfcf4484156dd4ebd7ede811ef72e8929593e59a5fa27\"" Sep 12 22:02:38.768912 containerd[1535]: time="2025-09-12T22:02:38.768829384Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\"" Sep 12 22:02:39.478328 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 12 22:02:39.479645 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:02:39.616544 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:39.620517 (kubelet)[2061]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 22:02:39.665904 kubelet[2061]: E0912 22:02:39.665859 2061 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 22:02:39.668638 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 22:02:39.668752 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 22:02:39.670037 systemd[1]: kubelet.service: Consumed 141ms CPU time, 108.1M memory peak. Sep 12 22:02:39.898883 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2413286554.mount: Deactivated successfully. Sep 12 22:02:40.233039 containerd[1535]: time="2025-09-12T22:02:40.232990043Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:40.233515 containerd[1535]: time="2025-09-12T22:02:40.233482565Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.13: active requests=0, bytes read=26954909" Sep 12 22:02:40.234338 containerd[1535]: time="2025-09-12T22:02:40.234299309Z" level=info msg="ImageCreate event name:\"sha256:c15699f0b7002450249485b10f20211982dfd2bec4d61c86c35acebc659e794e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:40.236113 containerd[1535]: time="2025-09-12T22:02:40.236090822Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:40.236931 containerd[1535]: time="2025-09-12T22:02:40.236584058Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.13\" with image id \"sha256:c15699f0b7002450249485b10f20211982dfd2bec4d61c86c35acebc659e794e\", repo tag \"registry.k8s.io/kube-proxy:v1.31.13\", repo digest \"registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae\", size \"26953926\" in 1.467731541s" Sep 12 22:02:40.236931 containerd[1535]: time="2025-09-12T22:02:40.236609529Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\" returns image reference \"sha256:c15699f0b7002450249485b10f20211982dfd2bec4d61c86c35acebc659e794e\"" Sep 12 22:02:40.237095 containerd[1535]: time="2025-09-12T22:02:40.237072612Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 12 22:02:40.729548 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3652589701.mount: Deactivated successfully. Sep 12 22:02:41.346944 containerd[1535]: time="2025-09-12T22:02:41.346899461Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:41.349649 containerd[1535]: time="2025-09-12T22:02:41.349618853Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951624" Sep 12 22:02:41.355303 containerd[1535]: time="2025-09-12T22:02:41.355256730Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:41.358685 containerd[1535]: time="2025-09-12T22:02:41.358647947Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:41.360589 containerd[1535]: time="2025-09-12T22:02:41.360471003Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.123368547s" Sep 12 22:02:41.360589 containerd[1535]: time="2025-09-12T22:02:41.360503689Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Sep 12 22:02:41.361141 containerd[1535]: time="2025-09-12T22:02:41.361118880Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 12 22:02:41.803298 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2195502781.mount: Deactivated successfully. Sep 12 22:02:41.808537 containerd[1535]: time="2025-09-12T22:02:41.808475938Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 22:02:41.809760 containerd[1535]: time="2025-09-12T22:02:41.809729118Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Sep 12 22:02:41.810659 containerd[1535]: time="2025-09-12T22:02:41.810625732Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 22:02:41.813236 containerd[1535]: time="2025-09-12T22:02:41.813194682Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 22:02:41.813661 containerd[1535]: time="2025-09-12T22:02:41.813629685Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 452.47947ms" Sep 12 22:02:41.813661 containerd[1535]: time="2025-09-12T22:02:41.813657049Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Sep 12 22:02:41.814196 containerd[1535]: time="2025-09-12T22:02:41.814030771Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 12 22:02:42.334928 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3695049514.mount: Deactivated successfully. Sep 12 22:02:43.663561 containerd[1535]: time="2025-09-12T22:02:43.662615739Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:43.663561 containerd[1535]: time="2025-09-12T22:02:43.663040692Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66537163" Sep 12 22:02:43.664060 containerd[1535]: time="2025-09-12T22:02:43.664032738Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:43.667052 containerd[1535]: time="2025-09-12T22:02:43.667021487Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:02:43.668451 containerd[1535]: time="2025-09-12T22:02:43.668406938Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 1.854349189s" Sep 12 22:02:43.668451 containerd[1535]: time="2025-09-12T22:02:43.668449147Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Sep 12 22:02:48.397173 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:48.397319 systemd[1]: kubelet.service: Consumed 141ms CPU time, 108.1M memory peak. Sep 12 22:02:48.399101 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:02:48.418745 systemd[1]: Reload requested from client PID 2217 ('systemctl') (unit session-7.scope)... Sep 12 22:02:48.418864 systemd[1]: Reloading... Sep 12 22:02:48.488991 zram_generator::config[2259]: No configuration found. Sep 12 22:02:48.662145 systemd[1]: Reloading finished in 242 ms. Sep 12 22:02:48.718385 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 12 22:02:48.718457 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 12 22:02:48.718692 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:48.718735 systemd[1]: kubelet.service: Consumed 86ms CPU time, 95.1M memory peak. Sep 12 22:02:48.721155 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:02:48.852454 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:48.870300 (kubelet)[2304]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 22:02:48.901842 kubelet[2304]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:02:48.901842 kubelet[2304]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 12 22:02:48.901842 kubelet[2304]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:02:48.902211 kubelet[2304]: I0912 22:02:48.902038 2304 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 22:02:49.637511 kubelet[2304]: I0912 22:02:49.637463 2304 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 12 22:02:49.637511 kubelet[2304]: I0912 22:02:49.637494 2304 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 22:02:49.637763 kubelet[2304]: I0912 22:02:49.637744 2304 server.go:934] "Client rotation is on, will bootstrap in background" Sep 12 22:02:49.656314 kubelet[2304]: E0912 22:02:49.656273 2304 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.29:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" Sep 12 22:02:49.657006 kubelet[2304]: I0912 22:02:49.656991 2304 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 22:02:49.665388 kubelet[2304]: I0912 22:02:49.665266 2304 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 12 22:02:49.668923 kubelet[2304]: I0912 22:02:49.668877 2304 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 22:02:49.669670 kubelet[2304]: I0912 22:02:49.669647 2304 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 12 22:02:49.669817 kubelet[2304]: I0912 22:02:49.669788 2304 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 22:02:49.670081 kubelet[2304]: I0912 22:02:49.669816 2304 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 22:02:49.670171 kubelet[2304]: I0912 22:02:49.670153 2304 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 22:02:49.670171 kubelet[2304]: I0912 22:02:49.670163 2304 container_manager_linux.go:300] "Creating device plugin manager" Sep 12 22:02:49.670415 kubelet[2304]: I0912 22:02:49.670399 2304 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:02:49.672499 kubelet[2304]: I0912 22:02:49.672342 2304 kubelet.go:408] "Attempting to sync node with API server" Sep 12 22:02:49.672499 kubelet[2304]: I0912 22:02:49.672377 2304 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 22:02:49.672499 kubelet[2304]: I0912 22:02:49.672396 2304 kubelet.go:314] "Adding apiserver pod source" Sep 12 22:02:49.672499 kubelet[2304]: I0912 22:02:49.672474 2304 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 22:02:49.673966 kubelet[2304]: W0912 22:02:49.673921 2304 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.29:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.29:6443: connect: connection refused Sep 12 22:02:49.674109 kubelet[2304]: E0912 22:02:49.674092 2304 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.29:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" Sep 12 22:02:49.674206 kubelet[2304]: W0912 22:02:49.674162 2304 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.29:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.29:6443: connect: connection refused Sep 12 22:02:49.674244 kubelet[2304]: E0912 22:02:49.674212 2304 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.29:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" Sep 12 22:02:49.678326 kubelet[2304]: I0912 22:02:49.678309 2304 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 12 22:02:49.679007 kubelet[2304]: I0912 22:02:49.678989 2304 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 12 22:02:49.679173 kubelet[2304]: W0912 22:02:49.679160 2304 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 12 22:02:49.680020 kubelet[2304]: I0912 22:02:49.680002 2304 server.go:1274] "Started kubelet" Sep 12 22:02:49.680364 kubelet[2304]: I0912 22:02:49.680317 2304 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 22:02:49.680735 kubelet[2304]: I0912 22:02:49.680699 2304 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 22:02:49.680799 kubelet[2304]: I0912 22:02:49.680715 2304 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 22:02:49.681745 kubelet[2304]: I0912 22:02:49.681703 2304 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 22:02:49.682247 kubelet[2304]: I0912 22:02:49.682215 2304 server.go:449] "Adding debug handlers to kubelet server" Sep 12 22:02:49.682323 kubelet[2304]: I0912 22:02:49.682224 2304 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 22:02:49.683437 kubelet[2304]: I0912 22:02:49.683407 2304 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 12 22:02:49.683508 kubelet[2304]: I0912 22:02:49.683490 2304 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 12 22:02:49.683543 kubelet[2304]: I0912 22:02:49.683534 2304 reconciler.go:26] "Reconciler: start to sync state" Sep 12 22:02:49.683853 kubelet[2304]: E0912 22:02:49.683833 2304 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 22:02:49.683925 kubelet[2304]: W0912 22:02:49.683849 2304 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.29:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.29:6443: connect: connection refused Sep 12 22:02:49.684025 kubelet[2304]: E0912 22:02:49.683998 2304 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.29:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" Sep 12 22:02:49.684180 kubelet[2304]: E0912 22:02:49.684146 2304 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="200ms" Sep 12 22:02:49.684270 kubelet[2304]: E0912 22:02:49.684254 2304 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 12 22:02:49.687427 kubelet[2304]: I0912 22:02:49.687363 2304 factory.go:221] Registration of the containerd container factory successfully Sep 12 22:02:49.687660 kubelet[2304]: I0912 22:02:49.687568 2304 factory.go:221] Registration of the systemd container factory successfully Sep 12 22:02:49.687762 kubelet[2304]: I0912 22:02:49.687647 2304 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 22:02:49.689166 kubelet[2304]: E0912 22:02:49.688162 2304 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.29:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.29:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1864a8169e4d8afb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-12 22:02:49.679964923 +0000 UTC m=+0.806691989,LastTimestamp:2025-09-12 22:02:49.679964923 +0000 UTC m=+0.806691989,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 12 22:02:49.700321 kubelet[2304]: I0912 22:02:49.700235 2304 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 12 22:02:49.702137 kubelet[2304]: I0912 22:02:49.701874 2304 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 12 22:02:49.702137 kubelet[2304]: I0912 22:02:49.701890 2304 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 12 22:02:49.702137 kubelet[2304]: I0912 22:02:49.701906 2304 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:02:49.702275 kubelet[2304]: I0912 22:02:49.702168 2304 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 12 22:02:49.702275 kubelet[2304]: I0912 22:02:49.702191 2304 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 12 22:02:49.702275 kubelet[2304]: I0912 22:02:49.702212 2304 kubelet.go:2321] "Starting kubelet main sync loop" Sep 12 22:02:49.702275 kubelet[2304]: E0912 22:02:49.702252 2304 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 22:02:49.703214 kubelet[2304]: W0912 22:02:49.703173 2304 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.29:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.29:6443: connect: connection refused Sep 12 22:02:49.703281 kubelet[2304]: E0912 22:02:49.703219 2304 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.29:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" Sep 12 22:02:49.779373 kubelet[2304]: I0912 22:02:49.779323 2304 policy_none.go:49] "None policy: Start" Sep 12 22:02:49.780181 kubelet[2304]: I0912 22:02:49.780162 2304 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 12 22:02:49.780250 kubelet[2304]: I0912 22:02:49.780188 2304 state_mem.go:35] "Initializing new in-memory state store" Sep 12 22:02:49.784711 kubelet[2304]: E0912 22:02:49.784671 2304 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 12 22:02:49.787960 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 12 22:02:49.802542 kubelet[2304]: E0912 22:02:49.802400 2304 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 12 22:02:49.802758 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 12 22:02:49.805402 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 12 22:02:49.821790 kubelet[2304]: I0912 22:02:49.821692 2304 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 12 22:02:49.822209 kubelet[2304]: I0912 22:02:49.822179 2304 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 22:02:49.822249 kubelet[2304]: I0912 22:02:49.822198 2304 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 22:02:49.822437 kubelet[2304]: I0912 22:02:49.822382 2304 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 22:02:49.823534 kubelet[2304]: E0912 22:02:49.823512 2304 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 12 22:02:49.885083 kubelet[2304]: E0912 22:02:49.885041 2304 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="400ms" Sep 12 22:02:49.924256 kubelet[2304]: I0912 22:02:49.924167 2304 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 12 22:02:49.924684 kubelet[2304]: E0912 22:02:49.924657 2304 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 12 22:02:50.010949 systemd[1]: Created slice kubepods-burstable-pod71d8bf7bd9b7c7432927bee9d50592b5.slice - libcontainer container kubepods-burstable-pod71d8bf7bd9b7c7432927bee9d50592b5.slice. Sep 12 22:02:50.039227 systemd[1]: Created slice kubepods-burstable-pod3a240974f5c00dd9c8fb15f2a883738a.slice - libcontainer container kubepods-burstable-pod3a240974f5c00dd9c8fb15f2a883738a.slice. Sep 12 22:02:50.063732 systemd[1]: Created slice kubepods-burstable-podfe5e332fba00ba0b5b33a25fe2e8fd7b.slice - libcontainer container kubepods-burstable-podfe5e332fba00ba0b5b33a25fe2e8fd7b.slice. Sep 12 22:02:50.126734 kubelet[2304]: I0912 22:02:50.126687 2304 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 12 22:02:50.127169 kubelet[2304]: E0912 22:02:50.127144 2304 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 12 22:02:50.185116 kubelet[2304]: I0912 22:02:50.184993 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3a240974f5c00dd9c8fb15f2a883738a-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3a240974f5c00dd9c8fb15f2a883738a\") " pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:50.185116 kubelet[2304]: I0912 22:02:50.185067 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:50.185116 kubelet[2304]: I0912 22:02:50.185090 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:50.185116 kubelet[2304]: I0912 22:02:50.185108 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:50.185281 kubelet[2304]: I0912 22:02:50.185150 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:50.185281 kubelet[2304]: I0912 22:02:50.185168 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3a240974f5c00dd9c8fb15f2a883738a-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3a240974f5c00dd9c8fb15f2a883738a\") " pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:50.185281 kubelet[2304]: I0912 22:02:50.185183 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3a240974f5c00dd9c8fb15f2a883738a-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3a240974f5c00dd9c8fb15f2a883738a\") " pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:50.185281 kubelet[2304]: I0912 22:02:50.185199 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:50.185281 kubelet[2304]: I0912 22:02:50.185214 2304 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fe5e332fba00ba0b5b33a25fe2e8fd7b-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"fe5e332fba00ba0b5b33a25fe2e8fd7b\") " pod="kube-system/kube-scheduler-localhost" Sep 12 22:02:50.285590 kubelet[2304]: E0912 22:02:50.285541 2304 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="800ms" Sep 12 22:02:50.337924 kubelet[2304]: E0912 22:02:50.337885 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.338597 containerd[1535]: time="2025-09-12T22:02:50.338558116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:71d8bf7bd9b7c7432927bee9d50592b5,Namespace:kube-system,Attempt:0,}" Sep 12 22:02:50.354691 containerd[1535]: time="2025-09-12T22:02:50.354640238Z" level=info msg="connecting to shim 40e1203542face4a641f5ee6bda9ec988b937a139f41ebe87678c2c82f7775ff" address="unix:///run/containerd/s/87c2c217767af0b3fd25623aecc1a745d4af05b72e452b7b07ef5341fa14bcda" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:02:50.362995 kubelet[2304]: E0912 22:02:50.362813 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.363354 containerd[1535]: time="2025-09-12T22:02:50.363326376Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3a240974f5c00dd9c8fb15f2a883738a,Namespace:kube-system,Attempt:0,}" Sep 12 22:02:50.366196 kubelet[2304]: E0912 22:02:50.366174 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.367153 containerd[1535]: time="2025-09-12T22:02:50.367124034Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:fe5e332fba00ba0b5b33a25fe2e8fd7b,Namespace:kube-system,Attempt:0,}" Sep 12 22:02:50.375122 systemd[1]: Started cri-containerd-40e1203542face4a641f5ee6bda9ec988b937a139f41ebe87678c2c82f7775ff.scope - libcontainer container 40e1203542face4a641f5ee6bda9ec988b937a139f41ebe87678c2c82f7775ff. Sep 12 22:02:50.400024 containerd[1535]: time="2025-09-12T22:02:50.398305346Z" level=info msg="connecting to shim c59a150b8fdfcf02ff9b9fb9346379fece0fc47415952b4c7ec0db4333147061" address="unix:///run/containerd/s/f4bddb19900685c5234f0a24e4ea59a9aedb7a73317ef51d2ea3fd4cdaa899a7" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:02:50.401534 containerd[1535]: time="2025-09-12T22:02:50.401496026Z" level=info msg="connecting to shim 86139d748d2436abf57c3c66f27ab7606dbc9a4f83b66c66030d2990f4524cd6" address="unix:///run/containerd/s/3347df9ce007ee541f4f8a347d07464a25a878ac9c0edbaca3f5ee0a58bd3062" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:02:50.439175 systemd[1]: Started cri-containerd-86139d748d2436abf57c3c66f27ab7606dbc9a4f83b66c66030d2990f4524cd6.scope - libcontainer container 86139d748d2436abf57c3c66f27ab7606dbc9a4f83b66c66030d2990f4524cd6. Sep 12 22:02:50.442563 systemd[1]: Started cri-containerd-c59a150b8fdfcf02ff9b9fb9346379fece0fc47415952b4c7ec0db4333147061.scope - libcontainer container c59a150b8fdfcf02ff9b9fb9346379fece0fc47415952b4c7ec0db4333147061. Sep 12 22:02:50.443918 containerd[1535]: time="2025-09-12T22:02:50.443850144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:71d8bf7bd9b7c7432927bee9d50592b5,Namespace:kube-system,Attempt:0,} returns sandbox id \"40e1203542face4a641f5ee6bda9ec988b937a139f41ebe87678c2c82f7775ff\"" Sep 12 22:02:50.445245 kubelet[2304]: E0912 22:02:50.445220 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.449064 containerd[1535]: time="2025-09-12T22:02:50.449018703Z" level=info msg="CreateContainer within sandbox \"40e1203542face4a641f5ee6bda9ec988b937a139f41ebe87678c2c82f7775ff\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 12 22:02:50.458588 containerd[1535]: time="2025-09-12T22:02:50.458556695Z" level=info msg="Container 541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:02:50.465758 containerd[1535]: time="2025-09-12T22:02:50.465720657Z" level=info msg="CreateContainer within sandbox \"40e1203542face4a641f5ee6bda9ec988b937a139f41ebe87678c2c82f7775ff\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd\"" Sep 12 22:02:50.466697 containerd[1535]: time="2025-09-12T22:02:50.466666229Z" level=info msg="StartContainer for \"541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd\"" Sep 12 22:02:50.467919 containerd[1535]: time="2025-09-12T22:02:50.467886853Z" level=info msg="connecting to shim 541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd" address="unix:///run/containerd/s/87c2c217767af0b3fd25623aecc1a745d4af05b72e452b7b07ef5341fa14bcda" protocol=ttrpc version=3 Sep 12 22:02:50.483628 containerd[1535]: time="2025-09-12T22:02:50.483589349Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3a240974f5c00dd9c8fb15f2a883738a,Namespace:kube-system,Attempt:0,} returns sandbox id \"86139d748d2436abf57c3c66f27ab7606dbc9a4f83b66c66030d2990f4524cd6\"" Sep 12 22:02:50.484381 kubelet[2304]: E0912 22:02:50.484336 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.486643 containerd[1535]: time="2025-09-12T22:02:50.486543855Z" level=info msg="CreateContainer within sandbox \"86139d748d2436abf57c3c66f27ab7606dbc9a4f83b66c66030d2990f4524cd6\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 12 22:02:50.493139 systemd[1]: Started cri-containerd-541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd.scope - libcontainer container 541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd. Sep 12 22:02:50.494645 containerd[1535]: time="2025-09-12T22:02:50.494617186Z" level=info msg="Container a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:02:50.496209 containerd[1535]: time="2025-09-12T22:02:50.495785522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:fe5e332fba00ba0b5b33a25fe2e8fd7b,Namespace:kube-system,Attempt:0,} returns sandbox id \"c59a150b8fdfcf02ff9b9fb9346379fece0fc47415952b4c7ec0db4333147061\"" Sep 12 22:02:50.496965 kubelet[2304]: E0912 22:02:50.496940 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.499429 containerd[1535]: time="2025-09-12T22:02:50.499395063Z" level=info msg="CreateContainer within sandbox \"c59a150b8fdfcf02ff9b9fb9346379fece0fc47415952b4c7ec0db4333147061\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 12 22:02:50.500716 containerd[1535]: time="2025-09-12T22:02:50.500668214Z" level=info msg="CreateContainer within sandbox \"86139d748d2436abf57c3c66f27ab7606dbc9a4f83b66c66030d2990f4524cd6\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25\"" Sep 12 22:02:50.501289 containerd[1535]: time="2025-09-12T22:02:50.501261342Z" level=info msg="StartContainer for \"a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25\"" Sep 12 22:02:50.502457 containerd[1535]: time="2025-09-12T22:02:50.502427482Z" level=info msg="connecting to shim a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25" address="unix:///run/containerd/s/3347df9ce007ee541f4f8a347d07464a25a878ac9c0edbaca3f5ee0a58bd3062" protocol=ttrpc version=3 Sep 12 22:02:50.505608 containerd[1535]: time="2025-09-12T22:02:50.505545957Z" level=info msg="Container f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:02:50.512502 containerd[1535]: time="2025-09-12T22:02:50.512450394Z" level=info msg="CreateContainer within sandbox \"c59a150b8fdfcf02ff9b9fb9346379fece0fc47415952b4c7ec0db4333147061\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd\"" Sep 12 22:02:50.514150 containerd[1535]: time="2025-09-12T22:02:50.513099004Z" level=info msg="StartContainer for \"f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd\"" Sep 12 22:02:50.514150 containerd[1535]: time="2025-09-12T22:02:50.514023861Z" level=info msg="connecting to shim f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd" address="unix:///run/containerd/s/f4bddb19900685c5234f0a24e4ea59a9aedb7a73317ef51d2ea3fd4cdaa899a7" protocol=ttrpc version=3 Sep 12 22:02:50.525143 systemd[1]: Started cri-containerd-a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25.scope - libcontainer container a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25. Sep 12 22:02:50.529276 kubelet[2304]: I0912 22:02:50.529222 2304 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 12 22:02:50.531055 kubelet[2304]: E0912 22:02:50.529605 2304 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 12 22:02:50.540165 systemd[1]: Started cri-containerd-f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd.scope - libcontainer container f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd. Sep 12 22:02:50.546810 containerd[1535]: time="2025-09-12T22:02:50.546733696Z" level=info msg="StartContainer for \"541c35b81308434ba6124e2a9542366ae33fe16e2d175c7b1759b9c34bd10afd\" returns successfully" Sep 12 22:02:50.578174 containerd[1535]: time="2025-09-12T22:02:50.578131463Z" level=info msg="StartContainer for \"a592a915268a516c83cea139ed99f61e4dfb1279ea4d3d00d0a19ca091c0bb25\" returns successfully" Sep 12 22:02:50.589837 containerd[1535]: time="2025-09-12T22:02:50.589802682Z" level=info msg="StartContainer for \"f16565f5e4a50f7458ae7eebfdfafb5cbeaa47900ff386ebaeea92a719870ddd\" returns successfully" Sep 12 22:02:50.708515 kubelet[2304]: E0912 22:02:50.708416 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.711530 kubelet[2304]: E0912 22:02:50.711509 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:50.714924 kubelet[2304]: E0912 22:02:50.714905 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:51.332215 kubelet[2304]: I0912 22:02:51.332183 2304 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 12 22:02:51.718325 kubelet[2304]: E0912 22:02:51.718229 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:52.401144 kubelet[2304]: E0912 22:02:52.401101 2304 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 12 22:02:52.417384 kubelet[2304]: E0912 22:02:52.417290 2304 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1864a8169e4d8afb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-12 22:02:49.679964923 +0000 UTC m=+0.806691989,LastTimestamp:2025-09-12 22:02:49.679964923 +0000 UTC m=+0.806691989,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 12 22:02:52.448377 kubelet[2304]: I0912 22:02:52.448342 2304 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Sep 12 22:02:52.476060 kubelet[2304]: E0912 22:02:52.476013 2304 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 12 22:02:52.476219 kubelet[2304]: E0912 22:02:52.476197 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:52.674340 kubelet[2304]: I0912 22:02:52.674244 2304 apiserver.go:52] "Watching apiserver" Sep 12 22:02:52.684360 kubelet[2304]: I0912 22:02:52.684327 2304 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 12 22:02:52.721381 kubelet[2304]: E0912 22:02:52.721351 2304 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:52.721530 kubelet[2304]: E0912 22:02:52.721514 2304 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:54.290453 systemd[1]: Reload requested from client PID 2580 ('systemctl') (unit session-7.scope)... Sep 12 22:02:54.290468 systemd[1]: Reloading... Sep 12 22:02:54.358044 zram_generator::config[2623]: No configuration found. Sep 12 22:02:54.607101 systemd[1]: Reloading finished in 316 ms. Sep 12 22:02:54.626336 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:02:54.635791 systemd[1]: kubelet.service: Deactivated successfully. Sep 12 22:02:54.636232 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:54.636369 systemd[1]: kubelet.service: Consumed 1.200s CPU time, 127.9M memory peak. Sep 12 22:02:54.638099 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:02:54.769458 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:02:54.772694 (kubelet)[2665]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 22:02:54.811998 kubelet[2665]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:02:54.811998 kubelet[2665]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 12 22:02:54.811998 kubelet[2665]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:02:54.812346 kubelet[2665]: I0912 22:02:54.812048 2665 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 22:02:54.819462 kubelet[2665]: I0912 22:02:54.819425 2665 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 12 22:02:54.819462 kubelet[2665]: I0912 22:02:54.819455 2665 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 22:02:54.819716 kubelet[2665]: I0912 22:02:54.819699 2665 server.go:934] "Client rotation is on, will bootstrap in background" Sep 12 22:02:54.821041 kubelet[2665]: I0912 22:02:54.821023 2665 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 12 22:02:54.825633 kubelet[2665]: I0912 22:02:54.825599 2665 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 22:02:54.830004 kubelet[2665]: I0912 22:02:54.829207 2665 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 12 22:02:54.832191 kubelet[2665]: I0912 22:02:54.831539 2665 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 22:02:54.832191 kubelet[2665]: I0912 22:02:54.831667 2665 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 12 22:02:54.832191 kubelet[2665]: I0912 22:02:54.831767 2665 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 22:02:54.832191 kubelet[2665]: I0912 22:02:54.831790 2665 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.831951 2665 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.831959 2665 container_manager_linux.go:300] "Creating device plugin manager" Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.832007 2665 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.832110 2665 kubelet.go:408] "Attempting to sync node with API server" Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.832123 2665 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.832141 2665 kubelet.go:314] "Adding apiserver pod source" Sep 12 22:02:54.832364 kubelet[2665]: I0912 22:02:54.832153 2665 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 22:02:54.836366 kubelet[2665]: I0912 22:02:54.836113 2665 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 12 22:02:54.839990 kubelet[2665]: I0912 22:02:54.839362 2665 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 12 22:02:54.839990 kubelet[2665]: I0912 22:02:54.839789 2665 server.go:1274] "Started kubelet" Sep 12 22:02:54.840214 kubelet[2665]: I0912 22:02:54.840184 2665 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 22:02:54.842098 kubelet[2665]: I0912 22:02:54.842078 2665 server.go:449] "Adding debug handlers to kubelet server" Sep 12 22:02:54.844333 kubelet[2665]: I0912 22:02:54.844285 2665 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 22:02:54.845205 kubelet[2665]: I0912 22:02:54.845179 2665 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 22:02:54.845255 kubelet[2665]: I0912 22:02:54.844770 2665 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 22:02:54.845328 kubelet[2665]: I0912 22:02:54.844667 2665 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 22:02:54.852551 kubelet[2665]: I0912 22:02:54.851675 2665 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 12 22:02:54.854096 kubelet[2665]: E0912 22:02:54.852668 2665 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 22:02:54.854096 kubelet[2665]: I0912 22:02:54.852906 2665 factory.go:221] Registration of the systemd container factory successfully Sep 12 22:02:54.855715 kubelet[2665]: I0912 22:02:54.854170 2665 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 22:02:54.855715 kubelet[2665]: I0912 22:02:54.853413 2665 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 12 22:02:54.855715 kubelet[2665]: I0912 22:02:54.855146 2665 reconciler.go:26] "Reconciler: start to sync state" Sep 12 22:02:54.858404 kubelet[2665]: I0912 22:02:54.857644 2665 factory.go:221] Registration of the containerd container factory successfully Sep 12 22:02:54.861161 kubelet[2665]: I0912 22:02:54.861118 2665 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 12 22:02:54.862068 kubelet[2665]: I0912 22:02:54.862025 2665 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 12 22:02:54.862068 kubelet[2665]: I0912 22:02:54.862060 2665 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 12 22:02:54.862139 kubelet[2665]: I0912 22:02:54.862086 2665 kubelet.go:2321] "Starting kubelet main sync loop" Sep 12 22:02:54.862139 kubelet[2665]: E0912 22:02:54.862127 2665 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 22:02:54.891802 kubelet[2665]: I0912 22:02:54.891603 2665 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 12 22:02:54.891802 kubelet[2665]: I0912 22:02:54.891621 2665 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 12 22:02:54.891802 kubelet[2665]: I0912 22:02:54.891640 2665 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:02:54.892694 kubelet[2665]: I0912 22:02:54.892670 2665 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 12 22:02:54.892846 kubelet[2665]: I0912 22:02:54.892818 2665 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 12 22:02:54.893080 kubelet[2665]: I0912 22:02:54.893067 2665 policy_none.go:49] "None policy: Start" Sep 12 22:02:54.894336 kubelet[2665]: I0912 22:02:54.894316 2665 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 12 22:02:54.894461 kubelet[2665]: I0912 22:02:54.894449 2665 state_mem.go:35] "Initializing new in-memory state store" Sep 12 22:02:54.894655 kubelet[2665]: I0912 22:02:54.894641 2665 state_mem.go:75] "Updated machine memory state" Sep 12 22:02:54.899292 kubelet[2665]: I0912 22:02:54.899259 2665 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 12 22:02:54.899913 kubelet[2665]: I0912 22:02:54.899883 2665 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 22:02:54.899979 kubelet[2665]: I0912 22:02:54.899904 2665 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 22:02:54.900342 kubelet[2665]: I0912 22:02:54.900265 2665 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 22:02:55.006769 kubelet[2665]: I0912 22:02:55.006733 2665 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 12 22:02:55.012991 kubelet[2665]: I0912 22:02:55.012929 2665 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Sep 12 22:02:55.013105 kubelet[2665]: I0912 22:02:55.013025 2665 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Sep 12 22:02:55.056589 kubelet[2665]: I0912 22:02:55.056543 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3a240974f5c00dd9c8fb15f2a883738a-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3a240974f5c00dd9c8fb15f2a883738a\") " pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:55.056589 kubelet[2665]: I0912 22:02:55.056588 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3a240974f5c00dd9c8fb15f2a883738a-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3a240974f5c00dd9c8fb15f2a883738a\") " pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:55.056743 kubelet[2665]: I0912 22:02:55.056635 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3a240974f5c00dd9c8fb15f2a883738a-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3a240974f5c00dd9c8fb15f2a883738a\") " pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:55.056743 kubelet[2665]: I0912 22:02:55.056654 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:55.056743 kubelet[2665]: I0912 22:02:55.056668 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:55.056743 kubelet[2665]: I0912 22:02:55.056684 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:55.056743 kubelet[2665]: I0912 22:02:55.056710 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:55.056861 kubelet[2665]: I0912 22:02:55.056725 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 12 22:02:55.056861 kubelet[2665]: I0912 22:02:55.056739 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fe5e332fba00ba0b5b33a25fe2e8fd7b-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"fe5e332fba00ba0b5b33a25fe2e8fd7b\") " pod="kube-system/kube-scheduler-localhost" Sep 12 22:02:55.269753 kubelet[2665]: E0912 22:02:55.269708 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:55.270692 kubelet[2665]: E0912 22:02:55.269793 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:55.270692 kubelet[2665]: E0912 22:02:55.270009 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:55.833432 kubelet[2665]: I0912 22:02:55.833380 2665 apiserver.go:52] "Watching apiserver" Sep 12 22:02:55.854556 kubelet[2665]: I0912 22:02:55.854492 2665 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 12 22:02:55.879052 kubelet[2665]: E0912 22:02:55.877961 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:55.879189 kubelet[2665]: E0912 22:02:55.879113 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:55.884108 kubelet[2665]: E0912 22:02:55.884028 2665 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 12 22:02:55.884485 kubelet[2665]: E0912 22:02:55.884239 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:55.900991 kubelet[2665]: I0912 22:02:55.900760 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.900722005 podStartE2EDuration="1.900722005s" podCreationTimestamp="2025-09-12 22:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:02:55.900702906 +0000 UTC m=+1.125182381" watchObservedRunningTime="2025-09-12 22:02:55.900722005 +0000 UTC m=+1.125201480" Sep 12 22:02:55.917389 kubelet[2665]: I0912 22:02:55.917323 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.91730459 podStartE2EDuration="1.91730459s" podCreationTimestamp="2025-09-12 22:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:02:55.916273003 +0000 UTC m=+1.140752478" watchObservedRunningTime="2025-09-12 22:02:55.91730459 +0000 UTC m=+1.141784065" Sep 12 22:02:55.917602 kubelet[2665]: I0912 22:02:55.917414 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.917409235 podStartE2EDuration="1.917409235s" podCreationTimestamp="2025-09-12 22:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:02:55.908280462 +0000 UTC m=+1.132759897" watchObservedRunningTime="2025-09-12 22:02:55.917409235 +0000 UTC m=+1.141888710" Sep 12 22:02:56.880153 kubelet[2665]: E0912 22:02:56.880116 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:02:59.891427 kubelet[2665]: E0912 22:02:59.891387 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:00.797589 kubelet[2665]: E0912 22:03:00.797552 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:00.887552 kubelet[2665]: E0912 22:03:00.886697 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:01.461382 kubelet[2665]: I0912 22:03:01.461339 2665 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 12 22:03:01.461788 containerd[1535]: time="2025-09-12T22:03:01.461733252Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 12 22:03:01.462030 kubelet[2665]: I0912 22:03:01.461951 2665 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 12 22:03:02.412650 systemd[1]: Created slice kubepods-besteffort-pod09294417_5d2f_4090_b201_efb0e5ca0d6c.slice - libcontainer container kubepods-besteffort-pod09294417_5d2f_4090_b201_efb0e5ca0d6c.slice. Sep 12 22:03:02.499927 kubelet[2665]: I0912 22:03:02.499874 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/09294417-5d2f-4090-b201-efb0e5ca0d6c-kube-proxy\") pod \"kube-proxy-dz7mp\" (UID: \"09294417-5d2f-4090-b201-efb0e5ca0d6c\") " pod="kube-system/kube-proxy-dz7mp" Sep 12 22:03:02.499927 kubelet[2665]: I0912 22:03:02.499919 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09294417-5d2f-4090-b201-efb0e5ca0d6c-lib-modules\") pod \"kube-proxy-dz7mp\" (UID: \"09294417-5d2f-4090-b201-efb0e5ca0d6c\") " pod="kube-system/kube-proxy-dz7mp" Sep 12 22:03:02.499927 kubelet[2665]: I0912 22:03:02.499936 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f57lk\" (UniqueName: \"kubernetes.io/projected/09294417-5d2f-4090-b201-efb0e5ca0d6c-kube-api-access-f57lk\") pod \"kube-proxy-dz7mp\" (UID: \"09294417-5d2f-4090-b201-efb0e5ca0d6c\") " pod="kube-system/kube-proxy-dz7mp" Sep 12 22:03:02.500340 kubelet[2665]: I0912 22:03:02.499956 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/09294417-5d2f-4090-b201-efb0e5ca0d6c-xtables-lock\") pod \"kube-proxy-dz7mp\" (UID: \"09294417-5d2f-4090-b201-efb0e5ca0d6c\") " pod="kube-system/kube-proxy-dz7mp" Sep 12 22:03:02.638833 systemd[1]: Created slice kubepods-besteffort-pod17c7219b_dc71_4320_a176_d249cb7197d1.slice - libcontainer container kubepods-besteffort-pod17c7219b_dc71_4320_a176_d249cb7197d1.slice. Sep 12 22:03:02.701325 kubelet[2665]: I0912 22:03:02.701196 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdxcj\" (UniqueName: \"kubernetes.io/projected/17c7219b-dc71-4320-a176-d249cb7197d1-kube-api-access-xdxcj\") pod \"tigera-operator-58fc44c59b-mchkf\" (UID: \"17c7219b-dc71-4320-a176-d249cb7197d1\") " pod="tigera-operator/tigera-operator-58fc44c59b-mchkf" Sep 12 22:03:02.701325 kubelet[2665]: I0912 22:03:02.701272 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/17c7219b-dc71-4320-a176-d249cb7197d1-var-lib-calico\") pod \"tigera-operator-58fc44c59b-mchkf\" (UID: \"17c7219b-dc71-4320-a176-d249cb7197d1\") " pod="tigera-operator/tigera-operator-58fc44c59b-mchkf" Sep 12 22:03:02.731480 kubelet[2665]: E0912 22:03:02.731418 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:02.732277 containerd[1535]: time="2025-09-12T22:03:02.732004237Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-dz7mp,Uid:09294417-5d2f-4090-b201-efb0e5ca0d6c,Namespace:kube-system,Attempt:0,}" Sep 12 22:03:02.750161 containerd[1535]: time="2025-09-12T22:03:02.750068072Z" level=info msg="connecting to shim 479c6d44e3d355f54f6ad2865298ffa1ac60db24361a139e5079abb8aae5d602" address="unix:///run/containerd/s/4ab5a6703bbe70c6a565b0585d253fb319233b4c148b1343f9a667e3581eb79e" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:02.782194 systemd[1]: Started cri-containerd-479c6d44e3d355f54f6ad2865298ffa1ac60db24361a139e5079abb8aae5d602.scope - libcontainer container 479c6d44e3d355f54f6ad2865298ffa1ac60db24361a139e5079abb8aae5d602. Sep 12 22:03:02.803000 containerd[1535]: time="2025-09-12T22:03:02.802884605Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-dz7mp,Uid:09294417-5d2f-4090-b201-efb0e5ca0d6c,Namespace:kube-system,Attempt:0,} returns sandbox id \"479c6d44e3d355f54f6ad2865298ffa1ac60db24361a139e5079abb8aae5d602\"" Sep 12 22:03:02.804339 kubelet[2665]: E0912 22:03:02.804311 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:02.807983 containerd[1535]: time="2025-09-12T22:03:02.807232183Z" level=info msg="CreateContainer within sandbox \"479c6d44e3d355f54f6ad2865298ffa1ac60db24361a139e5079abb8aae5d602\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 12 22:03:02.822158 containerd[1535]: time="2025-09-12T22:03:02.822041627Z" level=info msg="Container 4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:02.830005 containerd[1535]: time="2025-09-12T22:03:02.829853768Z" level=info msg="CreateContainer within sandbox \"479c6d44e3d355f54f6ad2865298ffa1ac60db24361a139e5079abb8aae5d602\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041\"" Sep 12 22:03:02.830915 containerd[1535]: time="2025-09-12T22:03:02.830887940Z" level=info msg="StartContainer for \"4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041\"" Sep 12 22:03:02.835000 containerd[1535]: time="2025-09-12T22:03:02.834927483Z" level=info msg="connecting to shim 4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041" address="unix:///run/containerd/s/4ab5a6703bbe70c6a565b0585d253fb319233b4c148b1343f9a667e3581eb79e" protocol=ttrpc version=3 Sep 12 22:03:02.855171 systemd[1]: Started cri-containerd-4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041.scope - libcontainer container 4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041. Sep 12 22:03:02.890528 containerd[1535]: time="2025-09-12T22:03:02.890434430Z" level=info msg="StartContainer for \"4c5867476084b5b86d4120c0f5cfa13af75e057a1b18d704529c26a2565a3041\" returns successfully" Sep 12 22:03:02.943690 containerd[1535]: time="2025-09-12T22:03:02.943640778Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-58fc44c59b-mchkf,Uid:17c7219b-dc71-4320-a176-d249cb7197d1,Namespace:tigera-operator,Attempt:0,}" Sep 12 22:03:02.959598 containerd[1535]: time="2025-09-12T22:03:02.959180119Z" level=info msg="connecting to shim 1cc15fd6c24441b63127d92d3cf1e822c02e8c04b440d3ba848153e20a4e03be" address="unix:///run/containerd/s/79970375adc6cece2a6c2b2025ac0ba3b2b0880f07bb3bfcd489fea94f2c27dc" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:02.984202 systemd[1]: Started cri-containerd-1cc15fd6c24441b63127d92d3cf1e822c02e8c04b440d3ba848153e20a4e03be.scope - libcontainer container 1cc15fd6c24441b63127d92d3cf1e822c02e8c04b440d3ba848153e20a4e03be. Sep 12 22:03:03.023501 containerd[1535]: time="2025-09-12T22:03:03.023458784Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-58fc44c59b-mchkf,Uid:17c7219b-dc71-4320-a176-d249cb7197d1,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"1cc15fd6c24441b63127d92d3cf1e822c02e8c04b440d3ba848153e20a4e03be\"" Sep 12 22:03:03.026284 containerd[1535]: time="2025-09-12T22:03:03.026214779Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 12 22:03:03.621037 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1624124117.mount: Deactivated successfully. Sep 12 22:03:03.897967 kubelet[2665]: E0912 22:03:03.897391 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:04.742044 kubelet[2665]: E0912 22:03:04.741630 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:04.759702 kubelet[2665]: I0912 22:03:04.759630 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-dz7mp" podStartSLOduration=2.759610923 podStartE2EDuration="2.759610923s" podCreationTimestamp="2025-09-12 22:03:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:03:03.910393107 +0000 UTC m=+9.134872582" watchObservedRunningTime="2025-09-12 22:03:04.759610923 +0000 UTC m=+9.984090398" Sep 12 22:03:04.772274 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2123130266.mount: Deactivated successfully. Sep 12 22:03:04.899236 kubelet[2665]: E0912 22:03:04.899202 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:05.161399 containerd[1535]: time="2025-09-12T22:03:05.161280573Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:05.162334 containerd[1535]: time="2025-09-12T22:03:05.161769234Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=22152365" Sep 12 22:03:05.162758 containerd[1535]: time="2025-09-12T22:03:05.162720710Z" level=info msg="ImageCreate event name:\"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:05.164664 containerd[1535]: time="2025-09-12T22:03:05.164631704Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:05.165715 containerd[1535]: time="2025-09-12T22:03:05.165169775Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"22148360\" in 2.138892382s" Sep 12 22:03:05.165715 containerd[1535]: time="2025-09-12T22:03:05.165195860Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:dd2e197838b00861b08ae5f480dfbfb9a519722e35ced99346315722309cbe9f\"" Sep 12 22:03:05.167227 containerd[1535]: time="2025-09-12T22:03:05.167191551Z" level=info msg="CreateContainer within sandbox \"1cc15fd6c24441b63127d92d3cf1e822c02e8c04b440d3ba848153e20a4e03be\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 12 22:03:05.177950 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1190613377.mount: Deactivated successfully. Sep 12 22:03:05.178663 containerd[1535]: time="2025-09-12T22:03:05.178631790Z" level=info msg="Container 199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:05.187370 containerd[1535]: time="2025-09-12T22:03:05.187334904Z" level=info msg="CreateContainer within sandbox \"1cc15fd6c24441b63127d92d3cf1e822c02e8c04b440d3ba848153e20a4e03be\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b\"" Sep 12 22:03:05.189014 containerd[1535]: time="2025-09-12T22:03:05.188618329Z" level=info msg="StartContainer for \"199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b\"" Sep 12 22:03:05.189650 containerd[1535]: time="2025-09-12T22:03:05.189622416Z" level=info msg="connecting to shim 199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b" address="unix:///run/containerd/s/79970375adc6cece2a6c2b2025ac0ba3b2b0880f07bb3bfcd489fea94f2c27dc" protocol=ttrpc version=3 Sep 12 22:03:05.210143 systemd[1]: Started cri-containerd-199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b.scope - libcontainer container 199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b. Sep 12 22:03:05.232438 containerd[1535]: time="2025-09-12T22:03:05.232391913Z" level=info msg="StartContainer for \"199d88a6d66538e7faa75bed8e6ff36dcf1d8dbcb31ffb6a30fbf8193523af3b\" returns successfully" Sep 12 22:03:05.910456 kubelet[2665]: I0912 22:03:05.910005 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-58fc44c59b-mchkf" podStartSLOduration=1.769329999 podStartE2EDuration="3.909967203s" podCreationTimestamp="2025-09-12 22:03:02 +0000 UTC" firstStartedPulling="2025-09-12 22:03:03.02530725 +0000 UTC m=+8.249786725" lastFinishedPulling="2025-09-12 22:03:05.165944454 +0000 UTC m=+10.390423929" observedRunningTime="2025-09-12 22:03:05.909802889 +0000 UTC m=+11.134282364" watchObservedRunningTime="2025-09-12 22:03:05.909967203 +0000 UTC m=+11.134446678" Sep 12 22:03:09.898461 kubelet[2665]: E0912 22:03:09.898409 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:10.699495 sudo[1744]: pam_unix(sudo:session): session closed for user root Sep 12 22:03:10.703036 sshd[1743]: Connection closed by 10.0.0.1 port 39486 Sep 12 22:03:10.703500 sshd-session[1739]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:10.707360 systemd[1]: sshd@6-10.0.0.29:22-10.0.0.1:39486.service: Deactivated successfully. Sep 12 22:03:10.710946 systemd[1]: session-7.scope: Deactivated successfully. Sep 12 22:03:10.713058 systemd[1]: session-7.scope: Consumed 6.375s CPU time, 219.9M memory peak. Sep 12 22:03:10.716233 systemd-logind[1513]: Session 7 logged out. Waiting for processes to exit. Sep 12 22:03:10.719279 systemd-logind[1513]: Removed session 7. Sep 12 22:03:12.573513 update_engine[1515]: I20250912 22:03:12.573019 1515 update_attempter.cc:509] Updating boot flags... Sep 12 22:03:16.132962 systemd[1]: Created slice kubepods-besteffort-podee4304ec_f504_4602_8ac8_eca9080193dc.slice - libcontainer container kubepods-besteffort-podee4304ec_f504_4602_8ac8_eca9080193dc.slice. Sep 12 22:03:16.287568 kubelet[2665]: I0912 22:03:16.287519 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfwxg\" (UniqueName: \"kubernetes.io/projected/ee4304ec-f504-4602-8ac8-eca9080193dc-kube-api-access-rfwxg\") pod \"calico-typha-7d4554c56d-mqdrs\" (UID: \"ee4304ec-f504-4602-8ac8-eca9080193dc\") " pod="calico-system/calico-typha-7d4554c56d-mqdrs" Sep 12 22:03:16.287938 kubelet[2665]: I0912 22:03:16.287591 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/ee4304ec-f504-4602-8ac8-eca9080193dc-typha-certs\") pod \"calico-typha-7d4554c56d-mqdrs\" (UID: \"ee4304ec-f504-4602-8ac8-eca9080193dc\") " pod="calico-system/calico-typha-7d4554c56d-mqdrs" Sep 12 22:03:16.287938 kubelet[2665]: I0912 22:03:16.287616 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee4304ec-f504-4602-8ac8-eca9080193dc-tigera-ca-bundle\") pod \"calico-typha-7d4554c56d-mqdrs\" (UID: \"ee4304ec-f504-4602-8ac8-eca9080193dc\") " pod="calico-system/calico-typha-7d4554c56d-mqdrs" Sep 12 22:03:16.440210 kubelet[2665]: E0912 22:03:16.439847 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:16.440447 containerd[1535]: time="2025-09-12T22:03:16.440409241Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7d4554c56d-mqdrs,Uid:ee4304ec-f504-4602-8ac8-eca9080193dc,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:16.466816 systemd[1]: Created slice kubepods-besteffort-pod3a8546cb_cf9e_4cdc_b9eb_41764943cd03.slice - libcontainer container kubepods-besteffort-pod3a8546cb_cf9e_4cdc_b9eb_41764943cd03.slice. Sep 12 22:03:16.483567 containerd[1535]: time="2025-09-12T22:03:16.483517153Z" level=info msg="connecting to shim 1e9f247838a964dd14aa176148b8f847309f9815e09ddc374011f5af7f1655c1" address="unix:///run/containerd/s/59100926c4bfaf31646761f5c5df9f154dd82163d1f73039baa4e680808a74ee" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:16.529035 systemd[1]: Started cri-containerd-1e9f247838a964dd14aa176148b8f847309f9815e09ddc374011f5af7f1655c1.scope - libcontainer container 1e9f247838a964dd14aa176148b8f847309f9815e09ddc374011f5af7f1655c1. Sep 12 22:03:16.573999 containerd[1535]: time="2025-09-12T22:03:16.573930392Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7d4554c56d-mqdrs,Uid:ee4304ec-f504-4602-8ac8-eca9080193dc,Namespace:calico-system,Attempt:0,} returns sandbox id \"1e9f247838a964dd14aa176148b8f847309f9815e09ddc374011f5af7f1655c1\"" Sep 12 22:03:16.577634 kubelet[2665]: E0912 22:03:16.577608 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:16.579654 containerd[1535]: time="2025-09-12T22:03:16.579626823Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 12 22:03:16.589133 kubelet[2665]: I0912 22:03:16.589090 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-node-certs\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589133 kubelet[2665]: I0912 22:03:16.589132 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-tigera-ca-bundle\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589265 kubelet[2665]: I0912 22:03:16.589164 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-var-run-calico\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589265 kubelet[2665]: I0912 22:03:16.589185 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-cni-bin-dir\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589265 kubelet[2665]: I0912 22:03:16.589202 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-policysync\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589265 kubelet[2665]: I0912 22:03:16.589216 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-xtables-lock\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589265 kubelet[2665]: I0912 22:03:16.589261 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-cni-log-dir\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589373 kubelet[2665]: I0912 22:03:16.589280 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-lib-modules\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589373 kubelet[2665]: I0912 22:03:16.589311 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwj8t\" (UniqueName: \"kubernetes.io/projected/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-kube-api-access-zwj8t\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589373 kubelet[2665]: I0912 22:03:16.589331 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-var-lib-calico\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589373 kubelet[2665]: I0912 22:03:16.589346 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-cni-net-dir\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.589373 kubelet[2665]: I0912 22:03:16.589361 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/3a8546cb-cf9e-4cdc-b9eb-41764943cd03-flexvol-driver-host\") pod \"calico-node-cjs4k\" (UID: \"3a8546cb-cf9e-4cdc-b9eb-41764943cd03\") " pod="calico-system/calico-node-cjs4k" Sep 12 22:03:16.697435 kubelet[2665]: E0912 22:03:16.696407 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.697435 kubelet[2665]: W0912 22:03:16.696428 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.697435 kubelet[2665]: E0912 22:03:16.696448 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.698084 kubelet[2665]: E0912 22:03:16.698050 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.698084 kubelet[2665]: W0912 22:03:16.698072 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.698084 kubelet[2665]: E0912 22:03:16.698087 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.698809 kubelet[2665]: E0912 22:03:16.698762 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wdhc9" podUID="9741ef86-7513-4899-85e2-cf42c12fa232" Sep 12 22:03:16.699346 kubelet[2665]: E0912 22:03:16.699324 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.699346 kubelet[2665]: W0912 22:03:16.699340 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.699752 kubelet[2665]: E0912 22:03:16.699361 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.702150 kubelet[2665]: E0912 22:03:16.700921 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.702150 kubelet[2665]: W0912 22:03:16.701812 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.702150 kubelet[2665]: E0912 22:03:16.701879 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.702796 kubelet[2665]: E0912 22:03:16.702756 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.702796 kubelet[2665]: W0912 22:03:16.702780 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.702796 kubelet[2665]: E0912 22:03:16.702795 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.703756 kubelet[2665]: E0912 22:03:16.703725 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.703756 kubelet[2665]: W0912 22:03:16.703742 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.703756 kubelet[2665]: E0912 22:03:16.703755 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.713141 kubelet[2665]: E0912 22:03:16.713113 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.713141 kubelet[2665]: W0912 22:03:16.713140 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.713305 kubelet[2665]: E0912 22:03:16.713184 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.772377 containerd[1535]: time="2025-09-12T22:03:16.772316857Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cjs4k,Uid:3a8546cb-cf9e-4cdc-b9eb-41764943cd03,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:16.791774 kubelet[2665]: E0912 22:03:16.791733 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.791774 kubelet[2665]: W0912 22:03:16.791758 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.791774 kubelet[2665]: E0912 22:03:16.791779 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.791962 kubelet[2665]: E0912 22:03:16.791936 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.791962 kubelet[2665]: W0912 22:03:16.791948 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.792040 kubelet[2665]: E0912 22:03:16.791982 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.792832 kubelet[2665]: E0912 22:03:16.792804 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.792832 kubelet[2665]: W0912 22:03:16.792820 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.792832 kubelet[2665]: E0912 22:03:16.792834 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.793752 kubelet[2665]: E0912 22:03:16.793319 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.793752 kubelet[2665]: W0912 22:03:16.793335 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.793752 kubelet[2665]: E0912 22:03:16.793347 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.794676 kubelet[2665]: E0912 22:03:16.793893 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.794676 kubelet[2665]: W0912 22:03:16.793910 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.794676 kubelet[2665]: E0912 22:03:16.793922 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.794676 kubelet[2665]: E0912 22:03:16.794395 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.794676 kubelet[2665]: W0912 22:03:16.794408 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.794676 kubelet[2665]: E0912 22:03:16.794537 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.795535 kubelet[2665]: E0912 22:03:16.795301 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.795535 kubelet[2665]: W0912 22:03:16.795317 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.795535 kubelet[2665]: E0912 22:03:16.795329 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.796094 kubelet[2665]: E0912 22:03:16.796068 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.796094 kubelet[2665]: W0912 22:03:16.796087 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.796204 kubelet[2665]: E0912 22:03:16.796102 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.796332 kubelet[2665]: E0912 22:03:16.796311 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.796332 kubelet[2665]: W0912 22:03:16.796326 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.796384 kubelet[2665]: E0912 22:03:16.796336 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.796484 kubelet[2665]: E0912 22:03:16.796467 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.796484 kubelet[2665]: W0912 22:03:16.796481 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.796553 kubelet[2665]: E0912 22:03:16.796490 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797010 kubelet[2665]: E0912 22:03:16.796632 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797010 kubelet[2665]: W0912 22:03:16.796643 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797010 kubelet[2665]: E0912 22:03:16.796651 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797010 kubelet[2665]: E0912 22:03:16.796773 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797010 kubelet[2665]: W0912 22:03:16.796780 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797010 kubelet[2665]: E0912 22:03:16.796787 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797010 kubelet[2665]: E0912 22:03:16.796906 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797010 kubelet[2665]: W0912 22:03:16.796913 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797010 kubelet[2665]: E0912 22:03:16.796933 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797258 kubelet[2665]: E0912 22:03:16.797049 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797258 kubelet[2665]: W0912 22:03:16.797056 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797258 kubelet[2665]: E0912 22:03:16.797064 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797258 kubelet[2665]: E0912 22:03:16.797221 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797258 kubelet[2665]: W0912 22:03:16.797229 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797258 kubelet[2665]: E0912 22:03:16.797236 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797368 kubelet[2665]: E0912 22:03:16.797360 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797389 kubelet[2665]: W0912 22:03:16.797367 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797389 kubelet[2665]: E0912 22:03:16.797374 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797542 kubelet[2665]: E0912 22:03:16.797516 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797542 kubelet[2665]: W0912 22:03:16.797538 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797598 kubelet[2665]: E0912 22:03:16.797548 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.797696 kubelet[2665]: E0912 22:03:16.797680 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.797696 kubelet[2665]: W0912 22:03:16.797692 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.797756 kubelet[2665]: E0912 22:03:16.797704 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.798648 kubelet[2665]: E0912 22:03:16.798046 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.798648 kubelet[2665]: W0912 22:03:16.798063 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.798648 kubelet[2665]: E0912 22:03:16.798073 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.798648 kubelet[2665]: E0912 22:03:16.798363 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.798648 kubelet[2665]: W0912 22:03:16.798374 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.798648 kubelet[2665]: E0912 22:03:16.798384 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.812655 containerd[1535]: time="2025-09-12T22:03:16.812609398Z" level=info msg="connecting to shim 1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79" address="unix:///run/containerd/s/a684c19efef5e689a6e07cc6551800318417f9668c42a95a72d09d4f66c733be" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:16.838186 systemd[1]: Started cri-containerd-1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79.scope - libcontainer container 1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79. Sep 12 22:03:16.868755 containerd[1535]: time="2025-09-12T22:03:16.868696118Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cjs4k,Uid:3a8546cb-cf9e-4cdc-b9eb-41764943cd03,Namespace:calico-system,Attempt:0,} returns sandbox id \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\"" Sep 12 22:03:16.892594 kubelet[2665]: E0912 22:03:16.892557 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.892594 kubelet[2665]: W0912 22:03:16.892588 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.892798 kubelet[2665]: E0912 22:03:16.892608 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.892798 kubelet[2665]: I0912 22:03:16.892636 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9741ef86-7513-4899-85e2-cf42c12fa232-kubelet-dir\") pod \"csi-node-driver-wdhc9\" (UID: \"9741ef86-7513-4899-85e2-cf42c12fa232\") " pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:16.892862 kubelet[2665]: E0912 22:03:16.892843 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.892862 kubelet[2665]: W0912 22:03:16.892854 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.892862 kubelet[2665]: E0912 22:03:16.892868 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.892862 kubelet[2665]: I0912 22:03:16.892893 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9741ef86-7513-4899-85e2-cf42c12fa232-registration-dir\") pod \"csi-node-driver-wdhc9\" (UID: \"9741ef86-7513-4899-85e2-cf42c12fa232\") " pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:16.893208 kubelet[2665]: E0912 22:03:16.893191 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.893269 kubelet[2665]: W0912 22:03:16.893254 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.893331 kubelet[2665]: E0912 22:03:16.893319 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.893555 kubelet[2665]: E0912 22:03:16.893543 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.893631 kubelet[2665]: W0912 22:03:16.893617 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.893698 kubelet[2665]: E0912 22:03:16.893687 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.893898 kubelet[2665]: E0912 22:03:16.893883 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.893898 kubelet[2665]: W0912 22:03:16.893898 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.893964 kubelet[2665]: E0912 22:03:16.893915 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.893964 kubelet[2665]: I0912 22:03:16.893931 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9741ef86-7513-4899-85e2-cf42c12fa232-socket-dir\") pod \"csi-node-driver-wdhc9\" (UID: \"9741ef86-7513-4899-85e2-cf42c12fa232\") " pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:16.894146 kubelet[2665]: E0912 22:03:16.894113 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.894146 kubelet[2665]: W0912 22:03:16.894139 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.894205 kubelet[2665]: E0912 22:03:16.894153 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.894349 kubelet[2665]: E0912 22:03:16.894325 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.894349 kubelet[2665]: W0912 22:03:16.894341 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.894666 kubelet[2665]: E0912 22:03:16.894634 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.894666 kubelet[2665]: W0912 22:03:16.894651 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.894752 kubelet[2665]: E0912 22:03:16.894741 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.894927 kubelet[2665]: I0912 22:03:16.894899 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/9741ef86-7513-4899-85e2-cf42c12fa232-varrun\") pod \"csi-node-driver-wdhc9\" (UID: \"9741ef86-7513-4899-85e2-cf42c12fa232\") " pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:16.895067 kubelet[2665]: E0912 22:03:16.895055 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.895112 kubelet[2665]: W0912 22:03:16.895103 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.895182 kubelet[2665]: E0912 22:03:16.895167 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.896540 kubelet[2665]: E0912 22:03:16.894354 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.896795 kubelet[2665]: E0912 22:03:16.896780 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.896795 kubelet[2665]: W0912 22:03:16.896795 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.896896 kubelet[2665]: E0912 22:03:16.896811 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.896990 kubelet[2665]: E0912 22:03:16.896963 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.897132 kubelet[2665]: W0912 22:03:16.896992 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.897132 kubelet[2665]: E0912 22:03:16.897001 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.897252 kubelet[2665]: E0912 22:03:16.897238 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.897320 kubelet[2665]: W0912 22:03:16.897308 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.897523 kubelet[2665]: E0912 22:03:16.897433 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.898276 kubelet[2665]: E0912 22:03:16.898260 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.899463 kubelet[2665]: W0912 22:03:16.899229 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.899463 kubelet[2665]: E0912 22:03:16.899257 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.899463 kubelet[2665]: I0912 22:03:16.899427 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czl8d\" (UniqueName: \"kubernetes.io/projected/9741ef86-7513-4899-85e2-cf42c12fa232-kube-api-access-czl8d\") pod \"csi-node-driver-wdhc9\" (UID: \"9741ef86-7513-4899-85e2-cf42c12fa232\") " pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:16.899589 kubelet[2665]: E0912 22:03:16.899516 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.899589 kubelet[2665]: W0912 22:03:16.899537 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.899589 kubelet[2665]: E0912 22:03:16.899551 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:16.900787 kubelet[2665]: E0912 22:03:16.899688 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:16.900787 kubelet[2665]: W0912 22:03:16.899699 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:16.900787 kubelet[2665]: E0912 22:03:16.899707 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.000526 kubelet[2665]: E0912 22:03:17.000480 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.000526 kubelet[2665]: W0912 22:03:17.000510 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.000526 kubelet[2665]: E0912 22:03:17.000528 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.000811 kubelet[2665]: E0912 22:03:17.000780 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.000811 kubelet[2665]: W0912 22:03:17.000795 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.000811 kubelet[2665]: E0912 22:03:17.000818 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.001023 kubelet[2665]: E0912 22:03:17.001011 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.001023 kubelet[2665]: W0912 22:03:17.001022 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.001134 kubelet[2665]: E0912 22:03:17.001118 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.001196 kubelet[2665]: E0912 22:03:17.001175 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.001196 kubelet[2665]: W0912 22:03:17.001195 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.001245 kubelet[2665]: E0912 22:03:17.001206 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.001402 kubelet[2665]: E0912 22:03:17.001388 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.001432 kubelet[2665]: W0912 22:03:17.001408 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.001432 kubelet[2665]: E0912 22:03:17.001421 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.001655 kubelet[2665]: E0912 22:03:17.001640 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.001655 kubelet[2665]: W0912 22:03:17.001652 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.001720 kubelet[2665]: E0912 22:03:17.001665 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.001887 kubelet[2665]: E0912 22:03:17.001872 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.001887 kubelet[2665]: W0912 22:03:17.001886 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.001945 kubelet[2665]: E0912 22:03:17.001899 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.002131 kubelet[2665]: E0912 22:03:17.002101 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.002131 kubelet[2665]: W0912 22:03:17.002115 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.002131 kubelet[2665]: E0912 22:03:17.002130 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.002306 kubelet[2665]: E0912 22:03:17.002291 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.002306 kubelet[2665]: W0912 22:03:17.002303 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.002367 kubelet[2665]: E0912 22:03:17.002344 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.002503 kubelet[2665]: E0912 22:03:17.002491 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.002532 kubelet[2665]: W0912 22:03:17.002502 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.002565 kubelet[2665]: E0912 22:03:17.002557 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.002714 kubelet[2665]: E0912 22:03:17.002700 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.002714 kubelet[2665]: W0912 22:03:17.002711 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.002806 kubelet[2665]: E0912 22:03:17.002752 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.002910 kubelet[2665]: E0912 22:03:17.002890 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.003020 kubelet[2665]: W0912 22:03:17.002910 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.003020 kubelet[2665]: E0912 22:03:17.002947 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.003106 kubelet[2665]: E0912 22:03:17.003091 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.003106 kubelet[2665]: W0912 22:03:17.003103 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.003163 kubelet[2665]: E0912 22:03:17.003151 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.003275 kubelet[2665]: E0912 22:03:17.003262 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.003275 kubelet[2665]: W0912 22:03:17.003273 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.003480 kubelet[2665]: E0912 22:03:17.003285 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.003572 kubelet[2665]: E0912 22:03:17.003555 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.003631 kubelet[2665]: W0912 22:03:17.003619 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.003688 kubelet[2665]: E0912 22:03:17.003678 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.003896 kubelet[2665]: E0912 22:03:17.003872 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.003896 kubelet[2665]: W0912 22:03:17.003885 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.003962 kubelet[2665]: E0912 22:03:17.003903 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.004088 kubelet[2665]: E0912 22:03:17.004068 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.004143 kubelet[2665]: W0912 22:03:17.004081 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.004143 kubelet[2665]: E0912 22:03:17.004113 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.004376 kubelet[2665]: E0912 22:03:17.004362 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.004376 kubelet[2665]: W0912 22:03:17.004375 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.004451 kubelet[2665]: E0912 22:03:17.004437 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.004625 kubelet[2665]: E0912 22:03:17.004608 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.004666 kubelet[2665]: W0912 22:03:17.004626 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.004697 kubelet[2665]: E0912 22:03:17.004688 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.004852 kubelet[2665]: E0912 22:03:17.004838 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.004852 kubelet[2665]: W0912 22:03:17.004851 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.004934 kubelet[2665]: E0912 22:03:17.004920 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.005095 kubelet[2665]: E0912 22:03:17.005079 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.005095 kubelet[2665]: W0912 22:03:17.005094 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.005162 kubelet[2665]: E0912 22:03:17.005152 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.005296 kubelet[2665]: E0912 22:03:17.005283 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.005296 kubelet[2665]: W0912 22:03:17.005295 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.005348 kubelet[2665]: E0912 22:03:17.005307 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.005515 kubelet[2665]: E0912 22:03:17.005503 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.005515 kubelet[2665]: W0912 22:03:17.005515 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.005589 kubelet[2665]: E0912 22:03:17.005531 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.005746 kubelet[2665]: E0912 22:03:17.005707 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.005746 kubelet[2665]: W0912 22:03:17.005742 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.005802 kubelet[2665]: E0912 22:03:17.005753 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.006345 kubelet[2665]: E0912 22:03:17.006323 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.006345 kubelet[2665]: W0912 22:03:17.006338 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.006345 kubelet[2665]: E0912 22:03:17.006350 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.014991 kubelet[2665]: E0912 22:03:17.014941 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:17.014991 kubelet[2665]: W0912 22:03:17.014958 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:17.015072 kubelet[2665]: E0912 22:03:17.015000 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:17.505322 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3861210025.mount: Deactivated successfully. Sep 12 22:03:17.888568 containerd[1535]: time="2025-09-12T22:03:17.888349513Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:17.888905 containerd[1535]: time="2025-09-12T22:03:17.888631785Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=33105775" Sep 12 22:03:17.889721 containerd[1535]: time="2025-09-12T22:03:17.889692704Z" level=info msg="ImageCreate event name:\"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:17.891941 containerd[1535]: time="2025-09-12T22:03:17.891898872Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:17.900986 containerd[1535]: time="2025-09-12T22:03:17.900926126Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"33105629\" in 1.321262619s" Sep 12 22:03:17.900986 containerd[1535]: time="2025-09-12T22:03:17.900963330Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:6a1496fdc48cc0b9ab3c10aef777497484efac5df9efbfbbdf9775e9583645cb\"" Sep 12 22:03:17.901714 containerd[1535]: time="2025-09-12T22:03:17.901687891Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 12 22:03:17.924400 containerd[1535]: time="2025-09-12T22:03:17.924367719Z" level=info msg="CreateContainer within sandbox \"1e9f247838a964dd14aa176148b8f847309f9815e09ddc374011f5af7f1655c1\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 12 22:03:17.930300 containerd[1535]: time="2025-09-12T22:03:17.930260821Z" level=info msg="Container 8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:17.936554 containerd[1535]: time="2025-09-12T22:03:17.936515083Z" level=info msg="CreateContainer within sandbox \"1e9f247838a964dd14aa176148b8f847309f9815e09ddc374011f5af7f1655c1\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4\"" Sep 12 22:03:17.939170 containerd[1535]: time="2025-09-12T22:03:17.939130017Z" level=info msg="StartContainer for \"8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4\"" Sep 12 22:03:17.940144 containerd[1535]: time="2025-09-12T22:03:17.940117488Z" level=info msg="connecting to shim 8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4" address="unix:///run/containerd/s/59100926c4bfaf31646761f5c5df9f154dd82163d1f73039baa4e680808a74ee" protocol=ttrpc version=3 Sep 12 22:03:17.963133 systemd[1]: Started cri-containerd-8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4.scope - libcontainer container 8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4. Sep 12 22:03:18.000746 containerd[1535]: time="2025-09-12T22:03:18.000702413Z" level=info msg="StartContainer for \"8c0fe5adb797ae8185b4f9e885c59aca0eca0a356ddf28f0974b8994a13e05f4\" returns successfully" Sep 12 22:03:18.862914 kubelet[2665]: E0912 22:03:18.862872 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wdhc9" podUID="9741ef86-7513-4899-85e2-cf42c12fa232" Sep 12 22:03:18.931436 kubelet[2665]: E0912 22:03:18.931392 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:18.943481 kubelet[2665]: I0912 22:03:18.943417 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7d4554c56d-mqdrs" podStartSLOduration=1.6212273270000002 podStartE2EDuration="2.943378807s" podCreationTimestamp="2025-09-12 22:03:16 +0000 UTC" firstStartedPulling="2025-09-12 22:03:16.579366112 +0000 UTC m=+21.803845587" lastFinishedPulling="2025-09-12 22:03:17.901517592 +0000 UTC m=+23.125997067" observedRunningTime="2025-09-12 22:03:18.943210709 +0000 UTC m=+24.167690184" watchObservedRunningTime="2025-09-12 22:03:18.943378807 +0000 UTC m=+24.167858282" Sep 12 22:03:18.959245 containerd[1535]: time="2025-09-12T22:03:18.959183903Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:18.960963 containerd[1535]: time="2025-09-12T22:03:18.960924450Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4266814" Sep 12 22:03:18.962066 containerd[1535]: time="2025-09-12T22:03:18.962037329Z" level=info msg="ImageCreate event name:\"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:18.970855 containerd[1535]: time="2025-09-12T22:03:18.970812951Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:18.971508 containerd[1535]: time="2025-09-12T22:03:18.971467461Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5636015\" in 1.069750207s" Sep 12 22:03:18.971508 containerd[1535]: time="2025-09-12T22:03:18.971504785Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:29e6f31ad72882b1b817dd257df6b7981e4d7d31d872b7fe2cf102c6e2af27a5\"" Sep 12 22:03:18.974634 containerd[1535]: time="2025-09-12T22:03:18.974605758Z" level=info msg="CreateContainer within sandbox \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 12 22:03:18.983246 containerd[1535]: time="2025-09-12T22:03:18.983170117Z" level=info msg="Container 6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:18.990751 containerd[1535]: time="2025-09-12T22:03:18.990717327Z" level=info msg="CreateContainer within sandbox \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\"" Sep 12 22:03:18.992001 containerd[1535]: time="2025-09-12T22:03:18.991270626Z" level=info msg="StartContainer for \"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\"" Sep 12 22:03:18.994924 containerd[1535]: time="2025-09-12T22:03:18.994865372Z" level=info msg="connecting to shim 6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb" address="unix:///run/containerd/s/a684c19efef5e689a6e07cc6551800318417f9668c42a95a72d09d4f66c733be" protocol=ttrpc version=3 Sep 12 22:03:19.010065 kubelet[2665]: E0912 22:03:19.010039 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.010150 kubelet[2665]: W0912 22:03:19.010068 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.010150 kubelet[2665]: E0912 22:03:19.010088 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.010263 kubelet[2665]: E0912 22:03:19.010248 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.010263 kubelet[2665]: W0912 22:03:19.010260 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.010310 kubelet[2665]: E0912 22:03:19.010268 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.010433 kubelet[2665]: E0912 22:03:19.010404 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.010433 kubelet[2665]: W0912 22:03:19.010415 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.010433 kubelet[2665]: E0912 22:03:19.010424 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.010684 kubelet[2665]: E0912 22:03:19.010637 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.010684 kubelet[2665]: W0912 22:03:19.010648 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.010684 kubelet[2665]: E0912 22:03:19.010655 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.010851 kubelet[2665]: E0912 22:03:19.010834 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.010851 kubelet[2665]: W0912 22:03:19.010845 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.010851 kubelet[2665]: E0912 22:03:19.010853 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011017 kubelet[2665]: E0912 22:03:19.010998 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011017 kubelet[2665]: W0912 22:03:19.011008 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011017 kubelet[2665]: E0912 22:03:19.011016 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011162 kubelet[2665]: E0912 22:03:19.011147 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011162 kubelet[2665]: W0912 22:03:19.011160 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011209 kubelet[2665]: E0912 22:03:19.011169 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011312 kubelet[2665]: E0912 22:03:19.011301 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011339 kubelet[2665]: W0912 22:03:19.011312 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011339 kubelet[2665]: E0912 22:03:19.011319 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011472 kubelet[2665]: E0912 22:03:19.011461 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011495 kubelet[2665]: W0912 22:03:19.011471 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011495 kubelet[2665]: E0912 22:03:19.011480 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011617 kubelet[2665]: E0912 22:03:19.011607 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011639 kubelet[2665]: W0912 22:03:19.011616 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011639 kubelet[2665]: E0912 22:03:19.011624 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011774 kubelet[2665]: E0912 22:03:19.011761 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011800 kubelet[2665]: W0912 22:03:19.011774 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011800 kubelet[2665]: E0912 22:03:19.011783 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.011987 kubelet[2665]: E0912 22:03:19.011948 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.011987 kubelet[2665]: W0912 22:03:19.011960 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.011987 kubelet[2665]: E0912 22:03:19.011969 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.012157 kubelet[2665]: E0912 22:03:19.012141 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.012157 kubelet[2665]: W0912 22:03:19.012152 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.012204 kubelet[2665]: E0912 22:03:19.012161 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.013016 kubelet[2665]: E0912 22:03:19.012317 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.013016 kubelet[2665]: W0912 22:03:19.012327 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.013016 kubelet[2665]: E0912 22:03:19.012334 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.013016 kubelet[2665]: E0912 22:03:19.012520 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.013016 kubelet[2665]: W0912 22:03:19.012533 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.013016 kubelet[2665]: E0912 22:03:19.012542 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.014316 systemd[1]: Started cri-containerd-6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb.scope - libcontainer container 6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb. Sep 12 22:03:19.016224 kubelet[2665]: E0912 22:03:19.016199 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.016224 kubelet[2665]: W0912 22:03:19.016219 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.016333 kubelet[2665]: E0912 22:03:19.016250 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.016333 kubelet[2665]: E0912 22:03:19.016450 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.016333 kubelet[2665]: W0912 22:03:19.016460 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.016333 kubelet[2665]: E0912 22:03:19.016485 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.016692 kubelet[2665]: E0912 22:03:19.016674 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.016692 kubelet[2665]: W0912 22:03:19.016687 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.016752 kubelet[2665]: E0912 22:03:19.016701 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.017169 kubelet[2665]: E0912 22:03:19.017150 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.017440 kubelet[2665]: W0912 22:03:19.017401 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.017440 kubelet[2665]: E0912 22:03:19.017437 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.017938 kubelet[2665]: E0912 22:03:19.017914 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.018024 kubelet[2665]: W0912 22:03:19.017948 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.018024 kubelet[2665]: E0912 22:03:19.017992 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.018199 kubelet[2665]: E0912 22:03:19.018187 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.018199 kubelet[2665]: W0912 22:03:19.018198 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.018304 kubelet[2665]: E0912 22:03:19.018277 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.018949 kubelet[2665]: E0912 22:03:19.018927 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.018949 kubelet[2665]: W0912 22:03:19.018948 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.019916 kubelet[2665]: E0912 22:03:19.019341 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.020957 kubelet[2665]: E0912 22:03:19.020716 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.020957 kubelet[2665]: W0912 22:03:19.020946 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.021080 kubelet[2665]: E0912 22:03:19.021000 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.022221 kubelet[2665]: E0912 22:03:19.022203 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.022221 kubelet[2665]: W0912 22:03:19.022219 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.022377 kubelet[2665]: E0912 22:03:19.022286 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.022735 kubelet[2665]: E0912 22:03:19.022712 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.022834 kubelet[2665]: W0912 22:03:19.022728 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.022885 kubelet[2665]: E0912 22:03:19.022861 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.023073 kubelet[2665]: E0912 22:03:19.023055 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.023138 kubelet[2665]: W0912 22:03:19.023068 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.023138 kubelet[2665]: E0912 22:03:19.023131 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.023547 kubelet[2665]: E0912 22:03:19.023297 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.023547 kubelet[2665]: W0912 22:03:19.023306 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.023547 kubelet[2665]: E0912 22:03:19.023318 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.023688 kubelet[2665]: E0912 22:03:19.023675 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.023834 kubelet[2665]: W0912 22:03:19.023732 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.023834 kubelet[2665]: E0912 22:03:19.023767 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.024074 kubelet[2665]: E0912 22:03:19.024060 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.024150 kubelet[2665]: W0912 22:03:19.024138 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.024217 kubelet[2665]: E0912 22:03:19.024205 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.024717 kubelet[2665]: E0912 22:03:19.024675 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.024717 kubelet[2665]: W0912 22:03:19.024714 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.024832 kubelet[2665]: E0912 22:03:19.024733 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.025277 kubelet[2665]: E0912 22:03:19.025182 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.025277 kubelet[2665]: W0912 22:03:19.025198 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.025277 kubelet[2665]: E0912 22:03:19.025218 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.025525 kubelet[2665]: E0912 22:03:19.025512 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.025656 kubelet[2665]: W0912 22:03:19.025578 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.025656 kubelet[2665]: E0912 22:03:19.025606 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.025903 kubelet[2665]: E0912 22:03:19.025890 2665 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:03:19.025963 kubelet[2665]: W0912 22:03:19.025953 2665 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:03:19.026051 kubelet[2665]: E0912 22:03:19.026034 2665 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:03:19.058932 containerd[1535]: time="2025-09-12T22:03:19.058893690Z" level=info msg="StartContainer for \"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\" returns successfully" Sep 12 22:03:19.072669 systemd[1]: cri-containerd-6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb.scope: Deactivated successfully. Sep 12 22:03:19.073043 systemd[1]: cri-containerd-6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb.scope: Consumed 30ms CPU time, 6.3M memory peak, 4.1M written to disk. Sep 12 22:03:19.085895 containerd[1535]: time="2025-09-12T22:03:19.085685759Z" level=info msg="received exit event container_id:\"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\" id:\"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\" pid:3352 exited_at:{seconds:1757714599 nanos:81488088}" Sep 12 22:03:19.085895 containerd[1535]: time="2025-09-12T22:03:19.085808092Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\" id:\"6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb\" pid:3352 exited_at:{seconds:1757714599 nanos:81488088}" Sep 12 22:03:19.122113 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6a588f986a03c191915da5b459c3f420b58def49d6a43f0f28687901ed6fa9eb-rootfs.mount: Deactivated successfully. Sep 12 22:03:19.934193 kubelet[2665]: I0912 22:03:19.934144 2665 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:03:19.935480 kubelet[2665]: E0912 22:03:19.934686 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:19.939281 containerd[1535]: time="2025-09-12T22:03:19.939248136Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 12 22:03:20.862571 kubelet[2665]: E0912 22:03:20.862519 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wdhc9" podUID="9741ef86-7513-4899-85e2-cf42c12fa232" Sep 12 22:03:22.350838 containerd[1535]: time="2025-09-12T22:03:22.350785711Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:22.351573 containerd[1535]: time="2025-09-12T22:03:22.351524098Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=65913477" Sep 12 22:03:22.352202 containerd[1535]: time="2025-09-12T22:03:22.352168316Z" level=info msg="ImageCreate event name:\"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:22.355366 containerd[1535]: time="2025-09-12T22:03:22.355317160Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:22.356531 containerd[1535]: time="2025-09-12T22:03:22.356502427Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"67282718\" in 2.417213527s" Sep 12 22:03:22.356608 containerd[1535]: time="2025-09-12T22:03:22.356534030Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:7077a1dc632ee598cbfa626f9e3c9bca5b20c0d1e1e557995890125b2e8d2e23\"" Sep 12 22:03:22.358915 containerd[1535]: time="2025-09-12T22:03:22.358868400Z" level=info msg="CreateContainer within sandbox \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 12 22:03:22.368022 containerd[1535]: time="2025-09-12T22:03:22.366431523Z" level=info msg="Container b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:22.398118 containerd[1535]: time="2025-09-12T22:03:22.398055695Z" level=info msg="CreateContainer within sandbox \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\"" Sep 12 22:03:22.398937 containerd[1535]: time="2025-09-12T22:03:22.398905371Z" level=info msg="StartContainer for \"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\"" Sep 12 22:03:22.400560 containerd[1535]: time="2025-09-12T22:03:22.400526597Z" level=info msg="connecting to shim b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7" address="unix:///run/containerd/s/a684c19efef5e689a6e07cc6551800318417f9668c42a95a72d09d4f66c733be" protocol=ttrpc version=3 Sep 12 22:03:22.421147 systemd[1]: Started cri-containerd-b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7.scope - libcontainer container b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7. Sep 12 22:03:22.461245 containerd[1535]: time="2025-09-12T22:03:22.461190548Z" level=info msg="StartContainer for \"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\" returns successfully" Sep 12 22:03:22.863001 kubelet[2665]: E0912 22:03:22.862747 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-wdhc9" podUID="9741ef86-7513-4899-85e2-cf42c12fa232" Sep 12 22:03:23.036578 systemd[1]: cri-containerd-b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7.scope: Deactivated successfully. Sep 12 22:03:23.036852 systemd[1]: cri-containerd-b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7.scope: Consumed 460ms CPU time, 178M memory peak, 3.7M read from disk, 165.8M written to disk. Sep 12 22:03:23.050573 kubelet[2665]: I0912 22:03:23.050523 2665 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 12 22:03:23.057651 containerd[1535]: time="2025-09-12T22:03:23.057599850Z" level=info msg="received exit event container_id:\"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\" id:\"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\" pid:3425 exited_at:{seconds:1757714603 nanos:57365270}" Sep 12 22:03:23.057774 containerd[1535]: time="2025-09-12T22:03:23.057725421Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\" id:\"b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7\" pid:3425 exited_at:{seconds:1757714603 nanos:57365270}" Sep 12 22:03:23.102236 systemd[1]: Created slice kubepods-burstable-pod9c729210_e125_4923_8a29_db7b0068bea7.slice - libcontainer container kubepods-burstable-pod9c729210_e125_4923_8a29_db7b0068bea7.slice. Sep 12 22:03:23.118247 systemd[1]: Created slice kubepods-burstable-pod41f762f2_5fa7_4a7a_b5c8_964bf56dfbee.slice - libcontainer container kubepods-burstable-pod41f762f2_5fa7_4a7a_b5c8_964bf56dfbee.slice. Sep 12 22:03:23.121535 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b88947ea9e4ebd1a213a3f49f904cb3d9acb4d05c5bd8ae75cdfb2a399a209e7-rootfs.mount: Deactivated successfully. Sep 12 22:03:23.127913 systemd[1]: Created slice kubepods-besteffort-pod04200bdf_66fa_47d4_acdd_9b4679fe0b2b.slice - libcontainer container kubepods-besteffort-pod04200bdf_66fa_47d4_acdd_9b4679fe0b2b.slice. Sep 12 22:03:23.134674 systemd[1]: Created slice kubepods-besteffort-pod3cdb4f52_c9af_495e_a27e_8dd84094cf42.slice - libcontainer container kubepods-besteffort-pod3cdb4f52_c9af_495e_a27e_8dd84094cf42.slice. Sep 12 22:03:23.140273 systemd[1]: Created slice kubepods-besteffort-poda69477ef_6c97_40b7_8dbb_52ed876ed57d.slice - libcontainer container kubepods-besteffort-poda69477ef_6c97_40b7_8dbb_52ed876ed57d.slice. Sep 12 22:03:23.149966 systemd[1]: Created slice kubepods-besteffort-pod3b21c912_c86c_4756_bba2_0197b8ea9324.slice - libcontainer container kubepods-besteffort-pod3b21c912_c86c_4756_bba2_0197b8ea9324.slice. Sep 12 22:03:23.181739 systemd[1]: Created slice kubepods-besteffort-pod5f1bb2b2_f6f1_4e4b_a7a3_ccf525f155d5.slice - libcontainer container kubepods-besteffort-pod5f1bb2b2_f6f1_4e4b_a7a3_ccf525f155d5.slice. Sep 12 22:03:23.251562 kubelet[2665]: I0912 22:03:23.251510 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdt2b\" (UniqueName: \"kubernetes.io/projected/9c729210-e125-4923-8a29-db7b0068bea7-kube-api-access-jdt2b\") pod \"coredns-7c65d6cfc9-2v4dd\" (UID: \"9c729210-e125-4923-8a29-db7b0068bea7\") " pod="kube-system/coredns-7c65d6cfc9-2v4dd" Sep 12 22:03:23.252008 kubelet[2665]: I0912 22:03:23.251828 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41f762f2-5fa7-4a7a-b5c8-964bf56dfbee-config-volume\") pod \"coredns-7c65d6cfc9-9cgtm\" (UID: \"41f762f2-5fa7-4a7a-b5c8-964bf56dfbee\") " pod="kube-system/coredns-7c65d6cfc9-9cgtm" Sep 12 22:03:23.252008 kubelet[2665]: I0912 22:03:23.251862 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7n28\" (UniqueName: \"kubernetes.io/projected/3b21c912-c86c-4756-bba2-0197b8ea9324-kube-api-access-n7n28\") pod \"whisker-5b7f94cf94-nf85x\" (UID: \"3b21c912-c86c-4756-bba2-0197b8ea9324\") " pod="calico-system/whisker-5b7f94cf94-nf85x" Sep 12 22:03:23.252008 kubelet[2665]: I0912 22:03:23.251878 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spt7s\" (UniqueName: \"kubernetes.io/projected/a69477ef-6c97-40b7-8dbb-52ed876ed57d-kube-api-access-spt7s\") pod \"goldmane-7988f88666-l6kcq\" (UID: \"a69477ef-6c97-40b7-8dbb-52ed876ed57d\") " pod="calico-system/goldmane-7988f88666-l6kcq" Sep 12 22:03:23.252008 kubelet[2665]: I0912 22:03:23.251894 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c729210-e125-4923-8a29-db7b0068bea7-config-volume\") pod \"coredns-7c65d6cfc9-2v4dd\" (UID: \"9c729210-e125-4923-8a29-db7b0068bea7\") " pod="kube-system/coredns-7c65d6cfc9-2v4dd" Sep 12 22:03:23.252008 kubelet[2665]: I0912 22:03:23.251912 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04200bdf-66fa-47d4-acdd-9b4679fe0b2b-tigera-ca-bundle\") pod \"calico-kube-controllers-86d86dd9d9-dgcmg\" (UID: \"04200bdf-66fa-47d4-acdd-9b4679fe0b2b\") " pod="calico-system/calico-kube-controllers-86d86dd9d9-dgcmg" Sep 12 22:03:23.252155 kubelet[2665]: I0912 22:03:23.251929 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-ca-bundle\") pod \"whisker-5b7f94cf94-nf85x\" (UID: \"3b21c912-c86c-4756-bba2-0197b8ea9324\") " pod="calico-system/whisker-5b7f94cf94-nf85x" Sep 12 22:03:23.252155 kubelet[2665]: I0912 22:03:23.251953 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3cdb4f52-c9af-495e-a27e-8dd84094cf42-calico-apiserver-certs\") pod \"calico-apiserver-57d459fd7c-7mhnw\" (UID: \"3cdb4f52-c9af-495e-a27e-8dd84094cf42\") " pod="calico-apiserver/calico-apiserver-57d459fd7c-7mhnw" Sep 12 22:03:23.252436 kubelet[2665]: I0912 22:03:23.252210 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a69477ef-6c97-40b7-8dbb-52ed876ed57d-goldmane-ca-bundle\") pod \"goldmane-7988f88666-l6kcq\" (UID: \"a69477ef-6c97-40b7-8dbb-52ed876ed57d\") " pod="calico-system/goldmane-7988f88666-l6kcq" Sep 12 22:03:23.252436 kubelet[2665]: I0912 22:03:23.252252 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5-calico-apiserver-certs\") pod \"calico-apiserver-57d459fd7c-n9cxf\" (UID: \"5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5\") " pod="calico-apiserver/calico-apiserver-57d459fd7c-n9cxf" Sep 12 22:03:23.252436 kubelet[2665]: I0912 22:03:23.252276 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a69477ef-6c97-40b7-8dbb-52ed876ed57d-config\") pod \"goldmane-7988f88666-l6kcq\" (UID: \"a69477ef-6c97-40b7-8dbb-52ed876ed57d\") " pod="calico-system/goldmane-7988f88666-l6kcq" Sep 12 22:03:23.252436 kubelet[2665]: I0912 22:03:23.252293 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/a69477ef-6c97-40b7-8dbb-52ed876ed57d-goldmane-key-pair\") pod \"goldmane-7988f88666-l6kcq\" (UID: \"a69477ef-6c97-40b7-8dbb-52ed876ed57d\") " pod="calico-system/goldmane-7988f88666-l6kcq" Sep 12 22:03:23.252436 kubelet[2665]: I0912 22:03:23.252310 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrdwh\" (UniqueName: \"kubernetes.io/projected/3cdb4f52-c9af-495e-a27e-8dd84094cf42-kube-api-access-qrdwh\") pod \"calico-apiserver-57d459fd7c-7mhnw\" (UID: \"3cdb4f52-c9af-495e-a27e-8dd84094cf42\") " pod="calico-apiserver/calico-apiserver-57d459fd7c-7mhnw" Sep 12 22:03:23.252565 kubelet[2665]: I0912 22:03:23.252332 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-backend-key-pair\") pod \"whisker-5b7f94cf94-nf85x\" (UID: \"3b21c912-c86c-4756-bba2-0197b8ea9324\") " pod="calico-system/whisker-5b7f94cf94-nf85x" Sep 12 22:03:23.252565 kubelet[2665]: I0912 22:03:23.252348 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfxqg\" (UniqueName: \"kubernetes.io/projected/5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5-kube-api-access-vfxqg\") pod \"calico-apiserver-57d459fd7c-n9cxf\" (UID: \"5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5\") " pod="calico-apiserver/calico-apiserver-57d459fd7c-n9cxf" Sep 12 22:03:23.252565 kubelet[2665]: I0912 22:03:23.252364 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9f2c\" (UniqueName: \"kubernetes.io/projected/04200bdf-66fa-47d4-acdd-9b4679fe0b2b-kube-api-access-r9f2c\") pod \"calico-kube-controllers-86d86dd9d9-dgcmg\" (UID: \"04200bdf-66fa-47d4-acdd-9b4679fe0b2b\") " pod="calico-system/calico-kube-controllers-86d86dd9d9-dgcmg" Sep 12 22:03:23.252565 kubelet[2665]: I0912 22:03:23.252379 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5xgv\" (UniqueName: \"kubernetes.io/projected/41f762f2-5fa7-4a7a-b5c8-964bf56dfbee-kube-api-access-b5xgv\") pod \"coredns-7c65d6cfc9-9cgtm\" (UID: \"41f762f2-5fa7-4a7a-b5c8-964bf56dfbee\") " pod="kube-system/coredns-7c65d6cfc9-9cgtm" Sep 12 22:03:23.411921 kubelet[2665]: E0912 22:03:23.411797 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:23.413354 containerd[1535]: time="2025-09-12T22:03:23.413293357Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-2v4dd,Uid:9c729210-e125-4923-8a29-db7b0068bea7,Namespace:kube-system,Attempt:0,}" Sep 12 22:03:23.427110 kubelet[2665]: E0912 22:03:23.427060 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:23.430198 containerd[1535]: time="2025-09-12T22:03:23.430091851Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-9cgtm,Uid:41f762f2-5fa7-4a7a-b5c8-964bf56dfbee,Namespace:kube-system,Attempt:0,}" Sep 12 22:03:23.443560 containerd[1535]: time="2025-09-12T22:03:23.443512252Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-7mhnw,Uid:3cdb4f52-c9af-495e-a27e-8dd84094cf42,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:03:23.443729 containerd[1535]: time="2025-09-12T22:03:23.443703789Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-86d86dd9d9-dgcmg,Uid:04200bdf-66fa-47d4-acdd-9b4679fe0b2b,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:23.445121 containerd[1535]: time="2025-09-12T22:03:23.445061666Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-l6kcq,Uid:a69477ef-6c97-40b7-8dbb-52ed876ed57d,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:23.454429 containerd[1535]: time="2025-09-12T22:03:23.454258902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5b7f94cf94-nf85x,Uid:3b21c912-c86c-4756-bba2-0197b8ea9324,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:23.485237 containerd[1535]: time="2025-09-12T22:03:23.485195660Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-n9cxf,Uid:5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:03:23.536302 containerd[1535]: time="2025-09-12T22:03:23.536250559Z" level=error msg="Failed to destroy network for sandbox \"e928003cf4f49191a028b72044dbfafead7ee45a29e53f29ab2193c6ebbf1553\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.538371 containerd[1535]: time="2025-09-12T22:03:23.538320938Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-9cgtm,Uid:41f762f2-5fa7-4a7a-b5c8-964bf56dfbee,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e928003cf4f49191a028b72044dbfafead7ee45a29e53f29ab2193c6ebbf1553\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.538643 kubelet[2665]: E0912 22:03:23.538590 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e928003cf4f49191a028b72044dbfafead7ee45a29e53f29ab2193c6ebbf1553\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.538709 kubelet[2665]: E0912 22:03:23.538672 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e928003cf4f49191a028b72044dbfafead7ee45a29e53f29ab2193c6ebbf1553\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-9cgtm" Sep 12 22:03:23.538709 kubelet[2665]: E0912 22:03:23.538693 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e928003cf4f49191a028b72044dbfafead7ee45a29e53f29ab2193c6ebbf1553\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-9cgtm" Sep 12 22:03:23.538769 kubelet[2665]: E0912 22:03:23.538742 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-9cgtm_kube-system(41f762f2-5fa7-4a7a-b5c8-964bf56dfbee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-9cgtm_kube-system(41f762f2-5fa7-4a7a-b5c8-964bf56dfbee)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e928003cf4f49191a028b72044dbfafead7ee45a29e53f29ab2193c6ebbf1553\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-9cgtm" podUID="41f762f2-5fa7-4a7a-b5c8-964bf56dfbee" Sep 12 22:03:23.549220 containerd[1535]: time="2025-09-12T22:03:23.549160436Z" level=error msg="Failed to destroy network for sandbox \"eab49012891971bbd66e95c012fc8cde524e4db8f5c58d0f7f4f5e8facbef935\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.550719 containerd[1535]: time="2025-09-12T22:03:23.550597961Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-7mhnw,Uid:3cdb4f52-c9af-495e-a27e-8dd84094cf42,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"eab49012891971bbd66e95c012fc8cde524e4db8f5c58d0f7f4f5e8facbef935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.550883 kubelet[2665]: E0912 22:03:23.550844 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eab49012891971bbd66e95c012fc8cde524e4db8f5c58d0f7f4f5e8facbef935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.550939 kubelet[2665]: E0912 22:03:23.550905 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eab49012891971bbd66e95c012fc8cde524e4db8f5c58d0f7f4f5e8facbef935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-57d459fd7c-7mhnw" Sep 12 22:03:23.550939 kubelet[2665]: E0912 22:03:23.550928 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eab49012891971bbd66e95c012fc8cde524e4db8f5c58d0f7f4f5e8facbef935\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-57d459fd7c-7mhnw" Sep 12 22:03:23.551233 kubelet[2665]: E0912 22:03:23.550982 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-57d459fd7c-7mhnw_calico-apiserver(3cdb4f52-c9af-495e-a27e-8dd84094cf42)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-57d459fd7c-7mhnw_calico-apiserver(3cdb4f52-c9af-495e-a27e-8dd84094cf42)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"eab49012891971bbd66e95c012fc8cde524e4db8f5c58d0f7f4f5e8facbef935\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-57d459fd7c-7mhnw" podUID="3cdb4f52-c9af-495e-a27e-8dd84094cf42" Sep 12 22:03:23.559611 containerd[1535]: time="2025-09-12T22:03:23.559544975Z" level=error msg="Failed to destroy network for sandbox \"c105053a2e7ea83d47f9bb29bc065cfb469e76aed61ec9724faaf89ae411de74\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.562400 containerd[1535]: time="2025-09-12T22:03:23.562279172Z" level=error msg="Failed to destroy network for sandbox \"e81b73729c9a51fb953a4f41586d869a29362cde8eb56e680bd1ec3d586e58ff\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.562522 containerd[1535]: time="2025-09-12T22:03:23.562483990Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-86d86dd9d9-dgcmg,Uid:04200bdf-66fa-47d4-acdd-9b4679fe0b2b,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c105053a2e7ea83d47f9bb29bc065cfb469e76aed61ec9724faaf89ae411de74\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.562748 kubelet[2665]: E0912 22:03:23.562711 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c105053a2e7ea83d47f9bb29bc065cfb469e76aed61ec9724faaf89ae411de74\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.562805 kubelet[2665]: E0912 22:03:23.562768 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c105053a2e7ea83d47f9bb29bc065cfb469e76aed61ec9724faaf89ae411de74\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-86d86dd9d9-dgcmg" Sep 12 22:03:23.562805 kubelet[2665]: E0912 22:03:23.562789 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c105053a2e7ea83d47f9bb29bc065cfb469e76aed61ec9724faaf89ae411de74\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-86d86dd9d9-dgcmg" Sep 12 22:03:23.562851 kubelet[2665]: E0912 22:03:23.562829 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-86d86dd9d9-dgcmg_calico-system(04200bdf-66fa-47d4-acdd-9b4679fe0b2b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-86d86dd9d9-dgcmg_calico-system(04200bdf-66fa-47d4-acdd-9b4679fe0b2b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c105053a2e7ea83d47f9bb29bc065cfb469e76aed61ec9724faaf89ae411de74\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-86d86dd9d9-dgcmg" podUID="04200bdf-66fa-47d4-acdd-9b4679fe0b2b" Sep 12 22:03:23.565170 containerd[1535]: time="2025-09-12T22:03:23.565049092Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-2v4dd,Uid:9c729210-e125-4923-8a29-db7b0068bea7,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e81b73729c9a51fb953a4f41586d869a29362cde8eb56e680bd1ec3d586e58ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.565478 kubelet[2665]: E0912 22:03:23.565443 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e81b73729c9a51fb953a4f41586d869a29362cde8eb56e680bd1ec3d586e58ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.565478 kubelet[2665]: E0912 22:03:23.565496 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e81b73729c9a51fb953a4f41586d869a29362cde8eb56e680bd1ec3d586e58ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-2v4dd" Sep 12 22:03:23.566067 kubelet[2665]: E0912 22:03:23.565520 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e81b73729c9a51fb953a4f41586d869a29362cde8eb56e680bd1ec3d586e58ff\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-2v4dd" Sep 12 22:03:23.566067 kubelet[2665]: E0912 22:03:23.565740 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-2v4dd_kube-system(9c729210-e125-4923-8a29-db7b0068bea7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-2v4dd_kube-system(9c729210-e125-4923-8a29-db7b0068bea7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e81b73729c9a51fb953a4f41586d869a29362cde8eb56e680bd1ec3d586e58ff\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-2v4dd" podUID="9c729210-e125-4923-8a29-db7b0068bea7" Sep 12 22:03:23.572969 containerd[1535]: time="2025-09-12T22:03:23.572913972Z" level=error msg="Failed to destroy network for sandbox \"32c5e87378335021771b0e5172507a967edb6de3a30d23f0f1380904c9c17677\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.574019 containerd[1535]: time="2025-09-12T22:03:23.573914219Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-l6kcq,Uid:a69477ef-6c97-40b7-8dbb-52ed876ed57d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"32c5e87378335021771b0e5172507a967edb6de3a30d23f0f1380904c9c17677\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.574771 kubelet[2665]: E0912 22:03:23.574721 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"32c5e87378335021771b0e5172507a967edb6de3a30d23f0f1380904c9c17677\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.574825 kubelet[2665]: E0912 22:03:23.574783 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"32c5e87378335021771b0e5172507a967edb6de3a30d23f0f1380904c9c17677\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7988f88666-l6kcq" Sep 12 22:03:23.574825 kubelet[2665]: E0912 22:03:23.574807 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"32c5e87378335021771b0e5172507a967edb6de3a30d23f0f1380904c9c17677\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7988f88666-l6kcq" Sep 12 22:03:23.574881 kubelet[2665]: E0912 22:03:23.574852 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7988f88666-l6kcq_calico-system(a69477ef-6c97-40b7-8dbb-52ed876ed57d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7988f88666-l6kcq_calico-system(a69477ef-6c97-40b7-8dbb-52ed876ed57d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"32c5e87378335021771b0e5172507a967edb6de3a30d23f0f1380904c9c17677\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7988f88666-l6kcq" podUID="a69477ef-6c97-40b7-8dbb-52ed876ed57d" Sep 12 22:03:23.577306 containerd[1535]: time="2025-09-12T22:03:23.577262749Z" level=error msg="Failed to destroy network for sandbox \"ae8d7c8d8c33581322426527deb0b1153b6744f1b222e30828d25249c857e186\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.578831 containerd[1535]: time="2025-09-12T22:03:23.578771079Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5b7f94cf94-nf85x,Uid:3b21c912-c86c-4756-bba2-0197b8ea9324,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae8d7c8d8c33581322426527deb0b1153b6744f1b222e30828d25249c857e186\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.579068 kubelet[2665]: E0912 22:03:23.579038 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae8d7c8d8c33581322426527deb0b1153b6744f1b222e30828d25249c857e186\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.579119 kubelet[2665]: E0912 22:03:23.579088 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae8d7c8d8c33581322426527deb0b1153b6744f1b222e30828d25249c857e186\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5b7f94cf94-nf85x" Sep 12 22:03:23.579119 kubelet[2665]: E0912 22:03:23.579106 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae8d7c8d8c33581322426527deb0b1153b6744f1b222e30828d25249c857e186\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5b7f94cf94-nf85x" Sep 12 22:03:23.579165 kubelet[2665]: E0912 22:03:23.579145 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-5b7f94cf94-nf85x_calico-system(3b21c912-c86c-4756-bba2-0197b8ea9324)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-5b7f94cf94-nf85x_calico-system(3b21c912-c86c-4756-bba2-0197b8ea9324)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ae8d7c8d8c33581322426527deb0b1153b6744f1b222e30828d25249c857e186\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5b7f94cf94-nf85x" podUID="3b21c912-c86c-4756-bba2-0197b8ea9324" Sep 12 22:03:23.585548 containerd[1535]: time="2025-09-12T22:03:23.585444097Z" level=error msg="Failed to destroy network for sandbox \"5c7bcafe1d5a167663f505157985f39968a1c89f028e0dd04266e9ced62e7891\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.586523 containerd[1535]: time="2025-09-12T22:03:23.586490027Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-n9cxf,Uid:5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c7bcafe1d5a167663f505157985f39968a1c89f028e0dd04266e9ced62e7891\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.586804 kubelet[2665]: E0912 22:03:23.586773 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c7bcafe1d5a167663f505157985f39968a1c89f028e0dd04266e9ced62e7891\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:23.587011 kubelet[2665]: E0912 22:03:23.586897 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c7bcafe1d5a167663f505157985f39968a1c89f028e0dd04266e9ced62e7891\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-57d459fd7c-n9cxf" Sep 12 22:03:23.587011 kubelet[2665]: E0912 22:03:23.586918 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c7bcafe1d5a167663f505157985f39968a1c89f028e0dd04266e9ced62e7891\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-57d459fd7c-n9cxf" Sep 12 22:03:23.587011 kubelet[2665]: E0912 22:03:23.586962 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-57d459fd7c-n9cxf_calico-apiserver(5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-57d459fd7c-n9cxf_calico-apiserver(5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5c7bcafe1d5a167663f505157985f39968a1c89f028e0dd04266e9ced62e7891\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-57d459fd7c-n9cxf" podUID="5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5" Sep 12 22:03:23.961072 containerd[1535]: time="2025-09-12T22:03:23.960961559Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 12 22:03:24.367888 systemd[1]: run-netns-cni\x2d9654ca9f\x2d0dcb\x2db6bd\x2d95d1\x2dfa99ae22e542.mount: Deactivated successfully. Sep 12 22:03:24.367996 systemd[1]: run-netns-cni\x2d89e82cb9\x2dd73b\x2db0e9\x2d877d\x2dcb3e4a1f6ec2.mount: Deactivated successfully. Sep 12 22:03:24.868624 systemd[1]: Created slice kubepods-besteffort-pod9741ef86_7513_4899_85e2_cf42c12fa232.slice - libcontainer container kubepods-besteffort-pod9741ef86_7513_4899_85e2_cf42c12fa232.slice. Sep 12 22:03:24.878679 containerd[1535]: time="2025-09-12T22:03:24.878632117Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wdhc9,Uid:9741ef86-7513-4899-85e2-cf42c12fa232,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:24.948011 containerd[1535]: time="2025-09-12T22:03:24.947568569Z" level=error msg="Failed to destroy network for sandbox \"3532f74739f197a7eb65c0055fe4290a8654504aaaa23e2af67fcd1b9481b5a9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:24.948925 containerd[1535]: time="2025-09-12T22:03:24.948808592Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wdhc9,Uid:9741ef86-7513-4899-85e2-cf42c12fa232,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3532f74739f197a7eb65c0055fe4290a8654504aaaa23e2af67fcd1b9481b5a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:24.949177 kubelet[2665]: E0912 22:03:24.949116 2665 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3532f74739f197a7eb65c0055fe4290a8654504aaaa23e2af67fcd1b9481b5a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:03:24.949442 kubelet[2665]: E0912 22:03:24.949181 2665 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3532f74739f197a7eb65c0055fe4290a8654504aaaa23e2af67fcd1b9481b5a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:24.949442 kubelet[2665]: E0912 22:03:24.949200 2665 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3532f74739f197a7eb65c0055fe4290a8654504aaaa23e2af67fcd1b9481b5a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-wdhc9" Sep 12 22:03:24.949442 kubelet[2665]: E0912 22:03:24.949247 2665 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-wdhc9_calico-system(9741ef86-7513-4899-85e2-cf42c12fa232)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-wdhc9_calico-system(9741ef86-7513-4899-85e2-cf42c12fa232)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3532f74739f197a7eb65c0055fe4290a8654504aaaa23e2af67fcd1b9481b5a9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-wdhc9" podUID="9741ef86-7513-4899-85e2-cf42c12fa232" Sep 12 22:03:24.949632 systemd[1]: run-netns-cni\x2d945e9e73\x2da1fb\x2d843d\x2d96f9\x2d66a71dc437bb.mount: Deactivated successfully. Sep 12 22:03:27.045125 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4039505702.mount: Deactivated successfully. Sep 12 22:03:27.261625 containerd[1535]: time="2025-09-12T22:03:27.261566530Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:27.270864 containerd[1535]: time="2025-09-12T22:03:27.262011403Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=151100457" Sep 12 22:03:27.270921 containerd[1535]: time="2025-09-12T22:03:27.263027679Z" level=info msg="ImageCreate event name:\"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:27.270921 containerd[1535]: time="2025-09-12T22:03:27.265036828Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"151100319\" in 3.303966619s" Sep 12 22:03:27.270921 containerd[1535]: time="2025-09-12T22:03:27.270916624Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:2b8abd2140fc4464ed664d225fe38e5b90bbfcf62996b484b0fc0e0537b6a4a9\"" Sep 12 22:03:27.271615 containerd[1535]: time="2025-09-12T22:03:27.271568352Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:27.279510 containerd[1535]: time="2025-09-12T22:03:27.279469618Z" level=info msg="CreateContainer within sandbox \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 12 22:03:27.293134 containerd[1535]: time="2025-09-12T22:03:27.292137917Z" level=info msg="Container 1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:27.307832 containerd[1535]: time="2025-09-12T22:03:27.307724633Z" level=info msg="CreateContainer within sandbox \"1d11199753ed4f9932479eb5ec8ee9c067ae60f9fbaea9cc9798940e24344e79\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f\"" Sep 12 22:03:27.308301 containerd[1535]: time="2025-09-12T22:03:27.308263433Z" level=info msg="StartContainer for \"1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f\"" Sep 12 22:03:27.309840 containerd[1535]: time="2025-09-12T22:03:27.309809348Z" level=info msg="connecting to shim 1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f" address="unix:///run/containerd/s/a684c19efef5e689a6e07cc6551800318417f9668c42a95a72d09d4f66c733be" protocol=ttrpc version=3 Sep 12 22:03:27.329142 systemd[1]: Started cri-containerd-1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f.scope - libcontainer container 1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f. Sep 12 22:03:27.363006 containerd[1535]: time="2025-09-12T22:03:27.362159550Z" level=info msg="StartContainer for \"1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f\" returns successfully" Sep 12 22:03:27.480406 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 12 22:03:27.480506 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 12 22:03:27.779844 kubelet[2665]: I0912 22:03:27.779757 2665 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-backend-key-pair\") pod \"3b21c912-c86c-4756-bba2-0197b8ea9324\" (UID: \"3b21c912-c86c-4756-bba2-0197b8ea9324\") " Sep 12 22:03:27.779844 kubelet[2665]: I0912 22:03:27.779803 2665 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-ca-bundle\") pod \"3b21c912-c86c-4756-bba2-0197b8ea9324\" (UID: \"3b21c912-c86c-4756-bba2-0197b8ea9324\") " Sep 12 22:03:27.779844 kubelet[2665]: I0912 22:03:27.779827 2665 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7n28\" (UniqueName: \"kubernetes.io/projected/3b21c912-c86c-4756-bba2-0197b8ea9324-kube-api-access-n7n28\") pod \"3b21c912-c86c-4756-bba2-0197b8ea9324\" (UID: \"3b21c912-c86c-4756-bba2-0197b8ea9324\") " Sep 12 22:03:27.784865 kubelet[2665]: I0912 22:03:27.784793 2665 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "3b21c912-c86c-4756-bba2-0197b8ea9324" (UID: "3b21c912-c86c-4756-bba2-0197b8ea9324"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 12 22:03:27.785617 kubelet[2665]: I0912 22:03:27.785566 2665 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b21c912-c86c-4756-bba2-0197b8ea9324-kube-api-access-n7n28" (OuterVolumeSpecName: "kube-api-access-n7n28") pod "3b21c912-c86c-4756-bba2-0197b8ea9324" (UID: "3b21c912-c86c-4756-bba2-0197b8ea9324"). InnerVolumeSpecName "kube-api-access-n7n28". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 12 22:03:27.786660 kubelet[2665]: I0912 22:03:27.786626 2665 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "3b21c912-c86c-4756-bba2-0197b8ea9324" (UID: "3b21c912-c86c-4756-bba2-0197b8ea9324"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 12 22:03:27.880523 kubelet[2665]: I0912 22:03:27.880473 2665 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Sep 12 22:03:27.880523 kubelet[2665]: I0912 22:03:27.880511 2665 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b21c912-c86c-4756-bba2-0197b8ea9324-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Sep 12 22:03:27.880523 kubelet[2665]: I0912 22:03:27.880521 2665 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7n28\" (UniqueName: \"kubernetes.io/projected/3b21c912-c86c-4756-bba2-0197b8ea9324-kube-api-access-n7n28\") on node \"localhost\" DevicePath \"\"" Sep 12 22:03:27.982275 systemd[1]: Removed slice kubepods-besteffort-pod3b21c912_c86c_4756_bba2_0197b8ea9324.slice - libcontainer container kubepods-besteffort-pod3b21c912_c86c_4756_bba2_0197b8ea9324.slice. Sep 12 22:03:28.011214 kubelet[2665]: I0912 22:03:28.011118 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-cjs4k" podStartSLOduration=1.607617367 podStartE2EDuration="12.011100605s" podCreationTimestamp="2025-09-12 22:03:16 +0000 UTC" firstStartedPulling="2025-09-12 22:03:16.87059006 +0000 UTC m=+22.095069535" lastFinishedPulling="2025-09-12 22:03:27.274073338 +0000 UTC m=+32.498552773" observedRunningTime="2025-09-12 22:03:28.002044198 +0000 UTC m=+33.226523713" watchObservedRunningTime="2025-09-12 22:03:28.011100605 +0000 UTC m=+33.235580080" Sep 12 22:03:28.046178 systemd[1]: var-lib-kubelet-pods-3b21c912\x2dc86c\x2d4756\x2dbba2\x2d0197b8ea9324-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dn7n28.mount: Deactivated successfully. Sep 12 22:03:28.046893 systemd[1]: var-lib-kubelet-pods-3b21c912\x2dc86c\x2d4756\x2dbba2\x2d0197b8ea9324-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 12 22:03:28.054684 systemd[1]: Created slice kubepods-besteffort-pod2161f718_2754_4dbb_bfe9_02688bae41fb.slice - libcontainer container kubepods-besteffort-pod2161f718_2754_4dbb_bfe9_02688bae41fb.slice. Sep 12 22:03:28.182079 kubelet[2665]: I0912 22:03:28.182001 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs8r4\" (UniqueName: \"kubernetes.io/projected/2161f718-2754-4dbb-bfe9-02688bae41fb-kube-api-access-qs8r4\") pod \"whisker-75456b67bf-f8psn\" (UID: \"2161f718-2754-4dbb-bfe9-02688bae41fb\") " pod="calico-system/whisker-75456b67bf-f8psn" Sep 12 22:03:28.182079 kubelet[2665]: I0912 22:03:28.182059 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/2161f718-2754-4dbb-bfe9-02688bae41fb-whisker-backend-key-pair\") pod \"whisker-75456b67bf-f8psn\" (UID: \"2161f718-2754-4dbb-bfe9-02688bae41fb\") " pod="calico-system/whisker-75456b67bf-f8psn" Sep 12 22:03:28.182079 kubelet[2665]: I0912 22:03:28.182090 2665 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2161f718-2754-4dbb-bfe9-02688bae41fb-whisker-ca-bundle\") pod \"whisker-75456b67bf-f8psn\" (UID: \"2161f718-2754-4dbb-bfe9-02688bae41fb\") " pod="calico-system/whisker-75456b67bf-f8psn" Sep 12 22:03:28.364478 containerd[1535]: time="2025-09-12T22:03:28.364176499Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-75456b67bf-f8psn,Uid:2161f718-2754-4dbb-bfe9-02688bae41fb,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:28.510532 systemd-networkd[1435]: calie220b73daf3: Link UP Sep 12 22:03:28.510830 systemd-networkd[1435]: calie220b73daf3: Gained carrier Sep 12 22:03:28.523524 containerd[1535]: 2025-09-12 22:03:28.383 [INFO][3809] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 12 22:03:28.523524 containerd[1535]: 2025-09-12 22:03:28.411 [INFO][3809] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--75456b67bf--f8psn-eth0 whisker-75456b67bf- calico-system 2161f718-2754-4dbb-bfe9-02688bae41fb 915 0 2025-09-12 22:03:28 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:75456b67bf projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-75456b67bf-f8psn eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calie220b73daf3 [] [] }} ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-" Sep 12 22:03:28.523524 containerd[1535]: 2025-09-12 22:03:28.411 [INFO][3809] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.523524 containerd[1535]: 2025-09-12 22:03:28.466 [INFO][3825] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" HandleID="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Workload="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.466 [INFO][3825] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" HandleID="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Workload="localhost-k8s-whisker--75456b67bf--f8psn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c5e0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-75456b67bf-f8psn", "timestamp":"2025-09-12 22:03:28.466035985 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.466 [INFO][3825] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.466 [INFO][3825] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.466 [INFO][3825] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.476 [INFO][3825] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" host="localhost" Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.481 [INFO][3825] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.486 [INFO][3825] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.488 [INFO][3825] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.491 [INFO][3825] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:28.523741 containerd[1535]: 2025-09-12 22:03:28.491 [INFO][3825] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" host="localhost" Sep 12 22:03:28.523937 containerd[1535]: 2025-09-12 22:03:28.493 [INFO][3825] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437 Sep 12 22:03:28.523937 containerd[1535]: 2025-09-12 22:03:28.497 [INFO][3825] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" host="localhost" Sep 12 22:03:28.523937 containerd[1535]: 2025-09-12 22:03:28.502 [INFO][3825] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" host="localhost" Sep 12 22:03:28.523937 containerd[1535]: 2025-09-12 22:03:28.502 [INFO][3825] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" host="localhost" Sep 12 22:03:28.523937 containerd[1535]: 2025-09-12 22:03:28.502 [INFO][3825] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:28.523937 containerd[1535]: 2025-09-12 22:03:28.502 [INFO][3825] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" HandleID="k8s-pod-network.890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Workload="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.524074 containerd[1535]: 2025-09-12 22:03:28.504 [INFO][3809] cni-plugin/k8s.go 418: Populated endpoint ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--75456b67bf--f8psn-eth0", GenerateName:"whisker-75456b67bf-", Namespace:"calico-system", SelfLink:"", UID:"2161f718-2754-4dbb-bfe9-02688bae41fb", ResourceVersion:"915", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"75456b67bf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-75456b67bf-f8psn", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calie220b73daf3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:28.524074 containerd[1535]: 2025-09-12 22:03:28.504 [INFO][3809] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.524143 containerd[1535]: 2025-09-12 22:03:28.505 [INFO][3809] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie220b73daf3 ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.524143 containerd[1535]: 2025-09-12 22:03:28.511 [INFO][3809] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.524179 containerd[1535]: 2025-09-12 22:03:28.511 [INFO][3809] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--75456b67bf--f8psn-eth0", GenerateName:"whisker-75456b67bf-", Namespace:"calico-system", SelfLink:"", UID:"2161f718-2754-4dbb-bfe9-02688bae41fb", ResourceVersion:"915", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"75456b67bf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437", Pod:"whisker-75456b67bf-f8psn", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calie220b73daf3", MAC:"3a:b8:49:c2:e5:8c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:28.524223 containerd[1535]: 2025-09-12 22:03:28.521 [INFO][3809] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" Namespace="calico-system" Pod="whisker-75456b67bf-f8psn" WorkloadEndpoint="localhost-k8s-whisker--75456b67bf--f8psn-eth0" Sep 12 22:03:28.554859 containerd[1535]: time="2025-09-12T22:03:28.554804374Z" level=info msg="connecting to shim 890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437" address="unix:///run/containerd/s/b208c4cc4d8fca258c5ad43949d44cecdad9c82769a493027e7c71a483147600" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:28.586273 systemd[1]: Started cri-containerd-890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437.scope - libcontainer container 890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437. Sep 12 22:03:28.599416 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:28.620798 containerd[1535]: time="2025-09-12T22:03:28.620716408Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-75456b67bf-f8psn,Uid:2161f718-2754-4dbb-bfe9-02688bae41fb,Namespace:calico-system,Attempt:0,} returns sandbox id \"890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437\"" Sep 12 22:03:28.625368 containerd[1535]: time="2025-09-12T22:03:28.625291975Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 12 22:03:28.876018 kubelet[2665]: I0912 22:03:28.875883 2665 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b21c912-c86c-4756-bba2-0197b8ea9324" path="/var/lib/kubelet/pods/3b21c912-c86c-4756-bba2-0197b8ea9324/volumes" Sep 12 22:03:28.981171 kubelet[2665]: I0912 22:03:28.981141 2665 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:03:29.638186 containerd[1535]: time="2025-09-12T22:03:29.638134207Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:29.638719 containerd[1535]: time="2025-09-12T22:03:29.638691885Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4605606" Sep 12 22:03:29.639414 containerd[1535]: time="2025-09-12T22:03:29.639376813Z" level=info msg="ImageCreate event name:\"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:29.641387 containerd[1535]: time="2025-09-12T22:03:29.641352749Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:29.642248 containerd[1535]: time="2025-09-12T22:03:29.642214609Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"5974839\" in 1.01687419s" Sep 12 22:03:29.642279 containerd[1535]: time="2025-09-12T22:03:29.642248571Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:270a0129ec34c3ad6ae6d56c0afce111eb0baa25dfdacb63722ec5887bafd3c5\"" Sep 12 22:03:29.646530 containerd[1535]: time="2025-09-12T22:03:29.646194763Z" level=info msg="CreateContainer within sandbox \"890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 12 22:03:29.654167 containerd[1535]: time="2025-09-12T22:03:29.654137472Z" level=info msg="Container c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:29.660432 containerd[1535]: time="2025-09-12T22:03:29.660400784Z" level=info msg="CreateContainer within sandbox \"890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b\"" Sep 12 22:03:29.660839 containerd[1535]: time="2025-09-12T22:03:29.660813533Z" level=info msg="StartContainer for \"c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b\"" Sep 12 22:03:29.661996 containerd[1535]: time="2025-09-12T22:03:29.661930570Z" level=info msg="connecting to shim c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b" address="unix:///run/containerd/s/b208c4cc4d8fca258c5ad43949d44cecdad9c82769a493027e7c71a483147600" protocol=ttrpc version=3 Sep 12 22:03:29.683118 systemd[1]: Started cri-containerd-c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b.scope - libcontainer container c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b. Sep 12 22:03:29.716694 containerd[1535]: time="2025-09-12T22:03:29.716604866Z" level=info msg="StartContainer for \"c2b52822638a238c35ced80120463e4e63e0e151541849fcc16ae3433c826d5b\" returns successfully" Sep 12 22:03:29.719444 containerd[1535]: time="2025-09-12T22:03:29.719416660Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 12 22:03:29.756127 systemd-networkd[1435]: calie220b73daf3: Gained IPv6LL Sep 12 22:03:31.078243 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount24671084.mount: Deactivated successfully. Sep 12 22:03:31.169003 containerd[1535]: time="2025-09-12T22:03:31.168532230Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:31.169003 containerd[1535]: time="2025-09-12T22:03:31.168990460Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=30823700" Sep 12 22:03:31.169869 containerd[1535]: time="2025-09-12T22:03:31.169833634Z" level=info msg="ImageCreate event name:\"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:31.172415 containerd[1535]: time="2025-09-12T22:03:31.172381399Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:31.174818 containerd[1535]: time="2025-09-12T22:03:31.174005984Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"30823530\" in 1.454526759s" Sep 12 22:03:31.174818 containerd[1535]: time="2025-09-12T22:03:31.174042026Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:e210e86234bc99f018431b30477c5ca2ad6f7ecf67ef011498f7beb48fb0b21f\"" Sep 12 22:03:31.192463 containerd[1535]: time="2025-09-12T22:03:31.192396131Z" level=info msg="CreateContainer within sandbox \"890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 12 22:03:31.200995 containerd[1535]: time="2025-09-12T22:03:31.199480029Z" level=info msg="Container 07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:31.215406 containerd[1535]: time="2025-09-12T22:03:31.215358214Z" level=info msg="CreateContainer within sandbox \"890721cb6f8c52f0144db717bbd8fff8e6f38cea3fbd346bcf16cd7ff8290437\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4\"" Sep 12 22:03:31.215927 containerd[1535]: time="2025-09-12T22:03:31.215882128Z" level=info msg="StartContainer for \"07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4\"" Sep 12 22:03:31.218386 containerd[1535]: time="2025-09-12T22:03:31.218350567Z" level=info msg="connecting to shim 07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4" address="unix:///run/containerd/s/b208c4cc4d8fca258c5ad43949d44cecdad9c82769a493027e7c71a483147600" protocol=ttrpc version=3 Sep 12 22:03:31.241149 systemd[1]: Started cri-containerd-07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4.scope - libcontainer container 07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4. Sep 12 22:03:31.274579 containerd[1535]: time="2025-09-12T22:03:31.274531155Z" level=info msg="StartContainer for \"07d8f83a4a6945f9d9748bf50c056441e8fcd4e8976126a26f62d0136037d9f4\" returns successfully" Sep 12 22:03:32.004961 kubelet[2665]: I0912 22:03:32.004893 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-75456b67bf-f8psn" podStartSLOduration=1.454169333 podStartE2EDuration="4.004874032s" podCreationTimestamp="2025-09-12 22:03:28 +0000 UTC" firstStartedPulling="2025-09-12 22:03:28.62493343 +0000 UTC m=+33.849412865" lastFinishedPulling="2025-09-12 22:03:31.175638089 +0000 UTC m=+36.400117564" observedRunningTime="2025-09-12 22:03:32.003940013 +0000 UTC m=+37.228419488" watchObservedRunningTime="2025-09-12 22:03:32.004874032 +0000 UTC m=+37.229353507" Sep 12 22:03:34.508005 kubelet[2665]: I0912 22:03:34.507900 2665 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:03:34.633659 containerd[1535]: time="2025-09-12T22:03:34.633607247Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f\" id:\"c012d05ceb413f2e8feef60ee5249cd5f5274420273766ad9bb4cf375d887586\" pid:4205 exit_status:1 exited_at:{seconds:1757714614 nanos:633011852}" Sep 12 22:03:34.707054 containerd[1535]: time="2025-09-12T22:03:34.706999167Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f\" id:\"90b2795447a4f16ac598a13220e291f4bd2ca5bc45f7f8dff59b2735feb2326a\" pid:4230 exit_status:1 exited_at:{seconds:1757714614 nanos:706692869}" Sep 12 22:03:35.864820 kubelet[2665]: E0912 22:03:35.864698 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:35.866199 containerd[1535]: time="2025-09-12T22:03:35.866169906Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-9cgtm,Uid:41f762f2-5fa7-4a7a-b5c8-964bf56dfbee,Namespace:kube-system,Attempt:0,}" Sep 12 22:03:35.868714 kubelet[2665]: I0912 22:03:35.868656 2665 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:03:35.869127 kubelet[2665]: E0912 22:03:35.868967 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:36.003030 kubelet[2665]: E0912 22:03:36.002935 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:36.023694 systemd-networkd[1435]: calia89cb039a48: Link UP Sep 12 22:03:36.023893 systemd-networkd[1435]: calia89cb039a48: Gained carrier Sep 12 22:03:36.040886 containerd[1535]: 2025-09-12 22:03:35.902 [INFO][4271] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 12 22:03:36.040886 containerd[1535]: 2025-09-12 22:03:35.955 [INFO][4271] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0 coredns-7c65d6cfc9- kube-system 41f762f2-5fa7-4a7a-b5c8-964bf56dfbee 852 0 2025-09-12 22:03:02 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7c65d6cfc9-9cgtm eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calia89cb039a48 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-" Sep 12 22:03:36.040886 containerd[1535]: 2025-09-12 22:03:35.956 [INFO][4271] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.040886 containerd[1535]: 2025-09-12 22:03:35.985 [INFO][4289] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" HandleID="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Workload="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:35.986 [INFO][4289] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" HandleID="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Workload="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000323490), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7c65d6cfc9-9cgtm", "timestamp":"2025-09-12 22:03:35.985943994 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:35.986 [INFO][4289] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:35.986 [INFO][4289] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:35.986 [INFO][4289] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:35.995 [INFO][4289] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" host="localhost" Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:36.001 [INFO][4289] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:36.006 [INFO][4289] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:36.008 [INFO][4289] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:36.010 [INFO][4289] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:36.041168 containerd[1535]: 2025-09-12 22:03:36.010 [INFO][4289] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" host="localhost" Sep 12 22:03:36.041361 containerd[1535]: 2025-09-12 22:03:36.012 [INFO][4289] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba Sep 12 22:03:36.041361 containerd[1535]: 2025-09-12 22:03:36.014 [INFO][4289] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" host="localhost" Sep 12 22:03:36.041361 containerd[1535]: 2025-09-12 22:03:36.020 [INFO][4289] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" host="localhost" Sep 12 22:03:36.041361 containerd[1535]: 2025-09-12 22:03:36.020 [INFO][4289] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" host="localhost" Sep 12 22:03:36.041361 containerd[1535]: 2025-09-12 22:03:36.020 [INFO][4289] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:36.041361 containerd[1535]: 2025-09-12 22:03:36.020 [INFO][4289] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" HandleID="k8s-pod-network.fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Workload="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.041508 containerd[1535]: 2025-09-12 22:03:36.022 [INFO][4271] cni-plugin/k8s.go 418: Populated endpoint ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"41f762f2-5fa7-4a7a-b5c8-964bf56dfbee", ResourceVersion:"852", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7c65d6cfc9-9cgtm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia89cb039a48", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:36.041580 containerd[1535]: 2025-09-12 22:03:36.022 [INFO][4271] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.041580 containerd[1535]: 2025-09-12 22:03:36.022 [INFO][4271] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia89cb039a48 ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.041580 containerd[1535]: 2025-09-12 22:03:36.024 [INFO][4271] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.041646 containerd[1535]: 2025-09-12 22:03:36.024 [INFO][4271] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"41f762f2-5fa7-4a7a-b5c8-964bf56dfbee", ResourceVersion:"852", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba", Pod:"coredns-7c65d6cfc9-9cgtm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia89cb039a48", MAC:"22:c5:94:3f:51:1a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:36.041646 containerd[1535]: 2025-09-12 22:03:36.038 [INFO][4271] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" Namespace="kube-system" Pod="coredns-7c65d6cfc9-9cgtm" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--9cgtm-eth0" Sep 12 22:03:36.060646 containerd[1535]: time="2025-09-12T22:03:36.060610650Z" level=info msg="connecting to shim fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba" address="unix:///run/containerd/s/ddd0682e9dbe1f78557046b43431f246e74f58ebd90cb02366affad5951338c1" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:36.085128 systemd[1]: Started cri-containerd-fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba.scope - libcontainer container fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba. Sep 12 22:03:36.095681 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:36.117546 containerd[1535]: time="2025-09-12T22:03:36.117366806Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-9cgtm,Uid:41f762f2-5fa7-4a7a-b5c8-964bf56dfbee,Namespace:kube-system,Attempt:0,} returns sandbox id \"fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba\"" Sep 12 22:03:36.124900 kubelet[2665]: E0912 22:03:36.124871 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:36.127011 containerd[1535]: time="2025-09-12T22:03:36.126947499Z" level=info msg="CreateContainer within sandbox \"fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 22:03:36.152191 containerd[1535]: time="2025-09-12T22:03:36.152150261Z" level=info msg="Container 4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:36.153153 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount776947974.mount: Deactivated successfully. Sep 12 22:03:36.159189 containerd[1535]: time="2025-09-12T22:03:36.159150330Z" level=info msg="CreateContainer within sandbox \"fbcf3c27c2b70f6a0538023f987aa67c96f4be8aa037f4236a3e335a71c75fba\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636\"" Sep 12 22:03:36.159816 containerd[1535]: time="2025-09-12T22:03:36.159772805Z" level=info msg="StartContainer for \"4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636\"" Sep 12 22:03:36.160716 containerd[1535]: time="2025-09-12T22:03:36.160686335Z" level=info msg="connecting to shim 4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636" address="unix:///run/containerd/s/ddd0682e9dbe1f78557046b43431f246e74f58ebd90cb02366affad5951338c1" protocol=ttrpc version=3 Sep 12 22:03:36.181144 systemd[1]: Started cri-containerd-4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636.scope - libcontainer container 4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636. Sep 12 22:03:36.221738 containerd[1535]: time="2025-09-12T22:03:36.221700488Z" level=info msg="StartContainer for \"4fdfaeb28302c4fea3d379709525d956de82a1456fe33637b01360aa82a3e636\" returns successfully" Sep 12 22:03:36.562495 systemd-networkd[1435]: vxlan.calico: Link UP Sep 12 22:03:36.562501 systemd-networkd[1435]: vxlan.calico: Gained carrier Sep 12 22:03:36.822812 systemd[1]: Started sshd@7-10.0.0.29:22-10.0.0.1:41788.service - OpenSSH per-connection server daemon (10.0.0.1:41788). Sep 12 22:03:36.864588 kubelet[2665]: E0912 22:03:36.864420 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:36.865998 containerd[1535]: time="2025-09-12T22:03:36.865206074Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-2v4dd,Uid:9c729210-e125-4923-8a29-db7b0068bea7,Namespace:kube-system,Attempt:0,}" Sep 12 22:03:36.865998 containerd[1535]: time="2025-09-12T22:03:36.865425366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-86d86dd9d9-dgcmg,Uid:04200bdf-66fa-47d4-acdd-9b4679fe0b2b,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:36.865998 containerd[1535]: time="2025-09-12T22:03:36.865522171Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-n9cxf,Uid:5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:03:36.873984 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4242371582.mount: Deactivated successfully. Sep 12 22:03:36.895118 sshd[4524]: Accepted publickey for core from 10.0.0.1 port 41788 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:36.896937 sshd-session[4524]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:36.903036 systemd-logind[1513]: New session 8 of user core. Sep 12 22:03:36.910141 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 12 22:03:37.006070 systemd-networkd[1435]: cali9f76dceb77d: Link UP Sep 12 22:03:37.006967 systemd-networkd[1435]: cali9f76dceb77d: Gained carrier Sep 12 22:03:37.018846 kubelet[2665]: E0912 22:03:37.018801 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.926 [INFO][4552] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0 calico-kube-controllers-86d86dd9d9- calico-system 04200bdf-66fa-47d4-acdd-9b4679fe0b2b 854 0 2025-09-12 22:03:16 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:86d86dd9d9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-86d86dd9d9-dgcmg eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali9f76dceb77d [] [] }} ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.927 [INFO][4552] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.958 [INFO][4578] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" HandleID="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Workload="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.958 [INFO][4578] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" HandleID="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Workload="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001196b0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-86d86dd9d9-dgcmg", "timestamp":"2025-09-12 22:03:36.958585586 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.958 [INFO][4578] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.958 [INFO][4578] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.958 [INFO][4578] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.971 [INFO][4578] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.976 [INFO][4578] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.980 [INFO][4578] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.981 [INFO][4578] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.985 [INFO][4578] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.985 [INFO][4578] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.986 [INFO][4578] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8 Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:36.991 [INFO][4578] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:37.000 [INFO][4578] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:37.000 [INFO][4578] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" host="localhost" Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:37.000 [INFO][4578] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:37.025061 containerd[1535]: 2025-09-12 22:03:37.000 [INFO][4578] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" HandleID="k8s-pod-network.e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Workload="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.026398 containerd[1535]: 2025-09-12 22:03:37.003 [INFO][4552] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0", GenerateName:"calico-kube-controllers-86d86dd9d9-", Namespace:"calico-system", SelfLink:"", UID:"04200bdf-66fa-47d4-acdd-9b4679fe0b2b", ResourceVersion:"854", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"86d86dd9d9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-86d86dd9d9-dgcmg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali9f76dceb77d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.026398 containerd[1535]: 2025-09-12 22:03:37.003 [INFO][4552] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.026398 containerd[1535]: 2025-09-12 22:03:37.003 [INFO][4552] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9f76dceb77d ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.026398 containerd[1535]: 2025-09-12 22:03:37.005 [INFO][4552] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.026398 containerd[1535]: 2025-09-12 22:03:37.006 [INFO][4552] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0", GenerateName:"calico-kube-controllers-86d86dd9d9-", Namespace:"calico-system", SelfLink:"", UID:"04200bdf-66fa-47d4-acdd-9b4679fe0b2b", ResourceVersion:"854", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"86d86dd9d9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8", Pod:"calico-kube-controllers-86d86dd9d9-dgcmg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali9f76dceb77d", MAC:"76:4d:1b:70:e6:ec", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.026398 containerd[1535]: 2025-09-12 22:03:37.020 [INFO][4552] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" Namespace="calico-system" Pod="calico-kube-controllers-86d86dd9d9-dgcmg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--86d86dd9d9--dgcmg-eth0" Sep 12 22:03:37.033988 kubelet[2665]: I0912 22:03:37.033865 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-9cgtm" podStartSLOduration=35.03211691 podStartE2EDuration="35.03211691s" podCreationTimestamp="2025-09-12 22:03:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:03:37.031317026 +0000 UTC m=+42.255796501" watchObservedRunningTime="2025-09-12 22:03:37.03211691 +0000 UTC m=+42.256596345" Sep 12 22:03:37.078905 containerd[1535]: time="2025-09-12T22:03:37.078808397Z" level=info msg="connecting to shim e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8" address="unix:///run/containerd/s/46c99b8876e90842435373a1dfd2184fdff121450e5b72c52a2be72ef29a0d6f" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:37.122596 systemd[1]: Started cri-containerd-e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8.scope - libcontainer container e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8. Sep 12 22:03:37.133096 systemd-networkd[1435]: cali3eecf0aad72: Link UP Sep 12 22:03:37.133865 systemd-networkd[1435]: cali3eecf0aad72: Gained carrier Sep 12 22:03:37.144039 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:36.927 [INFO][4527] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0 coredns-7c65d6cfc9- kube-system 9c729210-e125-4923-8a29-db7b0068bea7 845 0 2025-09-12 22:03:02 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7c65d6cfc9-2v4dd eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali3eecf0aad72 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:36.927 [INFO][4527] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:36.969 [INFO][4585] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" HandleID="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Workload="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:36.969 [INFO][4585] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" HandleID="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Workload="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000137760), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7c65d6cfc9-2v4dd", "timestamp":"2025-09-12 22:03:36.969641241 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:36.969 [INFO][4585] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.000 [INFO][4585] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.000 [INFO][4585] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.075 [INFO][4585] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.083 [INFO][4585] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.091 [INFO][4585] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.095 [INFO][4585] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.097 [INFO][4585] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.097 [INFO][4585] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.099 [INFO][4585] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26 Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.107 [INFO][4585] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.116 [INFO][4585] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.116 [INFO][4585] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" host="localhost" Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.116 [INFO][4585] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:37.153415 containerd[1535]: 2025-09-12 22:03:37.116 [INFO][4585] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" HandleID="k8s-pod-network.00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Workload="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.153900 containerd[1535]: 2025-09-12 22:03:37.122 [INFO][4527] cni-plugin/k8s.go 418: Populated endpoint ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"9c729210-e125-4923-8a29-db7b0068bea7", ResourceVersion:"845", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7c65d6cfc9-2v4dd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3eecf0aad72", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.153900 containerd[1535]: 2025-09-12 22:03:37.127 [INFO][4527] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.153900 containerd[1535]: 2025-09-12 22:03:37.128 [INFO][4527] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3eecf0aad72 ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.153900 containerd[1535]: 2025-09-12 22:03:37.134 [INFO][4527] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.153900 containerd[1535]: 2025-09-12 22:03:37.135 [INFO][4527] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"9c729210-e125-4923-8a29-db7b0068bea7", ResourceVersion:"845", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26", Pod:"coredns-7c65d6cfc9-2v4dd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali3eecf0aad72", MAC:"2e:de:fc:9f:75:1e", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.153900 containerd[1535]: 2025-09-12 22:03:37.148 [INFO][4527] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" Namespace="kube-system" Pod="coredns-7c65d6cfc9-2v4dd" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--2v4dd-eth0" Sep 12 22:03:37.180946 containerd[1535]: time="2025-09-12T22:03:37.180903805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-86d86dd9d9-dgcmg,Uid:04200bdf-66fa-47d4-acdd-9b4679fe0b2b,Namespace:calico-system,Attempt:0,} returns sandbox id \"e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8\"" Sep 12 22:03:37.191058 containerd[1535]: time="2025-09-12T22:03:37.190840143Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 12 22:03:37.191331 containerd[1535]: time="2025-09-12T22:03:37.191294767Z" level=info msg="connecting to shim 00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26" address="unix:///run/containerd/s/80466f76580ae758af008c82f3a8484e6d0435dd3167ba541f2fb8b58a11997b" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:37.198609 sshd[4570]: Connection closed by 10.0.0.1 port 41788 Sep 12 22:03:37.199047 sshd-session[4524]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:37.204285 systemd[1]: sshd@7-10.0.0.29:22-10.0.0.1:41788.service: Deactivated successfully. Sep 12 22:03:37.206642 systemd[1]: session-8.scope: Deactivated successfully. Sep 12 22:03:37.207818 systemd-logind[1513]: Session 8 logged out. Waiting for processes to exit. Sep 12 22:03:37.209010 systemd-logind[1513]: Removed session 8. Sep 12 22:03:37.222149 systemd[1]: Started cri-containerd-00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26.scope - libcontainer container 00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26. Sep 12 22:03:37.233338 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:37.278528 systemd-networkd[1435]: calieab05fdc064: Link UP Sep 12 22:03:37.279311 systemd-networkd[1435]: calieab05fdc064: Gained carrier Sep 12 22:03:37.281332 containerd[1535]: time="2025-09-12T22:03:37.281293560Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-2v4dd,Uid:9c729210-e125-4923-8a29-db7b0068bea7,Namespace:kube-system,Attempt:0,} returns sandbox id \"00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26\"" Sep 12 22:03:37.283239 kubelet[2665]: E0912 22:03:37.283212 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:37.287931 containerd[1535]: time="2025-09-12T22:03:37.287902237Z" level=info msg="CreateContainer within sandbox \"00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:36.925 [INFO][4535] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0 calico-apiserver-57d459fd7c- calico-apiserver 5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5 855 0 2025-09-12 22:03:11 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:57d459fd7c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-57d459fd7c-n9cxf eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calieab05fdc064 [] [] }} ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:36.927 [INFO][4535] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:36.971 [INFO][4576] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" HandleID="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Workload="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:36.971 [INFO][4576] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" HandleID="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Workload="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000136520), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-57d459fd7c-n9cxf", "timestamp":"2025-09-12 22:03:36.971102763 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:36.972 [INFO][4576] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.116 [INFO][4576] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.116 [INFO][4576] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.175 [INFO][4576] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.186 [INFO][4576] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.200 [INFO][4576] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.202 [INFO][4576] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.206 [INFO][4576] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.206 [INFO][4576] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.207 [INFO][4576] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.261 [INFO][4576] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.273 [INFO][4576] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.273 [INFO][4576] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" host="localhost" Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.273 [INFO][4576] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:37.299295 containerd[1535]: 2025-09-12 22:03:37.274 [INFO][4576] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" HandleID="k8s-pod-network.c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Workload="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.300214 containerd[1535]: 2025-09-12 22:03:37.275 [INFO][4535] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0", GenerateName:"calico-apiserver-57d459fd7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5", ResourceVersion:"855", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57d459fd7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-57d459fd7c-n9cxf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieab05fdc064", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.300214 containerd[1535]: 2025-09-12 22:03:37.275 [INFO][4535] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.300214 containerd[1535]: 2025-09-12 22:03:37.276 [INFO][4535] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieab05fdc064 ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.300214 containerd[1535]: 2025-09-12 22:03:37.279 [INFO][4535] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.300214 containerd[1535]: 2025-09-12 22:03:37.280 [INFO][4535] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0", GenerateName:"calico-apiserver-57d459fd7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5", ResourceVersion:"855", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57d459fd7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e", Pod:"calico-apiserver-57d459fd7c-n9cxf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieab05fdc064", MAC:"26:b4:4e:90:05:60", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.300214 containerd[1535]: 2025-09-12 22:03:37.294 [INFO][4535] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-n9cxf" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--n9cxf-eth0" Sep 12 22:03:37.300214 containerd[1535]: time="2025-09-12T22:03:37.299243931Z" level=info msg="Container 5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:37.306132 containerd[1535]: time="2025-09-12T22:03:37.306081102Z" level=info msg="CreateContainer within sandbox \"00727764d0fee17ee91bffc5e4e941e6798cd3ef41a29f045c5eb06f94a52a26\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be\"" Sep 12 22:03:37.306649 containerd[1535]: time="2025-09-12T22:03:37.306611850Z" level=info msg="StartContainer for \"5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be\"" Sep 12 22:03:37.307762 containerd[1535]: time="2025-09-12T22:03:37.307732711Z" level=info msg="connecting to shim 5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be" address="unix:///run/containerd/s/80466f76580ae758af008c82f3a8484e6d0435dd3167ba541f2fb8b58a11997b" protocol=ttrpc version=3 Sep 12 22:03:37.329316 containerd[1535]: time="2025-09-12T22:03:37.329198873Z" level=info msg="connecting to shim c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e" address="unix:///run/containerd/s/d180c183c142de115b873d70783197b3196f316167d97e4db3d055f53f06a9e1" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:37.330352 systemd[1]: Started cri-containerd-5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be.scope - libcontainer container 5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be. Sep 12 22:03:37.352127 systemd[1]: Started cri-containerd-c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e.scope - libcontainer container c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e. Sep 12 22:03:37.366792 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:37.375008 containerd[1535]: time="2025-09-12T22:03:37.373673961Z" level=info msg="StartContainer for \"5ee1ae9963529147ba19eeacf2765621b05e15d73a6df9a117a1de7d3ef464be\" returns successfully" Sep 12 22:03:37.396894 containerd[1535]: time="2025-09-12T22:03:37.396846215Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-n9cxf,Uid:5f1bb2b2-f6f1-4e4b-a7a3-ccf525f155d5,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e\"" Sep 12 22:03:37.565859 systemd-networkd[1435]: calia89cb039a48: Gained IPv6LL Sep 12 22:03:37.863780 containerd[1535]: time="2025-09-12T22:03:37.863674609Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-l6kcq,Uid:a69477ef-6c97-40b7-8dbb-52ed876ed57d,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:37.863894 containerd[1535]: time="2025-09-12T22:03:37.863676689Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-7mhnw,Uid:3cdb4f52-c9af-495e-a27e-8dd84094cf42,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:03:37.964900 systemd-networkd[1435]: calia4b8110d9c0: Link UP Sep 12 22:03:37.965587 systemd-networkd[1435]: calia4b8110d9c0: Gained carrier Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.906 [INFO][4829] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0 calico-apiserver-57d459fd7c- calico-apiserver 3cdb4f52-c9af-495e-a27e-8dd84094cf42 851 0 2025-09-12 22:03:11 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:57d459fd7c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-57d459fd7c-7mhnw eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calia4b8110d9c0 [] [] }} ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.906 [INFO][4829] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.928 [INFO][4848] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" HandleID="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Workload="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.928 [INFO][4848] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" HandleID="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Workload="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40005232a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-57d459fd7c-7mhnw", "timestamp":"2025-09-12 22:03:37.92871021 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.928 [INFO][4848] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.928 [INFO][4848] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.929 [INFO][4848] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.937 [INFO][4848] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.942 [INFO][4848] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.947 [INFO][4848] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.948 [INFO][4848] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.950 [INFO][4848] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.950 [INFO][4848] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.952 [INFO][4848] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.955 [INFO][4848] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.960 [INFO][4848] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.960 [INFO][4848] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" host="localhost" Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.960 [INFO][4848] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:37.981059 containerd[1535]: 2025-09-12 22:03:37.960 [INFO][4848] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" HandleID="k8s-pod-network.1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Workload="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:37.981546 containerd[1535]: 2025-09-12 22:03:37.963 [INFO][4829] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0", GenerateName:"calico-apiserver-57d459fd7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"3cdb4f52-c9af-495e-a27e-8dd84094cf42", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57d459fd7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-57d459fd7c-7mhnw", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia4b8110d9c0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.981546 containerd[1535]: 2025-09-12 22:03:37.963 [INFO][4829] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:37.981546 containerd[1535]: 2025-09-12 22:03:37.963 [INFO][4829] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia4b8110d9c0 ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:37.981546 containerd[1535]: 2025-09-12 22:03:37.965 [INFO][4829] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:37.981546 containerd[1535]: 2025-09-12 22:03:37.965 [INFO][4829] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0", GenerateName:"calico-apiserver-57d459fd7c-", Namespace:"calico-apiserver", SelfLink:"", UID:"3cdb4f52-c9af-495e-a27e-8dd84094cf42", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 11, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"57d459fd7c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b", Pod:"calico-apiserver-57d459fd7c-7mhnw", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia4b8110d9c0", MAC:"9a:01:e8:79:16:d0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:37.981546 containerd[1535]: 2025-09-12 22:03:37.978 [INFO][4829] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" Namespace="calico-apiserver" Pod="calico-apiserver-57d459fd7c-7mhnw" WorkloadEndpoint="localhost-k8s-calico--apiserver--57d459fd7c--7mhnw-eth0" Sep 12 22:03:38.010008 containerd[1535]: time="2025-09-12T22:03:38.009846750Z" level=info msg="connecting to shim 1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b" address="unix:///run/containerd/s/77e0bb942f7fbb47ced026241678ce2e148eade6c2bb63f0e5cc4461a5ea6216" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:38.024214 kubelet[2665]: E0912 22:03:38.023802 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:38.025350 kubelet[2665]: E0912 22:03:38.025315 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:38.038396 kubelet[2665]: I0912 22:03:38.038259 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-2v4dd" podStartSLOduration=36.038242488 podStartE2EDuration="36.038242488s" podCreationTimestamp="2025-09-12 22:03:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:03:38.037597654 +0000 UTC m=+43.262077129" watchObservedRunningTime="2025-09-12 22:03:38.038242488 +0000 UTC m=+43.262721963" Sep 12 22:03:38.040496 systemd[1]: Started cri-containerd-1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b.scope - libcontainer container 1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b. Sep 12 22:03:38.076176 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:38.096465 systemd-networkd[1435]: cali1d203eb35b1: Link UP Sep 12 22:03:38.097481 systemd-networkd[1435]: cali1d203eb35b1: Gained carrier Sep 12 22:03:38.113377 containerd[1535]: time="2025-09-12T22:03:38.113327729Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-57d459fd7c-7mhnw,Uid:3cdb4f52-c9af-495e-a27e-8dd84094cf42,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b\"" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.903 [INFO][4817] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--7988f88666--l6kcq-eth0 goldmane-7988f88666- calico-system a69477ef-6c97-40b7-8dbb-52ed876ed57d 850 0 2025-09-12 22:03:16 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:7988f88666 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-7988f88666-l6kcq eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali1d203eb35b1 [] [] }} ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.903 [INFO][4817] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.929 [INFO][4846] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" HandleID="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Workload="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.929 [INFO][4846] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" HandleID="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Workload="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c610), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-7988f88666-l6kcq", "timestamp":"2025-09-12 22:03:37.929154474 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.929 [INFO][4846] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.960 [INFO][4846] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:37.960 [INFO][4846] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.041 [INFO][4846] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.050 [INFO][4846] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.056 [INFO][4846] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.058 [INFO][4846] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.063 [INFO][4846] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.064 [INFO][4846] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.067 [INFO][4846] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3 Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.074 [INFO][4846] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.085 [INFO][4846] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.085 [INFO][4846] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" host="localhost" Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.085 [INFO][4846] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:38.124413 containerd[1535]: 2025-09-12 22:03:38.085 [INFO][4846] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" HandleID="k8s-pod-network.26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Workload="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.124860 containerd[1535]: 2025-09-12 22:03:38.094 [INFO][4817] cni-plugin/k8s.go 418: Populated endpoint ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--7988f88666--l6kcq-eth0", GenerateName:"goldmane-7988f88666-", Namespace:"calico-system", SelfLink:"", UID:"a69477ef-6c97-40b7-8dbb-52ed876ed57d", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7988f88666", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-7988f88666-l6kcq", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali1d203eb35b1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:38.124860 containerd[1535]: 2025-09-12 22:03:38.094 [INFO][4817] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.124860 containerd[1535]: 2025-09-12 22:03:38.094 [INFO][4817] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1d203eb35b1 ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.124860 containerd[1535]: 2025-09-12 22:03:38.104 [INFO][4817] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.124860 containerd[1535]: 2025-09-12 22:03:38.104 [INFO][4817] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--7988f88666--l6kcq-eth0", GenerateName:"goldmane-7988f88666-", Namespace:"calico-system", SelfLink:"", UID:"a69477ef-6c97-40b7-8dbb-52ed876ed57d", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7988f88666", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3", Pod:"goldmane-7988f88666-l6kcq", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali1d203eb35b1", MAC:"2e:28:04:6a:20:b6", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:38.124860 containerd[1535]: 2025-09-12 22:03:38.119 [INFO][4817] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" Namespace="calico-system" Pod="goldmane-7988f88666-l6kcq" WorkloadEndpoint="localhost-k8s-goldmane--7988f88666--l6kcq-eth0" Sep 12 22:03:38.172982 containerd[1535]: time="2025-09-12T22:03:38.172920353Z" level=info msg="connecting to shim 26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3" address="unix:///run/containerd/s/7e07a61337ad3c84d6e7bec4cf3e8df5b0b2d0f32e582c3d31c03f75098f5487" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:38.193213 systemd[1]: Started cri-containerd-26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3.scope - libcontainer container 26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3. Sep 12 22:03:38.204145 systemd-networkd[1435]: vxlan.calico: Gained IPv6LL Sep 12 22:03:38.215967 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:38.308032 containerd[1535]: time="2025-09-12T22:03:38.307963478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7988f88666-l6kcq,Uid:a69477ef-6c97-40b7-8dbb-52ed876ed57d,Namespace:calico-system,Attempt:0,} returns sandbox id \"26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3\"" Sep 12 22:03:38.460249 systemd-networkd[1435]: calieab05fdc064: Gained IPv6LL Sep 12 22:03:38.716309 systemd-networkd[1435]: cali9f76dceb77d: Gained IPv6LL Sep 12 22:03:38.736273 containerd[1535]: time="2025-09-12T22:03:38.736226152Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:38.736804 containerd[1535]: time="2025-09-12T22:03:38.736725259Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=48134957" Sep 12 22:03:38.737928 containerd[1535]: time="2025-09-12T22:03:38.737689310Z" level=info msg="ImageCreate event name:\"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:38.739704 containerd[1535]: time="2025-09-12T22:03:38.739668974Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:38.740453 containerd[1535]: time="2025-09-12T22:03:38.740422414Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"49504166\" in 1.549542669s" Sep 12 22:03:38.740545 containerd[1535]: time="2025-09-12T22:03:38.740529819Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:34117caf92350e1565610f2254377d7455b11e36666b5ce11b4a13670720432a\"" Sep 12 22:03:38.742745 containerd[1535]: time="2025-09-12T22:03:38.742675373Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 12 22:03:38.747615 containerd[1535]: time="2025-09-12T22:03:38.747579071Z" level=info msg="CreateContainer within sandbox \"e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 12 22:03:38.752767 containerd[1535]: time="2025-09-12T22:03:38.752725943Z" level=info msg="Container 2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:38.758498 containerd[1535]: time="2025-09-12T22:03:38.758453005Z" level=info msg="CreateContainer within sandbox \"e39cb6a76ebe43e980eba83ff4e8830fa02142f15053fc646c09fb2d5d8714f8\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca\"" Sep 12 22:03:38.759084 containerd[1535]: time="2025-09-12T22:03:38.759056877Z" level=info msg="StartContainer for \"2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca\"" Sep 12 22:03:38.760259 containerd[1535]: time="2025-09-12T22:03:38.760230859Z" level=info msg="connecting to shim 2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca" address="unix:///run/containerd/s/46c99b8876e90842435373a1dfd2184fdff121450e5b72c52a2be72ef29a0d6f" protocol=ttrpc version=3 Sep 12 22:03:38.789164 systemd[1]: Started cri-containerd-2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca.scope - libcontainer container 2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca. Sep 12 22:03:38.824248 containerd[1535]: time="2025-09-12T22:03:38.824214755Z" level=info msg="StartContainer for \"2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca\" returns successfully" Sep 12 22:03:39.031961 kubelet[2665]: E0912 22:03:39.031794 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:39.033066 kubelet[2665]: E0912 22:03:39.032303 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:39.047539 kubelet[2665]: I0912 22:03:39.046500 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-86d86dd9d9-dgcmg" podStartSLOduration=21.495580521 podStartE2EDuration="23.046484742s" podCreationTimestamp="2025-09-12 22:03:16 +0000 UTC" firstStartedPulling="2025-09-12 22:03:37.19060733 +0000 UTC m=+42.415086805" lastFinishedPulling="2025-09-12 22:03:38.741511591 +0000 UTC m=+43.965991026" observedRunningTime="2025-09-12 22:03:39.045445289 +0000 UTC m=+44.269924804" watchObservedRunningTime="2025-09-12 22:03:39.046484742 +0000 UTC m=+44.270964217" Sep 12 22:03:39.064202 containerd[1535]: time="2025-09-12T22:03:39.064167372Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca\" id:\"fe0368d181db3a2c9c615349d4c1b32211b0f823fa1fe92c51e3e948a81bf331\" pid:5038 exited_at:{seconds:1757714619 nanos:63888118}" Sep 12 22:03:39.100124 systemd-networkd[1435]: cali3eecf0aad72: Gained IPv6LL Sep 12 22:03:39.484128 systemd-networkd[1435]: calia4b8110d9c0: Gained IPv6LL Sep 12 22:03:39.804768 systemd-networkd[1435]: cali1d203eb35b1: Gained IPv6LL Sep 12 22:03:40.033899 kubelet[2665]: E0912 22:03:40.033873 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:03:40.238571 containerd[1535]: time="2025-09-12T22:03:40.238521402Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:40.239757 containerd[1535]: time="2025-09-12T22:03:40.239726263Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=44530807" Sep 12 22:03:40.240702 containerd[1535]: time="2025-09-12T22:03:40.240678470Z" level=info msg="ImageCreate event name:\"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:40.243000 containerd[1535]: time="2025-09-12T22:03:40.242959745Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:40.243602 containerd[1535]: time="2025-09-12T22:03:40.243563295Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"45900064\" in 1.500821919s" Sep 12 22:03:40.243632 containerd[1535]: time="2025-09-12T22:03:40.243609938Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\"" Sep 12 22:03:40.245142 containerd[1535]: time="2025-09-12T22:03:40.245012768Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 12 22:03:40.245656 containerd[1535]: time="2025-09-12T22:03:40.245612318Z" level=info msg="CreateContainer within sandbox \"c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 12 22:03:40.252993 containerd[1535]: time="2025-09-12T22:03:40.252220570Z" level=info msg="Container e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:40.258725 containerd[1535]: time="2025-09-12T22:03:40.258693296Z" level=info msg="CreateContainer within sandbox \"c8697d0050b028aea683c7fdef9b8bc7d6f4adb39d7bff69ba2df69e9199745e\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28\"" Sep 12 22:03:40.259172 containerd[1535]: time="2025-09-12T22:03:40.259072515Z" level=info msg="StartContainer for \"e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28\"" Sep 12 22:03:40.260038 containerd[1535]: time="2025-09-12T22:03:40.260014962Z" level=info msg="connecting to shim e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28" address="unix:///run/containerd/s/d180c183c142de115b873d70783197b3196f316167d97e4db3d055f53f06a9e1" protocol=ttrpc version=3 Sep 12 22:03:40.281141 systemd[1]: Started cri-containerd-e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28.scope - libcontainer container e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28. Sep 12 22:03:40.319859 containerd[1535]: time="2025-09-12T22:03:40.316960144Z" level=info msg="StartContainer for \"e2fa2cd011042f638db57454fa83f620c3b3759f3bbe80e878cb2f9a5fbcea28\" returns successfully" Sep 12 22:03:40.528495 containerd[1535]: time="2025-09-12T22:03:40.528371688Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:40.529437 containerd[1535]: time="2025-09-12T22:03:40.529005599Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Sep 12 22:03:40.530968 containerd[1535]: time="2025-09-12T22:03:40.530930576Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"45900064\" in 285.886727ms" Sep 12 22:03:40.530968 containerd[1535]: time="2025-09-12T22:03:40.530965138Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:632fbde00b1918016ac07458e79cc438ccda83cb762bfd5fc50a26721abced08\"" Sep 12 22:03:40.532007 containerd[1535]: time="2025-09-12T22:03:40.531745977Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 12 22:03:40.532882 containerd[1535]: time="2025-09-12T22:03:40.532859793Z" level=info msg="CreateContainer within sandbox \"1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 12 22:03:40.542271 containerd[1535]: time="2025-09-12T22:03:40.542242385Z" level=info msg="Container 4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:40.545852 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3593510875.mount: Deactivated successfully. Sep 12 22:03:40.557378 containerd[1535]: time="2025-09-12T22:03:40.557340663Z" level=info msg="CreateContainer within sandbox \"1928b048ee3e378df292bdf54fea8c24156e87ff26ba9ed883cae61044f9553b\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1\"" Sep 12 22:03:40.557764 containerd[1535]: time="2025-09-12T22:03:40.557743444Z" level=info msg="StartContainer for \"4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1\"" Sep 12 22:03:40.558833 containerd[1535]: time="2025-09-12T22:03:40.558809817Z" level=info msg="connecting to shim 4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1" address="unix:///run/containerd/s/77e0bb942f7fbb47ced026241678ce2e148eade6c2bb63f0e5cc4461a5ea6216" protocol=ttrpc version=3 Sep 12 22:03:40.580125 systemd[1]: Started cri-containerd-4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1.scope - libcontainer container 4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1. Sep 12 22:03:40.626524 containerd[1535]: time="2025-09-12T22:03:40.626488258Z" level=info msg="StartContainer for \"4e81f6ddee00c68452a7ce6ff2111eb793aba60900c12d7d7302d8d1add6d1a1\" returns successfully" Sep 12 22:03:40.865247 containerd[1535]: time="2025-09-12T22:03:40.864877118Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wdhc9,Uid:9741ef86-7513-4899-85e2-cf42c12fa232,Namespace:calico-system,Attempt:0,}" Sep 12 22:03:41.049567 systemd-networkd[1435]: cali1364a10672f: Link UP Sep 12 22:03:41.050761 systemd-networkd[1435]: cali1364a10672f: Gained carrier Sep 12 22:03:41.055238 kubelet[2665]: I0912 22:03:41.055012 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-57d459fd7c-7mhnw" podStartSLOduration=27.638065601 podStartE2EDuration="30.05499641s" podCreationTimestamp="2025-09-12 22:03:11 +0000 UTC" firstStartedPulling="2025-09-12 22:03:38.114736524 +0000 UTC m=+43.339215959" lastFinishedPulling="2025-09-12 22:03:40.531667293 +0000 UTC m=+45.756146768" observedRunningTime="2025-09-12 22:03:41.053653504 +0000 UTC m=+46.278133019" watchObservedRunningTime="2025-09-12 22:03:41.05499641 +0000 UTC m=+46.279475885" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.964 [INFO][5134] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--wdhc9-eth0 csi-node-driver- calico-system 9741ef86-7513-4899-85e2-cf42c12fa232 751 0 2025-09-12 22:03:16 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:856c6b598f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-wdhc9 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali1364a10672f [] [] }} ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.964 [INFO][5134] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.999 [INFO][5148] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" HandleID="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Workload="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.999 [INFO][5148] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" HandleID="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Workload="localhost-k8s-csi--node--driver--wdhc9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40004953b0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-wdhc9", "timestamp":"2025-09-12 22:03:40.999224389 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.999 [INFO][5148] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.999 [INFO][5148] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:40.999 [INFO][5148] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.009 [INFO][5148] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.015 [INFO][5148] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.019 [INFO][5148] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.022 [INFO][5148] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.024 [INFO][5148] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.024 [INFO][5148] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.026 [INFO][5148] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6 Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.031 [INFO][5148] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.038 [INFO][5148] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.038 [INFO][5148] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" host="localhost" Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.038 [INFO][5148] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:03:41.069274 containerd[1535]: 2025-09-12 22:03:41.038 [INFO][5148] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" HandleID="k8s-pod-network.7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Workload="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.071223 containerd[1535]: 2025-09-12 22:03:41.045 [INFO][5134] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--wdhc9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"9741ef86-7513-4899-85e2-cf42c12fa232", ResourceVersion:"751", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"856c6b598f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-wdhc9", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali1364a10672f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:41.071223 containerd[1535]: 2025-09-12 22:03:41.045 [INFO][5134] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.071223 containerd[1535]: 2025-09-12 22:03:41.045 [INFO][5134] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1364a10672f ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.071223 containerd[1535]: 2025-09-12 22:03:41.051 [INFO][5134] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.071223 containerd[1535]: 2025-09-12 22:03:41.054 [INFO][5134] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--wdhc9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"9741ef86-7513-4899-85e2-cf42c12fa232", ResourceVersion:"751", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 3, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"856c6b598f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6", Pod:"csi-node-driver-wdhc9", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali1364a10672f", MAC:"56:ad:87:bf:2a:f8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:03:41.071223 containerd[1535]: 2025-09-12 22:03:41.065 [INFO][5134] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" Namespace="calico-system" Pod="csi-node-driver-wdhc9" WorkloadEndpoint="localhost-k8s-csi--node--driver--wdhc9-eth0" Sep 12 22:03:41.103204 containerd[1535]: time="2025-09-12T22:03:41.103154895Z" level=info msg="connecting to shim 7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6" address="unix:///run/containerd/s/f5fb72823f9b8e51000cbd7694578c27d8234b2ee1522781b1518b10204cea2d" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:03:41.124137 systemd[1]: Started cri-containerd-7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6.scope - libcontainer container 7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6. Sep 12 22:03:41.136187 systemd-resolved[1352]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 12 22:03:41.150698 containerd[1535]: time="2025-09-12T22:03:41.150658148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-wdhc9,Uid:9741ef86-7513-4899-85e2-cf42c12fa232,Namespace:calico-system,Attempt:0,} returns sandbox id \"7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6\"" Sep 12 22:03:42.050426 kubelet[2665]: I0912 22:03:42.050083 2665 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:03:42.116882 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3269497560.mount: Deactivated successfully. Sep 12 22:03:42.213637 systemd[1]: Started sshd@8-10.0.0.29:22-10.0.0.1:52854.service - OpenSSH per-connection server daemon (10.0.0.1:52854). Sep 12 22:03:42.322561 sshd[5227]: Accepted publickey for core from 10.0.0.1 port 52854 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:42.325244 sshd-session[5227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:42.333568 systemd-logind[1513]: New session 9 of user core. Sep 12 22:03:42.340142 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 12 22:03:42.364097 systemd-networkd[1435]: cali1364a10672f: Gained IPv6LL Sep 12 22:03:42.393735 kubelet[2665]: I0912 22:03:42.393672 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-57d459fd7c-n9cxf" podStartSLOduration=28.548087285 podStartE2EDuration="31.39365506s" podCreationTimestamp="2025-09-12 22:03:11 +0000 UTC" firstStartedPulling="2025-09-12 22:03:37.398714997 +0000 UTC m=+42.623194472" lastFinishedPulling="2025-09-12 22:03:40.244282772 +0000 UTC m=+45.468762247" observedRunningTime="2025-09-12 22:03:41.07455161 +0000 UTC m=+46.299031085" watchObservedRunningTime="2025-09-12 22:03:42.39365506 +0000 UTC m=+47.618134535" Sep 12 22:03:42.628054 sshd[5230]: Connection closed by 10.0.0.1 port 52854 Sep 12 22:03:42.628588 sshd-session[5227]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:42.635552 systemd[1]: sshd@8-10.0.0.29:22-10.0.0.1:52854.service: Deactivated successfully. Sep 12 22:03:42.638631 systemd[1]: session-9.scope: Deactivated successfully. Sep 12 22:03:42.641795 systemd-logind[1513]: Session 9 logged out. Waiting for processes to exit. Sep 12 22:03:42.644931 systemd-logind[1513]: Removed session 9. Sep 12 22:03:42.925967 containerd[1535]: time="2025-09-12T22:03:42.925867032Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:42.926818 containerd[1535]: time="2025-09-12T22:03:42.926788116Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=61845332" Sep 12 22:03:42.927946 containerd[1535]: time="2025-09-12T22:03:42.927900890Z" level=info msg="ImageCreate event name:\"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:42.930699 containerd[1535]: time="2025-09-12T22:03:42.930666943Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:42.931321 containerd[1535]: time="2025-09-12T22:03:42.931295693Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"61845178\" in 2.399523755s" Sep 12 22:03:42.931398 containerd[1535]: time="2025-09-12T22:03:42.931324774Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:14088376331a0622b7f6a2fbc2f2932806a6eafdd7b602f6139d3b985bf1e685\"" Sep 12 22:03:42.932482 containerd[1535]: time="2025-09-12T22:03:42.932456429Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 12 22:03:42.934082 containerd[1535]: time="2025-09-12T22:03:42.934015304Z" level=info msg="CreateContainer within sandbox \"26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 12 22:03:42.942845 containerd[1535]: time="2025-09-12T22:03:42.940637622Z" level=info msg="Container a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:42.947770 containerd[1535]: time="2025-09-12T22:03:42.947727762Z" level=info msg="CreateContainer within sandbox \"26a654129d97b62eb5a75bbff6e95d68d179a17077cb5e05f2ac660724b0f2c3\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\"" Sep 12 22:03:42.949111 containerd[1535]: time="2025-09-12T22:03:42.948225626Z" level=info msg="StartContainer for \"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\"" Sep 12 22:03:42.950870 containerd[1535]: time="2025-09-12T22:03:42.950833592Z" level=info msg="connecting to shim a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14" address="unix:///run/containerd/s/7e07a61337ad3c84d6e7bec4cf3e8df5b0b2d0f32e582c3d31c03f75098f5487" protocol=ttrpc version=3 Sep 12 22:03:42.971148 systemd[1]: Started cri-containerd-a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14.scope - libcontainer container a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14. Sep 12 22:03:43.008449 containerd[1535]: time="2025-09-12T22:03:43.008238143Z" level=info msg="StartContainer for \"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" returns successfully" Sep 12 22:03:43.076689 kubelet[2665]: I0912 22:03:43.076477 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-7988f88666-l6kcq" podStartSLOduration=22.453404277 podStartE2EDuration="27.076460113s" podCreationTimestamp="2025-09-12 22:03:16 +0000 UTC" firstStartedPulling="2025-09-12 22:03:38.309213424 +0000 UTC m=+43.533692899" lastFinishedPulling="2025-09-12 22:03:42.93226926 +0000 UTC m=+48.156748735" observedRunningTime="2025-09-12 22:03:43.076255223 +0000 UTC m=+48.300734698" watchObservedRunningTime="2025-09-12 22:03:43.076460113 +0000 UTC m=+48.300939588" Sep 12 22:03:43.150319 containerd[1535]: time="2025-09-12T22:03:43.150259385Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" id:\"271dd7f1ff425ad4b6bf39e4915893fd780f043a85d7b93d4b9bb062c77acc8e\" pid:5297 exit_status:1 exited_at:{seconds:1757714623 nanos:149888607}" Sep 12 22:03:43.919282 containerd[1535]: time="2025-09-12T22:03:43.919232485Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:43.920556 containerd[1535]: time="2025-09-12T22:03:43.920525586Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8227489" Sep 12 22:03:43.921529 containerd[1535]: time="2025-09-12T22:03:43.921497111Z" level=info msg="ImageCreate event name:\"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:43.924348 containerd[1535]: time="2025-09-12T22:03:43.924311964Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:43.925675 containerd[1535]: time="2025-09-12T22:03:43.925635746Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"9596730\" in 993.148916ms" Sep 12 22:03:43.925719 containerd[1535]: time="2025-09-12T22:03:43.925683908Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:5e2b30128ce4b607acd97d3edef62ce1a90be0259903090a51c360adbe4a8f3b\"" Sep 12 22:03:43.927955 containerd[1535]: time="2025-09-12T22:03:43.927929894Z" level=info msg="CreateContainer within sandbox \"7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 12 22:03:43.936594 containerd[1535]: time="2025-09-12T22:03:43.935953632Z" level=info msg="Container 408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:43.950133 containerd[1535]: time="2025-09-12T22:03:43.950093937Z" level=info msg="CreateContainer within sandbox \"7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae\"" Sep 12 22:03:43.951216 containerd[1535]: time="2025-09-12T22:03:43.951190108Z" level=info msg="StartContainer for \"408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae\"" Sep 12 22:03:43.952824 containerd[1535]: time="2025-09-12T22:03:43.952798304Z" level=info msg="connecting to shim 408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae" address="unix:///run/containerd/s/f5fb72823f9b8e51000cbd7694578c27d8234b2ee1522781b1518b10204cea2d" protocol=ttrpc version=3 Sep 12 22:03:43.971135 systemd[1]: Started cri-containerd-408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae.scope - libcontainer container 408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae. Sep 12 22:03:44.032266 containerd[1535]: time="2025-09-12T22:03:44.032225972Z" level=info msg="StartContainer for \"408dff59321b63c555c8b74483fea634235b9b4372fd86ae33046e91098942ae\" returns successfully" Sep 12 22:03:44.033837 containerd[1535]: time="2025-09-12T22:03:44.033810965Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 12 22:03:44.137530 containerd[1535]: time="2025-09-12T22:03:44.137471065Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" id:\"1c148e531972b3ff50d877fec9a0ef42d47f0ea11d591cbced653600b7b704ff\" pid:5357 exit_status:1 exited_at:{seconds:1757714624 nanos:137132969}" Sep 12 22:03:45.151024 containerd[1535]: time="2025-09-12T22:03:45.150507088Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" id:\"3bd432561d1ad5d6b2f26e1b4709808419e013ce8a487613e615ee0b8954fcad\" pid:5387 exited_at:{seconds:1757714625 nanos:149283473}" Sep 12 22:03:45.159077 containerd[1535]: time="2025-09-12T22:03:45.159045914Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:45.160484 containerd[1535]: time="2025-09-12T22:03:45.160460778Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=13761208" Sep 12 22:03:45.161376 containerd[1535]: time="2025-09-12T22:03:45.161352539Z" level=info msg="ImageCreate event name:\"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:45.163565 containerd[1535]: time="2025-09-12T22:03:45.163496916Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:03:45.164082 containerd[1535]: time="2025-09-12T22:03:45.164055061Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"15130401\" in 1.130208614s" Sep 12 22:03:45.164127 containerd[1535]: time="2025-09-12T22:03:45.164090982Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:a319b5bdc1001e98875b68e2943279adb74bcb19d09f1db857bc27959a078a65\"" Sep 12 22:03:45.168638 containerd[1535]: time="2025-09-12T22:03:45.168512622Z" level=info msg="CreateContainer within sandbox \"7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 12 22:03:45.179127 containerd[1535]: time="2025-09-12T22:03:45.179092341Z" level=info msg="Container 9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:03:45.189380 containerd[1535]: time="2025-09-12T22:03:45.189342245Z" level=info msg="CreateContainer within sandbox \"7df22f63ffb61c70f320abb14d10ec526111e535a7b656fa10467423d11ca5f6\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636\"" Sep 12 22:03:45.190040 containerd[1535]: time="2025-09-12T22:03:45.190013755Z" level=info msg="StartContainer for \"9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636\"" Sep 12 22:03:45.191392 containerd[1535]: time="2025-09-12T22:03:45.191366736Z" level=info msg="connecting to shim 9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636" address="unix:///run/containerd/s/f5fb72823f9b8e51000cbd7694578c27d8234b2ee1522781b1518b10204cea2d" protocol=ttrpc version=3 Sep 12 22:03:45.212263 systemd[1]: Started cri-containerd-9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636.scope - libcontainer container 9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636. Sep 12 22:03:45.264055 containerd[1535]: time="2025-09-12T22:03:45.263916578Z" level=info msg="StartContainer for \"9017038cf004b2c1ecd5df8db3d9746b4fca99f705792cfc302e20c7a5032636\" returns successfully" Sep 12 22:03:45.948957 kubelet[2665]: I0912 22:03:45.948903 2665 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 12 22:03:45.948957 kubelet[2665]: I0912 22:03:45.948952 2665 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 12 22:03:46.081008 kubelet[2665]: I0912 22:03:46.080850 2665 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-wdhc9" podStartSLOduration=26.068276916 podStartE2EDuration="30.080833105s" podCreationTimestamp="2025-09-12 22:03:16 +0000 UTC" firstStartedPulling="2025-09-12 22:03:41.152077458 +0000 UTC m=+46.376556893" lastFinishedPulling="2025-09-12 22:03:45.164633607 +0000 UTC m=+50.389113082" observedRunningTime="2025-09-12 22:03:46.080610335 +0000 UTC m=+51.305089810" watchObservedRunningTime="2025-09-12 22:03:46.080833105 +0000 UTC m=+51.305312580" Sep 12 22:03:47.644462 systemd[1]: Started sshd@9-10.0.0.29:22-10.0.0.1:52858.service - OpenSSH per-connection server daemon (10.0.0.1:52858). Sep 12 22:03:47.727050 sshd[5444]: Accepted publickey for core from 10.0.0.1 port 52858 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:47.730831 sshd-session[5444]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:47.735540 systemd-logind[1513]: New session 10 of user core. Sep 12 22:03:47.746180 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 12 22:03:48.005839 sshd[5447]: Connection closed by 10.0.0.1 port 52858 Sep 12 22:03:48.010173 sshd-session[5444]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:48.020774 systemd[1]: sshd@9-10.0.0.29:22-10.0.0.1:52858.service: Deactivated successfully. Sep 12 22:03:48.023760 systemd[1]: session-10.scope: Deactivated successfully. Sep 12 22:03:48.025014 systemd-logind[1513]: Session 10 logged out. Waiting for processes to exit. Sep 12 22:03:48.027494 systemd[1]: Started sshd@10-10.0.0.29:22-10.0.0.1:52872.service - OpenSSH per-connection server daemon (10.0.0.1:52872). Sep 12 22:03:48.029508 systemd-logind[1513]: Removed session 10. Sep 12 22:03:48.093965 sshd[5462]: Accepted publickey for core from 10.0.0.1 port 52872 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:48.095332 sshd-session[5462]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:48.099730 systemd-logind[1513]: New session 11 of user core. Sep 12 22:03:48.107395 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 12 22:03:48.304693 sshd[5465]: Connection closed by 10.0.0.1 port 52872 Sep 12 22:03:48.305328 sshd-session[5462]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:48.319907 systemd[1]: sshd@10-10.0.0.29:22-10.0.0.1:52872.service: Deactivated successfully. Sep 12 22:03:48.324070 systemd[1]: session-11.scope: Deactivated successfully. Sep 12 22:03:48.325815 systemd-logind[1513]: Session 11 logged out. Waiting for processes to exit. Sep 12 22:03:48.331381 systemd[1]: Started sshd@11-10.0.0.29:22-10.0.0.1:52880.service - OpenSSH per-connection server daemon (10.0.0.1:52880). Sep 12 22:03:48.332336 systemd-logind[1513]: Removed session 11. Sep 12 22:03:48.384841 sshd[5477]: Accepted publickey for core from 10.0.0.1 port 52880 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:48.386324 sshd-session[5477]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:48.390706 systemd-logind[1513]: New session 12 of user core. Sep 12 22:03:48.403179 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 12 22:03:48.559687 sshd[5480]: Connection closed by 10.0.0.1 port 52880 Sep 12 22:03:48.559997 sshd-session[5477]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:48.565194 systemd-logind[1513]: Session 12 logged out. Waiting for processes to exit. Sep 12 22:03:48.566601 systemd[1]: sshd@11-10.0.0.29:22-10.0.0.1:52880.service: Deactivated successfully. Sep 12 22:03:48.569250 systemd[1]: session-12.scope: Deactivated successfully. Sep 12 22:03:48.570822 systemd-logind[1513]: Removed session 12. Sep 12 22:03:51.726017 kubelet[2665]: I0912 22:03:51.725870 2665 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:03:53.575110 systemd[1]: Started sshd@12-10.0.0.29:22-10.0.0.1:51170.service - OpenSSH per-connection server daemon (10.0.0.1:51170). Sep 12 22:03:53.642428 sshd[5503]: Accepted publickey for core from 10.0.0.1 port 51170 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:53.644168 sshd-session[5503]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:53.648798 systemd-logind[1513]: New session 13 of user core. Sep 12 22:03:53.656148 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 12 22:03:53.825120 sshd[5506]: Connection closed by 10.0.0.1 port 51170 Sep 12 22:03:53.825712 sshd-session[5503]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:53.830065 systemd-logind[1513]: Session 13 logged out. Waiting for processes to exit. Sep 12 22:03:53.830267 systemd[1]: sshd@12-10.0.0.29:22-10.0.0.1:51170.service: Deactivated successfully. Sep 12 22:03:53.833612 systemd[1]: session-13.scope: Deactivated successfully. Sep 12 22:03:53.835063 systemd-logind[1513]: Removed session 13. Sep 12 22:03:53.921214 containerd[1535]: time="2025-09-12T22:03:53.921167992Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" id:\"21b5f94677be57014882619ab3324e707e5b0a84485cd1fae72028dfc22f7a18\" pid:5532 exited_at:{seconds:1757714633 nanos:920760295}" Sep 12 22:03:53.997818 containerd[1535]: time="2025-09-12T22:03:53.997753649Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca\" id:\"a7c86c957039365c7558efa32f40af72c08cdfdb1e9ea430d1f584204efeee0c\" pid:5555 exited_at:{seconds:1757714633 nanos:997362913}" Sep 12 22:03:58.840527 systemd[1]: Started sshd@13-10.0.0.29:22-10.0.0.1:51184.service - OpenSSH per-connection server daemon (10.0.0.1:51184). Sep 12 22:03:58.901780 sshd[5575]: Accepted publickey for core from 10.0.0.1 port 51184 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:03:58.903691 sshd-session[5575]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:03:58.908045 systemd-logind[1513]: New session 14 of user core. Sep 12 22:03:58.917181 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 12 22:03:59.069640 sshd[5578]: Connection closed by 10.0.0.1 port 51184 Sep 12 22:03:59.073020 sshd-session[5575]: pam_unix(sshd:session): session closed for user core Sep 12 22:03:59.076536 systemd-logind[1513]: Session 14 logged out. Waiting for processes to exit. Sep 12 22:03:59.077111 systemd[1]: sshd@13-10.0.0.29:22-10.0.0.1:51184.service: Deactivated successfully. Sep 12 22:03:59.079139 systemd[1]: session-14.scope: Deactivated successfully. Sep 12 22:03:59.082310 systemd-logind[1513]: Removed session 14. Sep 12 22:04:04.085613 systemd[1]: Started sshd@14-10.0.0.29:22-10.0.0.1:60464.service - OpenSSH per-connection server daemon (10.0.0.1:60464). Sep 12 22:04:04.158818 sshd[5598]: Accepted publickey for core from 10.0.0.1 port 60464 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:04.160334 sshd-session[5598]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:04.165060 systemd-logind[1513]: New session 15 of user core. Sep 12 22:04:04.177166 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 12 22:04:04.329429 sshd[5601]: Connection closed by 10.0.0.1 port 60464 Sep 12 22:04:04.330232 sshd-session[5598]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:04.335460 systemd-logind[1513]: Session 15 logged out. Waiting for processes to exit. Sep 12 22:04:04.335963 systemd[1]: sshd@14-10.0.0.29:22-10.0.0.1:60464.service: Deactivated successfully. Sep 12 22:04:04.338508 systemd[1]: session-15.scope: Deactivated successfully. Sep 12 22:04:04.340429 systemd-logind[1513]: Removed session 15. Sep 12 22:04:04.593295 containerd[1535]: time="2025-09-12T22:04:04.593152514Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1039365f3620bc1d6c349d84ef2e15952a8c784b3dac6cc5b3a3a1d7d795a20f\" id:\"c7891dd9cccd88157e2ff584c7d0214617d8355e25ccc1a8649bd50b7ada4532\" pid:5626 exited_at:{seconds:1757714644 nanos:592810564}" Sep 12 22:04:09.346321 systemd[1]: Started sshd@15-10.0.0.29:22-10.0.0.1:60480.service - OpenSSH per-connection server daemon (10.0.0.1:60480). Sep 12 22:04:09.414195 sshd[5641]: Accepted publickey for core from 10.0.0.1 port 60480 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:09.415565 sshd-session[5641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:09.419692 systemd-logind[1513]: New session 16 of user core. Sep 12 22:04:09.432387 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 12 22:04:09.562578 sshd[5644]: Connection closed by 10.0.0.1 port 60480 Sep 12 22:04:09.562919 sshd-session[5641]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:09.583485 systemd[1]: sshd@15-10.0.0.29:22-10.0.0.1:60480.service: Deactivated successfully. Sep 12 22:04:09.586478 systemd[1]: session-16.scope: Deactivated successfully. Sep 12 22:04:09.588221 systemd-logind[1513]: Session 16 logged out. Waiting for processes to exit. Sep 12 22:04:09.590550 systemd[1]: Started sshd@16-10.0.0.29:22-10.0.0.1:60494.service - OpenSSH per-connection server daemon (10.0.0.1:60494). Sep 12 22:04:09.591103 systemd-logind[1513]: Removed session 16. Sep 12 22:04:09.646486 sshd[5658]: Accepted publickey for core from 10.0.0.1 port 60494 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:09.647857 sshd-session[5658]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:09.652731 systemd-logind[1513]: New session 17 of user core. Sep 12 22:04:09.664114 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 12 22:04:09.880113 sshd[5661]: Connection closed by 10.0.0.1 port 60494 Sep 12 22:04:09.880846 sshd-session[5658]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:09.892493 systemd[1]: sshd@16-10.0.0.29:22-10.0.0.1:60494.service: Deactivated successfully. Sep 12 22:04:09.894280 systemd[1]: session-17.scope: Deactivated successfully. Sep 12 22:04:09.895100 systemd-logind[1513]: Session 17 logged out. Waiting for processes to exit. Sep 12 22:04:09.897505 systemd[1]: Started sshd@17-10.0.0.29:22-10.0.0.1:60504.service - OpenSSH per-connection server daemon (10.0.0.1:60504). Sep 12 22:04:09.899660 systemd-logind[1513]: Removed session 17. Sep 12 22:04:09.957310 sshd[5673]: Accepted publickey for core from 10.0.0.1 port 60504 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:09.958792 sshd-session[5673]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:09.963402 systemd-logind[1513]: New session 18 of user core. Sep 12 22:04:09.981186 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 12 22:04:11.598233 sshd[5676]: Connection closed by 10.0.0.1 port 60504 Sep 12 22:04:11.599203 sshd-session[5673]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:11.612163 systemd[1]: sshd@17-10.0.0.29:22-10.0.0.1:60504.service: Deactivated successfully. Sep 12 22:04:11.615156 systemd[1]: session-18.scope: Deactivated successfully. Sep 12 22:04:11.616279 systemd[1]: session-18.scope: Consumed 574ms CPU time, 74.9M memory peak. Sep 12 22:04:11.618177 systemd-logind[1513]: Session 18 logged out. Waiting for processes to exit. Sep 12 22:04:11.624891 systemd[1]: Started sshd@18-10.0.0.29:22-10.0.0.1:54028.service - OpenSSH per-connection server daemon (10.0.0.1:54028). Sep 12 22:04:11.629049 systemd-logind[1513]: Removed session 18. Sep 12 22:04:11.687479 sshd[5697]: Accepted publickey for core from 10.0.0.1 port 54028 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:11.688918 sshd-session[5697]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:11.693156 systemd-logind[1513]: New session 19 of user core. Sep 12 22:04:11.704167 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 12 22:04:11.977738 sshd[5700]: Connection closed by 10.0.0.1 port 54028 Sep 12 22:04:11.977175 sshd-session[5697]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:11.989408 systemd[1]: sshd@18-10.0.0.29:22-10.0.0.1:54028.service: Deactivated successfully. Sep 12 22:04:11.991028 systemd[1]: session-19.scope: Deactivated successfully. Sep 12 22:04:11.992851 systemd-logind[1513]: Session 19 logged out. Waiting for processes to exit. Sep 12 22:04:11.996246 systemd[1]: Started sshd@19-10.0.0.29:22-10.0.0.1:54036.service - OpenSSH per-connection server daemon (10.0.0.1:54036). Sep 12 22:04:11.997593 systemd-logind[1513]: Removed session 19. Sep 12 22:04:12.058103 sshd[5711]: Accepted publickey for core from 10.0.0.1 port 54036 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:12.059405 sshd-session[5711]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:12.064036 systemd-logind[1513]: New session 20 of user core. Sep 12 22:04:12.075139 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 12 22:04:12.213578 sshd[5714]: Connection closed by 10.0.0.1 port 54036 Sep 12 22:04:12.213906 sshd-session[5711]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:12.217663 systemd[1]: sshd@19-10.0.0.29:22-10.0.0.1:54036.service: Deactivated successfully. Sep 12 22:04:12.221474 systemd[1]: session-20.scope: Deactivated successfully. Sep 12 22:04:12.222191 systemd-logind[1513]: Session 20 logged out. Waiting for processes to exit. Sep 12 22:04:12.223498 systemd-logind[1513]: Removed session 20. Sep 12 22:04:17.232240 systemd[1]: Started sshd@20-10.0.0.29:22-10.0.0.1:54040.service - OpenSSH per-connection server daemon (10.0.0.1:54040). Sep 12 22:04:17.288062 sshd[5737]: Accepted publickey for core from 10.0.0.1 port 54040 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:17.289528 sshd-session[5737]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:17.294036 systemd-logind[1513]: New session 21 of user core. Sep 12 22:04:17.301146 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 12 22:04:17.427285 sshd[5740]: Connection closed by 10.0.0.1 port 54040 Sep 12 22:04:17.427986 sshd-session[5737]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:17.433770 systemd[1]: sshd@20-10.0.0.29:22-10.0.0.1:54040.service: Deactivated successfully. Sep 12 22:04:17.438353 systemd[1]: session-21.scope: Deactivated successfully. Sep 12 22:04:17.441118 systemd-logind[1513]: Session 21 logged out. Waiting for processes to exit. Sep 12 22:04:17.442664 systemd-logind[1513]: Removed session 21. Sep 12 22:04:22.439307 systemd[1]: Started sshd@21-10.0.0.29:22-10.0.0.1:35980.service - OpenSSH per-connection server daemon (10.0.0.1:35980). Sep 12 22:04:22.511096 sshd[5757]: Accepted publickey for core from 10.0.0.1 port 35980 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:22.512406 sshd-session[5757]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:22.515954 systemd-logind[1513]: New session 22 of user core. Sep 12 22:04:22.522235 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 12 22:04:22.592809 containerd[1535]: time="2025-09-12T22:04:22.592765281Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" id:\"d190158b74edaeb457b0acfc43cc7b0e6c6f0cd0ca6e81f3278ccb4ea9f58755\" pid:5773 exited_at:{seconds:1757714662 nanos:592460602}" Sep 12 22:04:22.694218 sshd[5762]: Connection closed by 10.0.0.1 port 35980 Sep 12 22:04:22.694933 sshd-session[5757]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:22.700096 systemd[1]: sshd@21-10.0.0.29:22-10.0.0.1:35980.service: Deactivated successfully. Sep 12 22:04:22.701698 systemd[1]: session-22.scope: Deactivated successfully. Sep 12 22:04:22.702320 systemd-logind[1513]: Session 22 logged out. Waiting for processes to exit. Sep 12 22:04:22.703582 systemd-logind[1513]: Removed session 22. Sep 12 22:04:22.863264 kubelet[2665]: E0912 22:04:22.863178 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:04:23.863038 kubelet[2665]: E0912 22:04:23.862954 2665 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 12 22:04:23.948120 containerd[1535]: time="2025-09-12T22:04:23.948070057Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a51c5f717a73d152ae09e60690b0dcb18245652e31ca045952b82d270739db14\" id:\"c1fb4997518a78bc110dfd5c4dbc6c73722304041c9fa77db4672729aa99c3f4\" pid:5808 exited_at:{seconds:1757714663 nanos:947496340}" Sep 12 22:04:23.995963 containerd[1535]: time="2025-09-12T22:04:23.995924111Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2a8e8e9b32b02c12e23e0127d7b1b467fcd818728b77f2bfad4400fe0de2d4ca\" id:\"69c234a63f62f15ee042467dc312e6d4f84954b3dfc6553f02ae95d76c7278f8\" pid:5832 exited_at:{seconds:1757714663 nanos:995704392}" Sep 12 22:04:27.710911 systemd[1]: Started sshd@22-10.0.0.29:22-10.0.0.1:35996.service - OpenSSH per-connection server daemon (10.0.0.1:35996). Sep 12 22:04:27.779281 sshd[5843]: Accepted publickey for core from 10.0.0.1 port 35996 ssh2: RSA SHA256:Yqy+ciIRp9tS6RxmRMX9+tv4H/mrc+u7L29C7Pz/5UI Sep 12 22:04:27.780656 sshd-session[5843]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:04:27.784969 systemd-logind[1513]: New session 23 of user core. Sep 12 22:04:27.792121 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 12 22:04:28.015311 sshd[5846]: Connection closed by 10.0.0.1 port 35996 Sep 12 22:04:28.018056 sshd-session[5843]: pam_unix(sshd:session): session closed for user core Sep 12 22:04:28.023223 systemd[1]: sshd@22-10.0.0.29:22-10.0.0.1:35996.service: Deactivated successfully. Sep 12 22:04:28.025627 systemd[1]: session-23.scope: Deactivated successfully. Sep 12 22:04:28.026300 systemd-logind[1513]: Session 23 logged out. Waiting for processes to exit. Sep 12 22:04:28.027685 systemd-logind[1513]: Removed session 23.