Jul 12 09:28:34.773188 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Jul 12 09:28:34.773209 kernel: Linux version 6.12.36-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT Sat Jul 12 08:24:03 -00 2025 Jul 12 09:28:34.773218 kernel: KASLR enabled Jul 12 09:28:34.773224 kernel: efi: EFI v2.7 by EDK II Jul 12 09:28:34.773229 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb832018 ACPI 2.0=0xdbfd0018 RNG=0xdbfd0a18 MEMRESERVE=0xdb838218 Jul 12 09:28:34.773235 kernel: random: crng init done Jul 12 09:28:34.773241 kernel: secureboot: Secure boot disabled Jul 12 09:28:34.773247 kernel: ACPI: Early table checksum verification disabled Jul 12 09:28:34.773253 kernel: ACPI: RSDP 0x00000000DBFD0018 000024 (v02 BOCHS ) Jul 12 09:28:34.773260 kernel: ACPI: XSDT 0x00000000DBFD0F18 000064 (v01 BOCHS BXPC 00000001 01000013) Jul 12 09:28:34.773265 kernel: ACPI: FACP 0x00000000DBFD0B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773271 kernel: ACPI: DSDT 0x00000000DBF0E018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773277 kernel: ACPI: APIC 0x00000000DBFD0C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773283 kernel: ACPI: PPTT 0x00000000DBFD0098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773289 kernel: ACPI: GTDT 0x00000000DBFD0818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773297 kernel: ACPI: MCFG 0x00000000DBFD0A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773303 kernel: ACPI: SPCR 0x00000000DBFD0918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773309 kernel: ACPI: DBG2 0x00000000DBFD0998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773315 kernel: ACPI: IORT 0x00000000DBFD0198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jul 12 09:28:34.773321 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Jul 12 09:28:34.773327 kernel: ACPI: Use ACPI SPCR as default console: Yes Jul 12 09:28:34.773333 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Jul 12 09:28:34.773339 kernel: NODE_DATA(0) allocated [mem 0xdc965a00-0xdc96cfff] Jul 12 09:28:34.773345 kernel: Zone ranges: Jul 12 09:28:34.773351 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Jul 12 09:28:34.773358 kernel: DMA32 empty Jul 12 09:28:34.773363 kernel: Normal empty Jul 12 09:28:34.773369 kernel: Device empty Jul 12 09:28:34.773375 kernel: Movable zone start for each node Jul 12 09:28:34.773381 kernel: Early memory node ranges Jul 12 09:28:34.773387 kernel: node 0: [mem 0x0000000040000000-0x00000000db81ffff] Jul 12 09:28:34.773393 kernel: node 0: [mem 0x00000000db820000-0x00000000db82ffff] Jul 12 09:28:34.773399 kernel: node 0: [mem 0x00000000db830000-0x00000000dc09ffff] Jul 12 09:28:34.773405 kernel: node 0: [mem 0x00000000dc0a0000-0x00000000dc2dffff] Jul 12 09:28:34.773411 kernel: node 0: [mem 0x00000000dc2e0000-0x00000000dc36ffff] Jul 12 09:28:34.773416 kernel: node 0: [mem 0x00000000dc370000-0x00000000dc45ffff] Jul 12 09:28:34.773422 kernel: node 0: [mem 0x00000000dc460000-0x00000000dc52ffff] Jul 12 09:28:34.773429 kernel: node 0: [mem 0x00000000dc530000-0x00000000dc5cffff] Jul 12 09:28:34.773435 kernel: node 0: [mem 0x00000000dc5d0000-0x00000000dce1ffff] Jul 12 09:28:34.773441 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Jul 12 09:28:34.773450 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Jul 12 09:28:34.773456 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Jul 12 09:28:34.773463 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Jul 12 09:28:34.773471 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Jul 12 09:28:34.773477 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Jul 12 09:28:34.773483 kernel: cma: Reserved 16 MiB at 0x00000000d8000000 on node -1 Jul 12 09:28:34.773490 kernel: psci: probing for conduit method from ACPI. Jul 12 09:28:34.773496 kernel: psci: PSCIv1.1 detected in firmware. Jul 12 09:28:34.773502 kernel: psci: Using standard PSCI v0.2 function IDs Jul 12 09:28:34.773508 kernel: psci: Trusted OS migration not required Jul 12 09:28:34.773515 kernel: psci: SMC Calling Convention v1.1 Jul 12 09:28:34.773521 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Jul 12 09:28:34.773528 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Jul 12 09:28:34.773535 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Jul 12 09:28:34.773542 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Jul 12 09:28:34.773548 kernel: Detected PIPT I-cache on CPU0 Jul 12 09:28:34.773554 kernel: CPU features: detected: GIC system register CPU interface Jul 12 09:28:34.773561 kernel: CPU features: detected: Spectre-v4 Jul 12 09:28:34.773567 kernel: CPU features: detected: Spectre-BHB Jul 12 09:28:34.773573 kernel: CPU features: kernel page table isolation forced ON by KASLR Jul 12 09:28:34.773580 kernel: CPU features: detected: Kernel page table isolation (KPTI) Jul 12 09:28:34.773586 kernel: CPU features: detected: ARM erratum 1418040 Jul 12 09:28:34.773592 kernel: CPU features: detected: SSBS not fully self-synchronizing Jul 12 09:28:34.773598 kernel: alternatives: applying boot alternatives Jul 12 09:28:34.773606 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=2eed6122ab9d95fa96c8f5511b96c1220a0caf18bbf7b84035ef573d9ba90496 Jul 12 09:28:34.773613 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 12 09:28:34.773620 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jul 12 09:28:34.773627 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 12 09:28:34.773633 kernel: Fallback order for Node 0: 0 Jul 12 09:28:34.773639 kernel: Built 1 zonelists, mobility grouping on. Total pages: 643072 Jul 12 09:28:34.773646 kernel: Policy zone: DMA Jul 12 09:28:34.773652 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 12 09:28:34.773658 kernel: software IO TLB: SWIOTLB bounce buffer size adjusted to 2MB Jul 12 09:28:34.773664 kernel: software IO TLB: area num 4. Jul 12 09:28:34.773671 kernel: software IO TLB: SWIOTLB bounce buffer size roundup to 4MB Jul 12 09:28:34.773677 kernel: software IO TLB: mapped [mem 0x00000000d7c00000-0x00000000d8000000] (4MB) Jul 12 09:28:34.773685 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jul 12 09:28:34.773691 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 12 09:28:34.773698 kernel: rcu: RCU event tracing is enabled. Jul 12 09:28:34.773704 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jul 12 09:28:34.773711 kernel: Trampoline variant of Tasks RCU enabled. Jul 12 09:28:34.773717 kernel: Tracing variant of Tasks RCU enabled. Jul 12 09:28:34.773723 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 12 09:28:34.773730 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jul 12 09:28:34.773736 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jul 12 09:28:34.773743 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jul 12 09:28:34.773749 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Jul 12 09:28:34.773757 kernel: GICv3: 256 SPIs implemented Jul 12 09:28:34.773763 kernel: GICv3: 0 Extended SPIs implemented Jul 12 09:28:34.773770 kernel: Root IRQ handler: gic_handle_irq Jul 12 09:28:34.773776 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Jul 12 09:28:34.773782 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Jul 12 09:28:34.773789 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Jul 12 09:28:34.773795 kernel: ITS [mem 0x08080000-0x0809ffff] Jul 12 09:28:34.773801 kernel: ITS@0x0000000008080000: allocated 8192 Devices @40110000 (indirect, esz 8, psz 64K, shr 1) Jul 12 09:28:34.773808 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @40120000 (flat, esz 8, psz 64K, shr 1) Jul 12 09:28:34.773814 kernel: GICv3: using LPI property table @0x0000000040130000 Jul 12 09:28:34.773821 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040140000 Jul 12 09:28:34.773827 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jul 12 09:28:34.773835 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jul 12 09:28:34.773841 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Jul 12 09:28:34.773848 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Jul 12 09:28:34.773854 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Jul 12 09:28:34.773861 kernel: arm-pv: using stolen time PV Jul 12 09:28:34.773867 kernel: Console: colour dummy device 80x25 Jul 12 09:28:34.773874 kernel: ACPI: Core revision 20240827 Jul 12 09:28:34.773881 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Jul 12 09:28:34.773887 kernel: pid_max: default: 32768 minimum: 301 Jul 12 09:28:34.773894 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jul 12 09:28:34.773901 kernel: landlock: Up and running. Jul 12 09:28:34.773908 kernel: SELinux: Initializing. Jul 12 09:28:34.773934 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 12 09:28:34.773941 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 12 09:28:34.773947 kernel: rcu: Hierarchical SRCU implementation. Jul 12 09:28:34.773955 kernel: rcu: Max phase no-delay instances is 400. Jul 12 09:28:34.773961 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jul 12 09:28:34.773968 kernel: Remapping and enabling EFI services. Jul 12 09:28:34.773974 kernel: smp: Bringing up secondary CPUs ... Jul 12 09:28:34.773987 kernel: Detected PIPT I-cache on CPU1 Jul 12 09:28:34.773994 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Jul 12 09:28:34.774001 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040150000 Jul 12 09:28:34.774009 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jul 12 09:28:34.774021 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Jul 12 09:28:34.774028 kernel: Detected PIPT I-cache on CPU2 Jul 12 09:28:34.774036 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Jul 12 09:28:34.774043 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040160000 Jul 12 09:28:34.774051 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jul 12 09:28:34.774058 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Jul 12 09:28:34.774065 kernel: Detected PIPT I-cache on CPU3 Jul 12 09:28:34.774072 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Jul 12 09:28:34.774079 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040170000 Jul 12 09:28:34.774085 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jul 12 09:28:34.774092 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Jul 12 09:28:34.774099 kernel: smp: Brought up 1 node, 4 CPUs Jul 12 09:28:34.774106 kernel: SMP: Total of 4 processors activated. Jul 12 09:28:34.774114 kernel: CPU: All CPU(s) started at EL1 Jul 12 09:28:34.774121 kernel: CPU features: detected: 32-bit EL0 Support Jul 12 09:28:34.774127 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Jul 12 09:28:34.774134 kernel: CPU features: detected: Common not Private translations Jul 12 09:28:34.774141 kernel: CPU features: detected: CRC32 instructions Jul 12 09:28:34.774148 kernel: CPU features: detected: Enhanced Virtualization Traps Jul 12 09:28:34.774155 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Jul 12 09:28:34.774162 kernel: CPU features: detected: LSE atomic instructions Jul 12 09:28:34.774168 kernel: CPU features: detected: Privileged Access Never Jul 12 09:28:34.774176 kernel: CPU features: detected: RAS Extension Support Jul 12 09:28:34.774183 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Jul 12 09:28:34.774190 kernel: alternatives: applying system-wide alternatives Jul 12 09:28:34.774197 kernel: CPU features: detected: Hardware dirty bit management on CPU0-3 Jul 12 09:28:34.774205 kernel: Memory: 2424032K/2572288K available (11136K kernel code, 2436K rwdata, 9056K rodata, 39424K init, 1038K bss, 125920K reserved, 16384K cma-reserved) Jul 12 09:28:34.774212 kernel: devtmpfs: initialized Jul 12 09:28:34.774218 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 12 09:28:34.774225 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jul 12 09:28:34.774232 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Jul 12 09:28:34.774240 kernel: 0 pages in range for non-PLT usage Jul 12 09:28:34.774247 kernel: 508448 pages in range for PLT usage Jul 12 09:28:34.774254 kernel: pinctrl core: initialized pinctrl subsystem Jul 12 09:28:34.774260 kernel: SMBIOS 3.0.0 present. Jul 12 09:28:34.774267 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Jul 12 09:28:34.774274 kernel: DMI: Memory slots populated: 1/1 Jul 12 09:28:34.774281 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 12 09:28:34.774288 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Jul 12 09:28:34.774295 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Jul 12 09:28:34.774303 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Jul 12 09:28:34.774310 kernel: audit: initializing netlink subsys (disabled) Jul 12 09:28:34.774317 kernel: audit: type=2000 audit(0.020:1): state=initialized audit_enabled=0 res=1 Jul 12 09:28:34.774324 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 12 09:28:34.774331 kernel: cpuidle: using governor menu Jul 12 09:28:34.774338 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Jul 12 09:28:34.774345 kernel: ASID allocator initialised with 32768 entries Jul 12 09:28:34.774351 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 12 09:28:34.774358 kernel: Serial: AMBA PL011 UART driver Jul 12 09:28:34.774366 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 12 09:28:34.774373 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Jul 12 09:28:34.774380 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Jul 12 09:28:34.774387 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Jul 12 09:28:34.774394 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 12 09:28:34.774401 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Jul 12 09:28:34.774407 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Jul 12 09:28:34.774414 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Jul 12 09:28:34.774421 kernel: ACPI: Added _OSI(Module Device) Jul 12 09:28:34.774429 kernel: ACPI: Added _OSI(Processor Device) Jul 12 09:28:34.774436 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 12 09:28:34.774442 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jul 12 09:28:34.774449 kernel: ACPI: Interpreter enabled Jul 12 09:28:34.774456 kernel: ACPI: Using GIC for interrupt routing Jul 12 09:28:34.774463 kernel: ACPI: MCFG table detected, 1 entries Jul 12 09:28:34.774469 kernel: ACPI: CPU0 has been hot-added Jul 12 09:28:34.774476 kernel: ACPI: CPU1 has been hot-added Jul 12 09:28:34.774483 kernel: ACPI: CPU2 has been hot-added Jul 12 09:28:34.774490 kernel: ACPI: CPU3 has been hot-added Jul 12 09:28:34.774498 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Jul 12 09:28:34.774505 kernel: printk: legacy console [ttyAMA0] enabled Jul 12 09:28:34.774512 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jul 12 09:28:34.774635 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jul 12 09:28:34.774699 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Jul 12 09:28:34.774757 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Jul 12 09:28:34.774814 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Jul 12 09:28:34.774872 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Jul 12 09:28:34.774881 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Jul 12 09:28:34.774888 kernel: PCI host bridge to bus 0000:00 Jul 12 09:28:34.774976 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Jul 12 09:28:34.775043 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Jul 12 09:28:34.775097 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Jul 12 09:28:34.775150 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jul 12 09:28:34.775230 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Jul 12 09:28:34.775300 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Jul 12 09:28:34.775360 kernel: pci 0000:00:01.0: BAR 0 [io 0x0000-0x001f] Jul 12 09:28:34.775425 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff] Jul 12 09:28:34.775483 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Jul 12 09:28:34.775541 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Jul 12 09:28:34.775599 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff]: assigned Jul 12 09:28:34.775661 kernel: pci 0000:00:01.0: BAR 0 [io 0x1000-0x101f]: assigned Jul 12 09:28:34.775713 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Jul 12 09:28:34.775764 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Jul 12 09:28:34.775816 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Jul 12 09:28:34.775825 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Jul 12 09:28:34.775833 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Jul 12 09:28:34.775840 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Jul 12 09:28:34.775848 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Jul 12 09:28:34.775855 kernel: iommu: Default domain type: Translated Jul 12 09:28:34.775862 kernel: iommu: DMA domain TLB invalidation policy: strict mode Jul 12 09:28:34.775869 kernel: efivars: Registered efivars operations Jul 12 09:28:34.775876 kernel: vgaarb: loaded Jul 12 09:28:34.775883 kernel: clocksource: Switched to clocksource arch_sys_counter Jul 12 09:28:34.775890 kernel: VFS: Disk quotas dquot_6.6.0 Jul 12 09:28:34.775897 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 12 09:28:34.775904 kernel: pnp: PnP ACPI init Jul 12 09:28:34.775994 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Jul 12 09:28:34.776004 kernel: pnp: PnP ACPI: found 1 devices Jul 12 09:28:34.776017 kernel: NET: Registered PF_INET protocol family Jul 12 09:28:34.776025 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jul 12 09:28:34.776032 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jul 12 09:28:34.776039 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 12 09:28:34.776046 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 12 09:28:34.776053 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jul 12 09:28:34.776062 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jul 12 09:28:34.776069 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 12 09:28:34.776076 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 12 09:28:34.776083 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 12 09:28:34.776090 kernel: PCI: CLS 0 bytes, default 64 Jul 12 09:28:34.776097 kernel: kvm [1]: HYP mode not available Jul 12 09:28:34.776103 kernel: Initialise system trusted keyrings Jul 12 09:28:34.776110 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jul 12 09:28:34.776117 kernel: Key type asymmetric registered Jul 12 09:28:34.776124 kernel: Asymmetric key parser 'x509' registered Jul 12 09:28:34.776132 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Jul 12 09:28:34.776139 kernel: io scheduler mq-deadline registered Jul 12 09:28:34.776146 kernel: io scheduler kyber registered Jul 12 09:28:34.776153 kernel: io scheduler bfq registered Jul 12 09:28:34.776160 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Jul 12 09:28:34.776167 kernel: ACPI: button: Power Button [PWRB] Jul 12 09:28:34.776174 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Jul 12 09:28:34.776237 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Jul 12 09:28:34.776247 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 12 09:28:34.776256 kernel: thunder_xcv, ver 1.0 Jul 12 09:28:34.776262 kernel: thunder_bgx, ver 1.0 Jul 12 09:28:34.776269 kernel: nicpf, ver 1.0 Jul 12 09:28:34.776276 kernel: nicvf, ver 1.0 Jul 12 09:28:34.776343 kernel: rtc-efi rtc-efi.0: registered as rtc0 Jul 12 09:28:34.776399 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-07-12T09:28:34 UTC (1752312514) Jul 12 09:28:34.776408 kernel: hid: raw HID events driver (C) Jiri Kosina Jul 12 09:28:34.776415 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Jul 12 09:28:34.776424 kernel: watchdog: NMI not fully supported Jul 12 09:28:34.776431 kernel: watchdog: Hard watchdog permanently disabled Jul 12 09:28:34.776438 kernel: NET: Registered PF_INET6 protocol family Jul 12 09:28:34.776445 kernel: Segment Routing with IPv6 Jul 12 09:28:34.776451 kernel: In-situ OAM (IOAM) with IPv6 Jul 12 09:28:34.776459 kernel: NET: Registered PF_PACKET protocol family Jul 12 09:28:34.776466 kernel: Key type dns_resolver registered Jul 12 09:28:34.776472 kernel: registered taskstats version 1 Jul 12 09:28:34.776479 kernel: Loading compiled-in X.509 certificates Jul 12 09:28:34.776487 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.36-flatcar: 5833903fd926e330df1283c2ccd9d99e7cfa4219' Jul 12 09:28:34.776494 kernel: Demotion targets for Node 0: null Jul 12 09:28:34.776501 kernel: Key type .fscrypt registered Jul 12 09:28:34.776508 kernel: Key type fscrypt-provisioning registered Jul 12 09:28:34.776515 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 12 09:28:34.776522 kernel: ima: Allocated hash algorithm: sha1 Jul 12 09:28:34.776529 kernel: ima: No architecture policies found Jul 12 09:28:34.776536 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Jul 12 09:28:34.776544 kernel: clk: Disabling unused clocks Jul 12 09:28:34.776550 kernel: PM: genpd: Disabling unused power domains Jul 12 09:28:34.776557 kernel: Warning: unable to open an initial console. Jul 12 09:28:34.776564 kernel: Freeing unused kernel memory: 39424K Jul 12 09:28:34.776571 kernel: Run /init as init process Jul 12 09:28:34.776578 kernel: with arguments: Jul 12 09:28:34.776585 kernel: /init Jul 12 09:28:34.776592 kernel: with environment: Jul 12 09:28:34.776598 kernel: HOME=/ Jul 12 09:28:34.776605 kernel: TERM=linux Jul 12 09:28:34.776613 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 12 09:28:34.776621 systemd[1]: Successfully made /usr/ read-only. Jul 12 09:28:34.776631 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 12 09:28:34.776639 systemd[1]: Detected virtualization kvm. Jul 12 09:28:34.776647 systemd[1]: Detected architecture arm64. Jul 12 09:28:34.776654 systemd[1]: Running in initrd. Jul 12 09:28:34.776662 systemd[1]: No hostname configured, using default hostname. Jul 12 09:28:34.776671 systemd[1]: Hostname set to . Jul 12 09:28:34.776678 systemd[1]: Initializing machine ID from VM UUID. Jul 12 09:28:34.776685 systemd[1]: Queued start job for default target initrd.target. Jul 12 09:28:34.776693 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 12 09:28:34.776700 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 12 09:28:34.776708 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 12 09:28:34.776716 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 12 09:28:34.776724 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 12 09:28:34.776733 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 12 09:28:34.776741 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 12 09:28:34.776749 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 12 09:28:34.776757 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 12 09:28:34.776764 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 12 09:28:34.776772 systemd[1]: Reached target paths.target - Path Units. Jul 12 09:28:34.776779 systemd[1]: Reached target slices.target - Slice Units. Jul 12 09:28:34.776788 systemd[1]: Reached target swap.target - Swaps. Jul 12 09:28:34.776795 systemd[1]: Reached target timers.target - Timer Units. Jul 12 09:28:34.776803 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 12 09:28:34.776810 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 12 09:28:34.776818 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 12 09:28:34.776825 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jul 12 09:28:34.776833 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 12 09:28:34.776841 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 12 09:28:34.776850 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 12 09:28:34.776858 systemd[1]: Reached target sockets.target - Socket Units. Jul 12 09:28:34.776866 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 12 09:28:34.776874 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 12 09:28:34.776881 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 12 09:28:34.776889 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jul 12 09:28:34.776897 systemd[1]: Starting systemd-fsck-usr.service... Jul 12 09:28:34.776905 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 12 09:28:34.776931 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 12 09:28:34.776942 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 09:28:34.776949 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 12 09:28:34.776958 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 12 09:28:34.776965 systemd[1]: Finished systemd-fsck-usr.service. Jul 12 09:28:34.776974 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 12 09:28:34.776997 systemd-journald[244]: Collecting audit messages is disabled. Jul 12 09:28:34.777020 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 09:28:34.777029 systemd-journald[244]: Journal started Jul 12 09:28:34.777048 systemd-journald[244]: Runtime Journal (/run/log/journal/42c498951c294b88a258513adefe4344) is 6M, max 48.5M, 42.4M free. Jul 12 09:28:34.768097 systemd-modules-load[245]: Inserted module 'overlay' Jul 12 09:28:34.780928 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 12 09:28:34.782923 systemd[1]: Started systemd-journald.service - Journal Service. Jul 12 09:28:34.784927 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 12 09:28:34.786525 systemd-modules-load[245]: Inserted module 'br_netfilter' Jul 12 09:28:34.787202 kernel: Bridge firewalling registered Jul 12 09:28:34.795083 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 12 09:28:34.796109 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 12 09:28:34.800032 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 12 09:28:34.801268 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 12 09:28:34.804032 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 12 09:28:34.811321 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 12 09:28:34.812390 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 12 09:28:34.814019 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 12 09:28:34.814374 systemd-tmpfiles[270]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jul 12 09:28:34.816679 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 12 09:28:34.818090 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 12 09:28:34.820832 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 12 09:28:34.835778 dracut-cmdline[284]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=2eed6122ab9d95fa96c8f5511b96c1220a0caf18bbf7b84035ef573d9ba90496 Jul 12 09:28:34.848829 systemd-resolved[286]: Positive Trust Anchors: Jul 12 09:28:34.848847 systemd-resolved[286]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 12 09:28:34.848877 systemd-resolved[286]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 12 09:28:34.853643 systemd-resolved[286]: Defaulting to hostname 'linux'. Jul 12 09:28:34.854797 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 12 09:28:34.856247 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 12 09:28:34.914943 kernel: SCSI subsystem initialized Jul 12 09:28:34.919929 kernel: Loading iSCSI transport class v2.0-870. Jul 12 09:28:34.930932 kernel: iscsi: registered transport (tcp) Jul 12 09:28:34.943939 kernel: iscsi: registered transport (qla4xxx) Jul 12 09:28:34.943991 kernel: QLogic iSCSI HBA Driver Jul 12 09:28:34.960046 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 12 09:28:34.980988 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 12 09:28:34.983357 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 12 09:28:35.035101 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 12 09:28:35.039020 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 12 09:28:35.097939 kernel: raid6: neonx8 gen() 15769 MB/s Jul 12 09:28:35.114938 kernel: raid6: neonx4 gen() 15026 MB/s Jul 12 09:28:35.131940 kernel: raid6: neonx2 gen() 12546 MB/s Jul 12 09:28:35.148932 kernel: raid6: neonx1 gen() 9736 MB/s Jul 12 09:28:35.166085 kernel: raid6: int64x8 gen() 6572 MB/s Jul 12 09:28:35.182930 kernel: raid6: int64x4 gen() 6764 MB/s Jul 12 09:28:35.199936 kernel: raid6: int64x2 gen() 5693 MB/s Jul 12 09:28:35.216928 kernel: raid6: int64x1 gen() 5009 MB/s Jul 12 09:28:35.216949 kernel: raid6: using algorithm neonx8 gen() 15769 MB/s Jul 12 09:28:35.233934 kernel: raid6: .... xor() 11998 MB/s, rmw enabled Jul 12 09:28:35.233948 kernel: raid6: using neon recovery algorithm Jul 12 09:28:35.239113 kernel: xor: measuring software checksum speed Jul 12 09:28:35.239132 kernel: 8regs : 21562 MB/sec Jul 12 09:28:35.240229 kernel: 32regs : 21676 MB/sec Jul 12 09:28:35.240242 kernel: arm64_neon : 28244 MB/sec Jul 12 09:28:35.240251 kernel: xor: using function: arm64_neon (28244 MB/sec) Jul 12 09:28:35.298940 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 12 09:28:35.305158 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 12 09:28:35.307312 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 12 09:28:35.338505 systemd-udevd[498]: Using default interface naming scheme 'v255'. Jul 12 09:28:35.342504 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 12 09:28:35.344108 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 12 09:28:35.372962 dracut-pre-trigger[506]: rd.md=0: removing MD RAID activation Jul 12 09:28:35.394394 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 12 09:28:35.396361 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 12 09:28:35.461471 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 12 09:28:35.465241 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 12 09:28:35.509032 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Jul 12 09:28:35.509804 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Jul 12 09:28:35.512450 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jul 12 09:28:35.512486 kernel: GPT:9289727 != 19775487 Jul 12 09:28:35.512496 kernel: GPT:Alternate GPT header not at the end of the disk. Jul 12 09:28:35.512507 kernel: GPT:9289727 != 19775487 Jul 12 09:28:35.513222 kernel: GPT: Use GNU Parted to correct GPT errors. Jul 12 09:28:35.513235 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jul 12 09:28:35.516080 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 12 09:28:35.516189 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 09:28:35.518321 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 09:28:35.520067 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 09:28:35.542659 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jul 12 09:28:35.546223 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 09:28:35.555271 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jul 12 09:28:35.556336 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 12 09:28:35.572938 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jul 12 09:28:35.578630 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jul 12 09:28:35.579514 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Jul 12 09:28:35.581123 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 12 09:28:35.583155 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 12 09:28:35.584823 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 12 09:28:35.586953 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 12 09:28:35.588484 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 12 09:28:35.606826 disk-uuid[590]: Primary Header is updated. Jul 12 09:28:35.606826 disk-uuid[590]: Secondary Entries is updated. Jul 12 09:28:35.606826 disk-uuid[590]: Secondary Header is updated. Jul 12 09:28:35.609770 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 12 09:28:35.612951 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jul 12 09:28:36.617934 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jul 12 09:28:36.618134 disk-uuid[595]: The operation has completed successfully. Jul 12 09:28:36.652396 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 12 09:28:36.652493 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 12 09:28:36.665957 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 12 09:28:36.692828 sh[610]: Success Jul 12 09:28:36.708923 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 12 09:28:36.710223 kernel: device-mapper: uevent: version 1.0.3 Jul 12 09:28:36.710248 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jul 12 09:28:36.716933 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Jul 12 09:28:36.742643 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 12 09:28:36.745148 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 12 09:28:36.764947 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 12 09:28:36.771759 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' Jul 12 09:28:36.771809 kernel: BTRFS: device fsid 61a6979b-5b23-4687-8775-cb04acb91b0a devid 1 transid 37 /dev/mapper/usr (253:0) scanned by mount (623) Jul 12 09:28:36.773498 kernel: BTRFS info (device dm-0): first mount of filesystem 61a6979b-5b23-4687-8775-cb04acb91b0a Jul 12 09:28:36.773515 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Jul 12 09:28:36.773526 kernel: BTRFS info (device dm-0): using free-space-tree Jul 12 09:28:36.777108 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 12 09:28:36.778131 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jul 12 09:28:36.779092 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jul 12 09:28:36.779811 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 12 09:28:36.782301 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 12 09:28:36.806353 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 (254:6) scanned by mount (655) Jul 12 09:28:36.806404 kernel: BTRFS info (device vda6): first mount of filesystem e5a719e8-42e4-4055-8ce0-9ce9f50475f2 Jul 12 09:28:36.806421 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Jul 12 09:28:36.807934 kernel: BTRFS info (device vda6): using free-space-tree Jul 12 09:28:36.813281 kernel: BTRFS info (device vda6): last unmount of filesystem e5a719e8-42e4-4055-8ce0-9ce9f50475f2 Jul 12 09:28:36.814535 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 12 09:28:36.816342 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 12 09:28:36.888970 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 12 09:28:36.891610 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 12 09:28:36.930036 systemd-networkd[797]: lo: Link UP Jul 12 09:28:36.930047 systemd-networkd[797]: lo: Gained carrier Jul 12 09:28:36.930861 systemd-networkd[797]: Enumeration completed Jul 12 09:28:36.931365 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 12 09:28:36.931483 systemd-networkd[797]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 09:28:36.931486 systemd-networkd[797]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 12 09:28:36.932344 systemd-networkd[797]: eth0: Link UP Jul 12 09:28:36.932347 systemd-networkd[797]: eth0: Gained carrier Jul 12 09:28:36.932354 systemd-networkd[797]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 09:28:36.932539 systemd[1]: Reached target network.target - Network. Jul 12 09:28:36.953478 systemd-networkd[797]: eth0: DHCPv4 address 10.0.0.27/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jul 12 09:28:36.985499 ignition[697]: Ignition 2.21.0 Jul 12 09:28:36.985514 ignition[697]: Stage: fetch-offline Jul 12 09:28:36.985544 ignition[697]: no configs at "/usr/lib/ignition/base.d" Jul 12 09:28:36.985552 ignition[697]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jul 12 09:28:36.985729 ignition[697]: parsed url from cmdline: "" Jul 12 09:28:36.985732 ignition[697]: no config URL provided Jul 12 09:28:36.985736 ignition[697]: reading system config file "/usr/lib/ignition/user.ign" Jul 12 09:28:36.985745 ignition[697]: no config at "/usr/lib/ignition/user.ign" Jul 12 09:28:36.985763 ignition[697]: op(1): [started] loading QEMU firmware config module Jul 12 09:28:36.985767 ignition[697]: op(1): executing: "modprobe" "qemu_fw_cfg" Jul 12 09:28:36.994362 ignition[697]: op(1): [finished] loading QEMU firmware config module Jul 12 09:28:37.033104 ignition[697]: parsing config with SHA512: cd7316046f7536d952344bec76e8a1efd918585c66c2c7d6eb9fcee8d7e6c4882070b764b6f6c2abc0ab7b32f833e8ffe7bde02bad4ecb5d371ab06df986612e Jul 12 09:28:37.038400 unknown[697]: fetched base config from "system" Jul 12 09:28:37.038967 unknown[697]: fetched user config from "qemu" Jul 12 09:28:37.039476 ignition[697]: fetch-offline: fetch-offline passed Jul 12 09:28:37.039538 ignition[697]: Ignition finished successfully Jul 12 09:28:37.041174 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 12 09:28:37.042183 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jul 12 09:28:37.044051 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 12 09:28:37.072272 ignition[812]: Ignition 2.21.0 Jul 12 09:28:37.072285 ignition[812]: Stage: kargs Jul 12 09:28:37.072418 ignition[812]: no configs at "/usr/lib/ignition/base.d" Jul 12 09:28:37.072426 ignition[812]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jul 12 09:28:37.073152 ignition[812]: kargs: kargs passed Jul 12 09:28:37.073198 ignition[812]: Ignition finished successfully Jul 12 09:28:37.077111 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 12 09:28:37.078704 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 12 09:28:37.104610 ignition[820]: Ignition 2.21.0 Jul 12 09:28:37.104627 ignition[820]: Stage: disks Jul 12 09:28:37.104773 ignition[820]: no configs at "/usr/lib/ignition/base.d" Jul 12 09:28:37.104781 ignition[820]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jul 12 09:28:37.106722 ignition[820]: disks: disks passed Jul 12 09:28:37.106779 ignition[820]: Ignition finished successfully Jul 12 09:28:37.108344 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 12 09:28:37.109483 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 12 09:28:37.110660 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 12 09:28:37.112130 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 12 09:28:37.113495 systemd[1]: Reached target sysinit.target - System Initialization. Jul 12 09:28:37.114735 systemd[1]: Reached target basic.target - Basic System. Jul 12 09:28:37.116782 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 12 09:28:37.138178 systemd-fsck[830]: ROOT: clean, 15/553520 files, 52789/553472 blocks Jul 12 09:28:37.142839 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 12 09:28:37.145061 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 12 09:28:37.216928 kernel: EXT4-fs (vda9): mounted filesystem 016d0f7f-22a0-4255-85cc-97a6d773acb9 r/w with ordered data mode. Quota mode: none. Jul 12 09:28:37.217565 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 12 09:28:37.218660 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 12 09:28:37.220659 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 12 09:28:37.222130 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 12 09:28:37.222896 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 12 09:28:37.222957 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 12 09:28:37.222981 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 12 09:28:37.238707 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 12 09:28:37.241633 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 12 09:28:37.246012 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 (254:6) scanned by mount (838) Jul 12 09:28:37.246036 kernel: BTRFS info (device vda6): first mount of filesystem e5a719e8-42e4-4055-8ce0-9ce9f50475f2 Jul 12 09:28:37.246046 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Jul 12 09:28:37.246055 kernel: BTRFS info (device vda6): using free-space-tree Jul 12 09:28:37.249331 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 12 09:28:37.304560 initrd-setup-root[864]: cut: /sysroot/etc/passwd: No such file or directory Jul 12 09:28:37.308306 initrd-setup-root[871]: cut: /sysroot/etc/group: No such file or directory Jul 12 09:28:37.312182 initrd-setup-root[878]: cut: /sysroot/etc/shadow: No such file or directory Jul 12 09:28:37.315860 initrd-setup-root[885]: cut: /sysroot/etc/gshadow: No such file or directory Jul 12 09:28:37.385513 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 12 09:28:37.387246 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 12 09:28:37.390264 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 12 09:28:37.406943 kernel: BTRFS info (device vda6): last unmount of filesystem e5a719e8-42e4-4055-8ce0-9ce9f50475f2 Jul 12 09:28:37.420017 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 12 09:28:37.430784 ignition[955]: INFO : Ignition 2.21.0 Jul 12 09:28:37.430784 ignition[955]: INFO : Stage: mount Jul 12 09:28:37.432023 ignition[955]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 12 09:28:37.432023 ignition[955]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jul 12 09:28:37.434590 ignition[955]: INFO : mount: mount passed Jul 12 09:28:37.434590 ignition[955]: INFO : Ignition finished successfully Jul 12 09:28:37.434942 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 12 09:28:37.436771 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 12 09:28:37.771233 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 12 09:28:37.772663 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 12 09:28:37.791610 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 (254:6) scanned by mount (967) Jul 12 09:28:37.793340 kernel: BTRFS info (device vda6): first mount of filesystem e5a719e8-42e4-4055-8ce0-9ce9f50475f2 Jul 12 09:28:37.793375 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Jul 12 09:28:37.793386 kernel: BTRFS info (device vda6): using free-space-tree Jul 12 09:28:37.796386 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 12 09:28:37.823008 ignition[984]: INFO : Ignition 2.21.0 Jul 12 09:28:37.823008 ignition[984]: INFO : Stage: files Jul 12 09:28:37.824728 ignition[984]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 12 09:28:37.824728 ignition[984]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jul 12 09:28:37.824728 ignition[984]: DEBUG : files: compiled without relabeling support, skipping Jul 12 09:28:37.827218 ignition[984]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 12 09:28:37.827218 ignition[984]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 12 09:28:37.827218 ignition[984]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 12 09:28:37.827218 ignition[984]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 12 09:28:37.831284 ignition[984]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 12 09:28:37.831284 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jul 12 09:28:37.831284 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Jul 12 09:28:37.827487 unknown[984]: wrote ssh authorized keys file for user: core Jul 12 09:28:37.918145 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jul 12 09:28:38.117235 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jul 12 09:28:38.117235 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 12 09:28:38.120325 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 12 09:28:38.128960 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 12 09:28:38.128960 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 12 09:28:38.128960 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 09:28:38.128960 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 09:28:38.128960 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 09:28:38.128960 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Jul 12 09:28:38.752104 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jul 12 09:28:38.775102 systemd-networkd[797]: eth0: Gained IPv6LL Jul 12 09:28:39.296234 ignition[984]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 09:28:39.296234 ignition[984]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jul 12 09:28:39.299323 ignition[984]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 12 09:28:39.301986 ignition[984]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 12 09:28:39.301986 ignition[984]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jul 12 09:28:39.301986 ignition[984]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jul 12 09:28:39.305711 ignition[984]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jul 12 09:28:39.305711 ignition[984]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jul 12 09:28:39.305711 ignition[984]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jul 12 09:28:39.305711 ignition[984]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Jul 12 09:28:39.322318 ignition[984]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Jul 12 09:28:39.329156 ignition[984]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jul 12 09:28:39.331071 ignition[984]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Jul 12 09:28:39.331071 ignition[984]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Jul 12 09:28:39.331071 ignition[984]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Jul 12 09:28:39.331071 ignition[984]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 12 09:28:39.331071 ignition[984]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 12 09:28:39.331071 ignition[984]: INFO : files: files passed Jul 12 09:28:39.331071 ignition[984]: INFO : Ignition finished successfully Jul 12 09:28:39.332696 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 12 09:28:39.335052 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 12 09:28:39.338077 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 12 09:28:39.357858 initrd-setup-root-after-ignition[1013]: grep: /sysroot/oem/oem-release: No such file or directory Jul 12 09:28:39.356877 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 12 09:28:39.356994 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 12 09:28:39.361730 initrd-setup-root-after-ignition[1015]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 12 09:28:39.361730 initrd-setup-root-after-ignition[1015]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 12 09:28:39.364078 initrd-setup-root-after-ignition[1019]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 12 09:28:39.363167 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 12 09:28:39.365325 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 12 09:28:39.367755 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 12 09:28:39.402043 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 12 09:28:39.402165 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 12 09:28:39.403983 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 12 09:28:39.405532 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 12 09:28:39.407042 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 12 09:28:39.407893 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 12 09:28:39.436143 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 12 09:28:39.439484 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 12 09:28:39.461243 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 12 09:28:39.462393 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 12 09:28:39.464022 systemd[1]: Stopped target timers.target - Timer Units. Jul 12 09:28:39.465425 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 12 09:28:39.465575 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 12 09:28:39.467520 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 12 09:28:39.469091 systemd[1]: Stopped target basic.target - Basic System. Jul 12 09:28:39.470408 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 12 09:28:39.471820 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 12 09:28:39.473372 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 12 09:28:39.474971 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jul 12 09:28:39.476530 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 12 09:28:39.477964 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 12 09:28:39.479593 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 12 09:28:39.481132 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 12 09:28:39.482479 systemd[1]: Stopped target swap.target - Swaps. Jul 12 09:28:39.483665 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 12 09:28:39.483828 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 12 09:28:39.485582 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 12 09:28:39.487129 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 12 09:28:39.488666 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 12 09:28:39.488967 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 12 09:28:39.491270 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 12 09:28:39.491428 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 12 09:28:39.493712 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 12 09:28:39.493873 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 12 09:28:39.495359 systemd[1]: Stopped target paths.target - Path Units. Jul 12 09:28:39.496544 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 12 09:28:39.499976 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 12 09:28:39.500894 systemd[1]: Stopped target slices.target - Slice Units. Jul 12 09:28:39.502426 systemd[1]: Stopped target sockets.target - Socket Units. Jul 12 09:28:39.503733 systemd[1]: iscsid.socket: Deactivated successfully. Jul 12 09:28:39.503836 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 12 09:28:39.505099 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 12 09:28:39.505212 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 12 09:28:39.506379 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 12 09:28:39.506537 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 12 09:28:39.507907 systemd[1]: ignition-files.service: Deactivated successfully. Jul 12 09:28:39.508066 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 12 09:28:39.510192 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 12 09:28:39.512036 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 12 09:28:39.513322 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 12 09:28:39.513428 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 12 09:28:39.514767 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 12 09:28:39.514864 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 12 09:28:39.521349 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 12 09:28:39.522046 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 12 09:28:39.531329 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 12 09:28:39.536009 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 12 09:28:39.536840 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 12 09:28:39.538299 ignition[1039]: INFO : Ignition 2.21.0 Jul 12 09:28:39.538299 ignition[1039]: INFO : Stage: umount Jul 12 09:28:39.538299 ignition[1039]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 12 09:28:39.538299 ignition[1039]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jul 12 09:28:39.538299 ignition[1039]: INFO : umount: umount passed Jul 12 09:28:39.538299 ignition[1039]: INFO : Ignition finished successfully Jul 12 09:28:39.540544 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 12 09:28:39.540638 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 12 09:28:39.541930 systemd[1]: Stopped target network.target - Network. Jul 12 09:28:39.543233 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 12 09:28:39.543295 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 12 09:28:39.544467 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 12 09:28:39.544511 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 12 09:28:39.545634 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 12 09:28:39.545677 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 12 09:28:39.546902 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 12 09:28:39.546955 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 12 09:28:39.548355 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 12 09:28:39.548401 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 12 09:28:39.553289 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 12 09:28:39.554451 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 12 09:28:39.561639 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 12 09:28:39.561747 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 12 09:28:39.565048 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Jul 12 09:28:39.565267 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 12 09:28:39.565359 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 12 09:28:39.568727 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Jul 12 09:28:39.569323 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jul 12 09:28:39.570751 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 12 09:28:39.570784 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 12 09:28:39.573340 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 12 09:28:39.575036 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 12 09:28:39.575106 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 12 09:28:39.576561 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 12 09:28:39.576674 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 12 09:28:39.582649 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 12 09:28:39.582701 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 12 09:28:39.585175 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 12 09:28:39.585230 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 12 09:28:39.588146 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 12 09:28:39.592481 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Jul 12 09:28:39.592567 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Jul 12 09:28:39.607985 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 12 09:28:39.608150 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 12 09:28:39.609866 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 12 09:28:39.610020 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 12 09:28:39.611644 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 12 09:28:39.611703 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 12 09:28:39.612529 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 12 09:28:39.612556 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 12 09:28:39.613978 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 12 09:28:39.614049 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 12 09:28:39.616110 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 12 09:28:39.616176 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 12 09:28:39.618364 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 12 09:28:39.618417 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 12 09:28:39.621581 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 12 09:28:39.622699 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jul 12 09:28:39.622747 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jul 12 09:28:39.625482 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 12 09:28:39.625528 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 12 09:28:39.627801 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jul 12 09:28:39.627844 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 12 09:28:39.630324 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 12 09:28:39.630365 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 12 09:28:39.632070 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 12 09:28:39.632110 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 09:28:39.635539 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Jul 12 09:28:39.635585 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Jul 12 09:28:39.635612 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Jul 12 09:28:39.635641 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Jul 12 09:28:39.636015 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 12 09:28:39.636093 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 12 09:28:39.637409 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 12 09:28:39.639341 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 12 09:28:39.655274 systemd[1]: Switching root. Jul 12 09:28:39.680975 systemd-journald[244]: Journal stopped Jul 12 09:28:40.460120 systemd-journald[244]: Received SIGTERM from PID 1 (systemd). Jul 12 09:28:40.460179 kernel: SELinux: policy capability network_peer_controls=1 Jul 12 09:28:40.460192 kernel: SELinux: policy capability open_perms=1 Jul 12 09:28:40.460201 kernel: SELinux: policy capability extended_socket_class=1 Jul 12 09:28:40.460210 kernel: SELinux: policy capability always_check_network=0 Jul 12 09:28:40.460222 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 12 09:28:40.460236 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 12 09:28:40.460245 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 12 09:28:40.460254 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 12 09:28:40.460264 kernel: SELinux: policy capability userspace_initial_context=0 Jul 12 09:28:40.460273 kernel: audit: type=1403 audit(1752312519.867:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 12 09:28:40.460288 systemd[1]: Successfully loaded SELinux policy in 75.393ms. Jul 12 09:28:40.460310 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.208ms. Jul 12 09:28:40.460322 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 12 09:28:40.460333 systemd[1]: Detected virtualization kvm. Jul 12 09:28:40.460344 systemd[1]: Detected architecture arm64. Jul 12 09:28:40.460354 systemd[1]: Detected first boot. Jul 12 09:28:40.460365 systemd[1]: Initializing machine ID from VM UUID. Jul 12 09:28:40.460375 zram_generator::config[1084]: No configuration found. Jul 12 09:28:40.460386 kernel: NET: Registered PF_VSOCK protocol family Jul 12 09:28:40.460398 systemd[1]: Populated /etc with preset unit settings. Jul 12 09:28:40.460409 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Jul 12 09:28:40.460419 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jul 12 09:28:40.460429 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jul 12 09:28:40.460440 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jul 12 09:28:40.460450 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jul 12 09:28:40.460461 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jul 12 09:28:40.460471 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jul 12 09:28:40.460482 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jul 12 09:28:40.460497 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jul 12 09:28:40.460507 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jul 12 09:28:40.460517 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jul 12 09:28:40.460527 systemd[1]: Created slice user.slice - User and Session Slice. Jul 12 09:28:40.460540 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 12 09:28:40.460552 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 12 09:28:40.460562 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jul 12 09:28:40.460572 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jul 12 09:28:40.460583 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jul 12 09:28:40.460595 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 12 09:28:40.460605 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Jul 12 09:28:40.460615 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 12 09:28:40.460625 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 12 09:28:40.460637 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jul 12 09:28:40.460647 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jul 12 09:28:40.460658 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jul 12 09:28:40.460669 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jul 12 09:28:40.460680 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 12 09:28:40.460690 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 12 09:28:40.460700 systemd[1]: Reached target slices.target - Slice Units. Jul 12 09:28:40.460710 systemd[1]: Reached target swap.target - Swaps. Jul 12 09:28:40.460721 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jul 12 09:28:40.460731 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jul 12 09:28:40.460741 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jul 12 09:28:40.460752 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 12 09:28:40.460763 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 12 09:28:40.460774 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 12 09:28:40.460785 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jul 12 09:28:40.460795 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jul 12 09:28:40.460806 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jul 12 09:28:40.460816 systemd[1]: Mounting media.mount - External Media Directory... Jul 12 09:28:40.460826 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jul 12 09:28:40.460836 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jul 12 09:28:40.460846 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jul 12 09:28:40.460859 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 12 09:28:40.460869 systemd[1]: Reached target machines.target - Containers. Jul 12 09:28:40.460879 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jul 12 09:28:40.460890 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 09:28:40.460900 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 12 09:28:40.460920 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jul 12 09:28:40.460931 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 09:28:40.460941 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 12 09:28:40.460954 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 09:28:40.460964 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jul 12 09:28:40.460973 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 09:28:40.460984 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 12 09:28:40.460999 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jul 12 09:28:40.461010 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jul 12 09:28:40.461019 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jul 12 09:28:40.461030 systemd[1]: Stopped systemd-fsck-usr.service. Jul 12 09:28:40.461040 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 12 09:28:40.461052 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 12 09:28:40.461063 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 12 09:28:40.461072 kernel: fuse: init (API version 7.41) Jul 12 09:28:40.461081 kernel: loop: module loaded Jul 12 09:28:40.461091 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 12 09:28:40.461101 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jul 12 09:28:40.461111 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jul 12 09:28:40.461122 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 12 09:28:40.461133 systemd[1]: verity-setup.service: Deactivated successfully. Jul 12 09:28:40.461144 systemd[1]: Stopped verity-setup.service. Jul 12 09:28:40.461153 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jul 12 09:28:40.461164 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jul 12 09:28:40.461174 systemd[1]: Mounted media.mount - External Media Directory. Jul 12 09:28:40.461186 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jul 12 09:28:40.461195 kernel: ACPI: bus type drm_connector registered Jul 12 09:28:40.461230 systemd-journald[1149]: Collecting audit messages is disabled. Jul 12 09:28:40.461252 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jul 12 09:28:40.461264 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jul 12 09:28:40.461275 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jul 12 09:28:40.461286 systemd-journald[1149]: Journal started Jul 12 09:28:40.461306 systemd-journald[1149]: Runtime Journal (/run/log/journal/42c498951c294b88a258513adefe4344) is 6M, max 48.5M, 42.4M free. Jul 12 09:28:40.264545 systemd[1]: Queued start job for default target multi-user.target. Jul 12 09:28:40.282960 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jul 12 09:28:40.283348 systemd[1]: systemd-journald.service: Deactivated successfully. Jul 12 09:28:40.464212 systemd[1]: Started systemd-journald.service - Journal Service. Jul 12 09:28:40.464901 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 12 09:28:40.466105 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 12 09:28:40.466275 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jul 12 09:28:40.467440 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 09:28:40.467590 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 09:28:40.468663 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 12 09:28:40.468819 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 12 09:28:40.469927 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 09:28:40.470115 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 09:28:40.471223 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 12 09:28:40.471371 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jul 12 09:28:40.472508 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 09:28:40.472664 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 09:28:40.475271 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 12 09:28:40.476506 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 12 09:28:40.477719 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jul 12 09:28:40.479162 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jul 12 09:28:40.490446 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 12 09:28:40.492511 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jul 12 09:28:40.494211 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jul 12 09:28:40.495062 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 12 09:28:40.495092 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 12 09:28:40.496749 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jul 12 09:28:40.502854 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jul 12 09:28:40.504082 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 09:28:40.505339 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jul 12 09:28:40.507103 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jul 12 09:28:40.508144 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 12 09:28:40.511019 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jul 12 09:28:40.511874 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 12 09:28:40.512689 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 12 09:28:40.514365 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jul 12 09:28:40.517114 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 12 09:28:40.519512 systemd-journald[1149]: Time spent on flushing to /var/log/journal/42c498951c294b88a258513adefe4344 is 23.230ms for 888 entries. Jul 12 09:28:40.519512 systemd-journald[1149]: System Journal (/var/log/journal/42c498951c294b88a258513adefe4344) is 8M, max 195.6M, 187.6M free. Jul 12 09:28:40.545950 systemd-journald[1149]: Received client request to flush runtime journal. Jul 12 09:28:40.545985 kernel: loop0: detected capacity change from 0 to 134232 Jul 12 09:28:40.533957 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 12 09:28:40.535176 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jul 12 09:28:40.536903 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jul 12 09:28:40.542550 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jul 12 09:28:40.548608 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jul 12 09:28:40.551598 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jul 12 09:28:40.555734 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jul 12 09:28:40.565851 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 12 09:28:40.565969 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 12 09:28:40.568722 systemd-tmpfiles[1201]: ACLs are not supported, ignoring. Jul 12 09:28:40.568738 systemd-tmpfiles[1201]: ACLs are not supported, ignoring. Jul 12 09:28:40.574168 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 12 09:28:40.576623 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jul 12 09:28:40.593059 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jul 12 09:28:40.598957 kernel: loop1: detected capacity change from 0 to 203944 Jul 12 09:28:40.602394 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jul 12 09:28:40.605055 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 12 09:28:40.621407 systemd-tmpfiles[1222]: ACLs are not supported, ignoring. Jul 12 09:28:40.621425 systemd-tmpfiles[1222]: ACLs are not supported, ignoring. Jul 12 09:28:40.625023 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 12 09:28:40.629002 kernel: loop2: detected capacity change from 0 to 105936 Jul 12 09:28:40.645933 kernel: loop3: detected capacity change from 0 to 134232 Jul 12 09:28:40.652933 kernel: loop4: detected capacity change from 0 to 203944 Jul 12 09:28:40.657926 kernel: loop5: detected capacity change from 0 to 105936 Jul 12 09:28:40.661026 (sd-merge)[1227]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Jul 12 09:28:40.661377 (sd-merge)[1227]: Merged extensions into '/usr'. Jul 12 09:28:40.664312 systemd[1]: Reload requested from client PID 1200 ('systemd-sysext') (unit systemd-sysext.service)... Jul 12 09:28:40.664328 systemd[1]: Reloading... Jul 12 09:28:40.703343 zram_generator::config[1252]: No configuration found. Jul 12 09:28:40.785130 ldconfig[1195]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 12 09:28:40.789729 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 09:28:40.850502 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 12 09:28:40.850638 systemd[1]: Reloading finished in 185 ms. Jul 12 09:28:40.865328 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jul 12 09:28:40.866472 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jul 12 09:28:40.877115 systemd[1]: Starting ensure-sysext.service... Jul 12 09:28:40.878634 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 12 09:28:40.893359 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jul 12 09:28:40.893665 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jul 12 09:28:40.893979 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 12 09:28:40.894259 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jul 12 09:28:40.894474 systemd[1]: Reload requested from client PID 1288 ('systemctl') (unit ensure-sysext.service)... Jul 12 09:28:40.894493 systemd[1]: Reloading... Jul 12 09:28:40.895075 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 12 09:28:40.895361 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Jul 12 09:28:40.895466 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Jul 12 09:28:40.898285 systemd-tmpfiles[1289]: Detected autofs mount point /boot during canonicalization of boot. Jul 12 09:28:40.898383 systemd-tmpfiles[1289]: Skipping /boot Jul 12 09:28:40.903882 systemd-tmpfiles[1289]: Detected autofs mount point /boot during canonicalization of boot. Jul 12 09:28:40.904022 systemd-tmpfiles[1289]: Skipping /boot Jul 12 09:28:40.937987 zram_generator::config[1316]: No configuration found. Jul 12 09:28:41.004530 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 09:28:41.064826 systemd[1]: Reloading finished in 170 ms. Jul 12 09:28:41.086415 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jul 12 09:28:41.087633 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 12 09:28:41.108012 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 12 09:28:41.110084 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jul 12 09:28:41.124716 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jul 12 09:28:41.128394 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 12 09:28:41.130875 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 12 09:28:41.132811 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jul 12 09:28:41.138220 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 09:28:41.144960 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 09:28:41.149120 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 09:28:41.151146 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 09:28:41.152024 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 09:28:41.152133 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 12 09:28:41.156974 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jul 12 09:28:41.158710 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 09:28:41.158945 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 09:28:41.160442 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 09:28:41.160644 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 09:28:41.162230 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 09:28:41.162412 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 09:28:41.167770 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 09:28:41.170190 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 09:28:41.171582 systemd-udevd[1357]: Using default interface naming scheme 'v255'. Jul 12 09:28:41.172288 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 09:28:41.174197 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 09:28:41.175082 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 09:28:41.175235 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 12 09:28:41.179419 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jul 12 09:28:41.185215 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jul 12 09:28:41.190344 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jul 12 09:28:41.191901 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jul 12 09:28:41.193351 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 09:28:41.193497 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 09:28:41.194849 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 09:28:41.194990 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 09:28:41.196409 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 09:28:41.196542 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 09:28:41.197844 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jul 12 09:28:41.201585 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 12 09:28:41.208310 augenrules[1398]: No rules Jul 12 09:28:41.209263 systemd[1]: audit-rules.service: Deactivated successfully. Jul 12 09:28:41.209732 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 12 09:28:41.215166 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 09:28:41.218042 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 09:28:41.222083 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 12 09:28:41.225159 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 09:28:41.228948 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 09:28:41.229757 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 09:28:41.229807 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 12 09:28:41.233103 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 12 09:28:41.235004 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 12 09:28:41.251414 systemd[1]: Finished ensure-sysext.service. Jul 12 09:28:41.253441 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 09:28:41.253785 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 09:28:41.255898 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 09:28:41.256601 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 09:28:41.257904 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 09:28:41.258102 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 09:28:41.262736 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 12 09:28:41.263539 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 12 09:28:41.271259 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 12 09:28:41.271309 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 12 09:28:41.273633 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jul 12 09:28:41.276356 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jul 12 09:28:41.311301 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Jul 12 09:28:41.316581 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jul 12 09:28:41.319722 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jul 12 09:28:41.340950 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jul 12 09:28:41.379196 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jul 12 09:28:41.380594 systemd[1]: Reached target time-set.target - System Time Set. Jul 12 09:28:41.384593 systemd-networkd[1432]: lo: Link UP Jul 12 09:28:41.384601 systemd-networkd[1432]: lo: Gained carrier Jul 12 09:28:41.385212 systemd-resolved[1356]: Positive Trust Anchors: Jul 12 09:28:41.385228 systemd-resolved[1356]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 12 09:28:41.385260 systemd-resolved[1356]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 12 09:28:41.385467 systemd-networkd[1432]: Enumeration completed Jul 12 09:28:41.385614 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 12 09:28:41.385842 systemd-networkd[1432]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 09:28:41.385851 systemd-networkd[1432]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 12 09:28:41.386425 systemd-networkd[1432]: eth0: Link UP Jul 12 09:28:41.386533 systemd-networkd[1432]: eth0: Gained carrier Jul 12 09:28:41.386546 systemd-networkd[1432]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 09:28:41.388901 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jul 12 09:28:41.393110 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jul 12 09:28:41.398854 systemd-resolved[1356]: Defaulting to hostname 'linux'. Jul 12 09:28:41.401935 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 12 09:28:41.402823 systemd[1]: Reached target network.target - Network. Jul 12 09:28:41.403975 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 12 09:28:41.404818 systemd[1]: Reached target sysinit.target - System Initialization. Jul 12 09:28:41.407041 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jul 12 09:28:41.407923 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jul 12 09:28:41.407985 systemd-networkd[1432]: eth0: DHCPv4 address 10.0.0.27/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jul 12 09:28:41.408455 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. Jul 12 09:28:41.408928 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jul 12 09:28:41.409824 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jul 12 09:28:41.410749 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jul 12 09:28:41.414050 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 12 09:28:41.414078 systemd[1]: Reached target paths.target - Path Units. Jul 12 09:28:41.414712 systemd[1]: Reached target timers.target - Timer Units. Jul 12 09:28:41.416084 systemd-timesyncd[1441]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jul 12 09:28:41.416143 systemd-timesyncd[1441]: Initial clock synchronization to Sat 2025-07-12 09:28:41.143532 UTC. Jul 12 09:28:41.416155 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jul 12 09:28:41.418100 systemd[1]: Starting docker.socket - Docker Socket for the API... Jul 12 09:28:41.421423 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jul 12 09:28:41.422468 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jul 12 09:28:41.423676 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jul 12 09:28:41.427539 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jul 12 09:28:41.428574 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jul 12 09:28:41.430860 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jul 12 09:28:41.432070 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jul 12 09:28:41.433238 systemd[1]: Reached target sockets.target - Socket Units. Jul 12 09:28:41.433959 systemd[1]: Reached target basic.target - Basic System. Jul 12 09:28:41.434650 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jul 12 09:28:41.434684 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jul 12 09:28:41.437061 systemd[1]: Starting containerd.service - containerd container runtime... Jul 12 09:28:41.440025 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jul 12 09:28:41.443149 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jul 12 09:28:41.448201 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jul 12 09:28:41.451707 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jul 12 09:28:41.452469 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jul 12 09:28:41.453354 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jul 12 09:28:41.458502 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jul 12 09:28:41.461178 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jul 12 09:28:41.463149 jq[1481]: false Jul 12 09:28:41.464149 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jul 12 09:28:41.472980 systemd[1]: Starting systemd-logind.service - User Login Management... Jul 12 09:28:41.473186 extend-filesystems[1482]: Found /dev/vda6 Jul 12 09:28:41.474501 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jul 12 09:28:41.474868 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 12 09:28:41.475424 systemd[1]: Starting update-engine.service - Update Engine... Jul 12 09:28:41.478064 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jul 12 09:28:41.480280 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jul 12 09:28:41.481478 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 12 09:28:41.481645 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jul 12 09:28:41.482371 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 12 09:28:41.482515 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jul 12 09:28:41.484324 systemd[1]: motdgen.service: Deactivated successfully. Jul 12 09:28:41.484502 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jul 12 09:28:41.491984 jq[1500]: true Jul 12 09:28:41.492323 extend-filesystems[1482]: Found /dev/vda9 Jul 12 09:28:41.506201 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 09:28:41.508488 jq[1511]: true Jul 12 09:28:41.509035 extend-filesystems[1482]: Checking size of /dev/vda9 Jul 12 09:28:41.517495 (ntainerd)[1514]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jul 12 09:28:41.532591 tar[1503]: linux-arm64/helm Jul 12 09:28:41.535405 extend-filesystems[1482]: Resized partition /dev/vda9 Jul 12 09:28:41.542260 extend-filesystems[1531]: resize2fs 1.47.2 (1-Jan-2025) Jul 12 09:28:41.553935 update_engine[1498]: I20250712 09:28:41.550824 1498 main.cc:92] Flatcar Update Engine starting Jul 12 09:28:41.559480 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Jul 12 09:28:41.560673 dbus-daemon[1478]: [system] SELinux support is enabled Jul 12 09:28:41.560882 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jul 12 09:28:41.565812 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 12 09:28:41.565842 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jul 12 09:28:41.577828 update_engine[1498]: I20250712 09:28:41.576784 1498 update_check_scheduler.cc:74] Next update check in 4m15s Jul 12 09:28:41.568374 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 12 09:28:41.568392 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jul 12 09:28:41.570013 systemd[1]: Started update-engine.service - Update Engine. Jul 12 09:28:41.582442 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jul 12 09:28:41.586353 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Jul 12 09:28:41.603666 extend-filesystems[1531]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jul 12 09:28:41.603666 extend-filesystems[1531]: old_desc_blocks = 1, new_desc_blocks = 1 Jul 12 09:28:41.603666 extend-filesystems[1531]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Jul 12 09:28:41.607113 bash[1539]: Updated "/home/core/.ssh/authorized_keys" Jul 12 09:28:41.605665 systemd-logind[1493]: Watching system buttons on /dev/input/event0 (Power Button) Jul 12 09:28:41.607406 extend-filesystems[1482]: Resized filesystem in /dev/vda9 Jul 12 09:28:41.607536 systemd-logind[1493]: New seat seat0. Jul 12 09:28:41.612107 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jul 12 09:28:41.613614 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 12 09:28:41.613899 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jul 12 09:28:41.642775 systemd[1]: Started systemd-logind.service - User Login Management. Jul 12 09:28:41.648008 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 09:28:41.654140 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jul 12 09:28:41.665370 locksmithd[1542]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 12 09:28:41.777366 containerd[1514]: time="2025-07-12T09:28:41Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jul 12 09:28:41.779630 containerd[1514]: time="2025-07-12T09:28:41.779586440Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Jul 12 09:28:41.792456 containerd[1514]: time="2025-07-12T09:28:41.792412800Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="8.48µs" Jul 12 09:28:41.792456 containerd[1514]: time="2025-07-12T09:28:41.792447200Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jul 12 09:28:41.792552 containerd[1514]: time="2025-07-12T09:28:41.792465680Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jul 12 09:28:41.792722 containerd[1514]: time="2025-07-12T09:28:41.792689000Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jul 12 09:28:41.792752 containerd[1514]: time="2025-07-12T09:28:41.792734960Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jul 12 09:28:41.792770 containerd[1514]: time="2025-07-12T09:28:41.792760600Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jul 12 09:28:41.792896 containerd[1514]: time="2025-07-12T09:28:41.792867800Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jul 12 09:28:41.792896 containerd[1514]: time="2025-07-12T09:28:41.792889760Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jul 12 09:28:41.794301 containerd[1514]: time="2025-07-12T09:28:41.794194360Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jul 12 09:28:41.794301 containerd[1514]: time="2025-07-12T09:28:41.794285120Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jul 12 09:28:41.794385 containerd[1514]: time="2025-07-12T09:28:41.794306920Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jul 12 09:28:41.794385 containerd[1514]: time="2025-07-12T09:28:41.794316040Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jul 12 09:28:41.794924 containerd[1514]: time="2025-07-12T09:28:41.794877640Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jul 12 09:28:41.795414 containerd[1514]: time="2025-07-12T09:28:41.795381960Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jul 12 09:28:41.795457 containerd[1514]: time="2025-07-12T09:28:41.795436200Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jul 12 09:28:41.795479 containerd[1514]: time="2025-07-12T09:28:41.795456440Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jul 12 09:28:41.795592 containerd[1514]: time="2025-07-12T09:28:41.795573920Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jul 12 09:28:41.796106 containerd[1514]: time="2025-07-12T09:28:41.796083200Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jul 12 09:28:41.796356 containerd[1514]: time="2025-07-12T09:28:41.796161480Z" level=info msg="metadata content store policy set" policy=shared Jul 12 09:28:41.799434 containerd[1514]: time="2025-07-12T09:28:41.799399640Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jul 12 09:28:41.799487 containerd[1514]: time="2025-07-12T09:28:41.799471680Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jul 12 09:28:41.799511 containerd[1514]: time="2025-07-12T09:28:41.799488760Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jul 12 09:28:41.799511 containerd[1514]: time="2025-07-12T09:28:41.799500600Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jul 12 09:28:41.799554 containerd[1514]: time="2025-07-12T09:28:41.799511640Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jul 12 09:28:41.799554 containerd[1514]: time="2025-07-12T09:28:41.799521360Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jul 12 09:28:41.799554 containerd[1514]: time="2025-07-12T09:28:41.799531520Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jul 12 09:28:41.799554 containerd[1514]: time="2025-07-12T09:28:41.799542600Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jul 12 09:28:41.799554 containerd[1514]: time="2025-07-12T09:28:41.799552440Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jul 12 09:28:41.799627 containerd[1514]: time="2025-07-12T09:28:41.799562440Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jul 12 09:28:41.799627 containerd[1514]: time="2025-07-12T09:28:41.799571120Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jul 12 09:28:41.799627 containerd[1514]: time="2025-07-12T09:28:41.799582200Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jul 12 09:28:41.799721 containerd[1514]: time="2025-07-12T09:28:41.799693600Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jul 12 09:28:41.799782 containerd[1514]: time="2025-07-12T09:28:41.799721120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jul 12 09:28:41.799870 containerd[1514]: time="2025-07-12T09:28:41.799744640Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jul 12 09:28:41.799896 containerd[1514]: time="2025-07-12T09:28:41.799870160Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jul 12 09:28:41.799896 containerd[1514]: time="2025-07-12T09:28:41.799882960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jul 12 09:28:41.799896 containerd[1514]: time="2025-07-12T09:28:41.799893000Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jul 12 09:28:41.799971 containerd[1514]: time="2025-07-12T09:28:41.799903280Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jul 12 09:28:41.799971 containerd[1514]: time="2025-07-12T09:28:41.799936560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jul 12 09:28:41.800021 containerd[1514]: time="2025-07-12T09:28:41.800005240Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jul 12 09:28:41.800040 containerd[1514]: time="2025-07-12T09:28:41.800029520Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jul 12 09:28:41.800057 containerd[1514]: time="2025-07-12T09:28:41.800042800Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jul 12 09:28:41.800292 containerd[1514]: time="2025-07-12T09:28:41.800262720Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jul 12 09:28:41.800292 containerd[1514]: time="2025-07-12T09:28:41.800287520Z" level=info msg="Start snapshots syncer" Jul 12 09:28:41.800338 containerd[1514]: time="2025-07-12T09:28:41.800320280Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jul 12 09:28:41.800604 containerd[1514]: time="2025-07-12T09:28:41.800509520Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jul 12 09:28:41.800713 containerd[1514]: time="2025-07-12T09:28:41.800624040Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jul 12 09:28:41.800713 containerd[1514]: time="2025-07-12T09:28:41.800694920Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jul 12 09:28:41.800869 containerd[1514]: time="2025-07-12T09:28:41.800792600Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jul 12 09:28:41.800896 containerd[1514]: time="2025-07-12T09:28:41.800880360Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jul 12 09:28:41.800956 containerd[1514]: time="2025-07-12T09:28:41.800893840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jul 12 09:28:41.800956 containerd[1514]: time="2025-07-12T09:28:41.800904640Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jul 12 09:28:41.800956 containerd[1514]: time="2025-07-12T09:28:41.800934360Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jul 12 09:28:41.800956 containerd[1514]: time="2025-07-12T09:28:41.800947200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jul 12 09:28:41.801043 containerd[1514]: time="2025-07-12T09:28:41.800957960Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jul 12 09:28:41.801043 containerd[1514]: time="2025-07-12T09:28:41.800980200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jul 12 09:28:41.801043 containerd[1514]: time="2025-07-12T09:28:41.800998800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jul 12 09:28:41.801043 containerd[1514]: time="2025-07-12T09:28:41.801013640Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jul 12 09:28:41.801957 containerd[1514]: time="2025-07-12T09:28:41.801927120Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jul 12 09:28:41.801995 containerd[1514]: time="2025-07-12T09:28:41.801958760Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jul 12 09:28:41.801995 containerd[1514]: time="2025-07-12T09:28:41.801971200Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jul 12 09:28:41.801995 containerd[1514]: time="2025-07-12T09:28:41.801980800Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jul 12 09:28:41.802047 containerd[1514]: time="2025-07-12T09:28:41.801988120Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jul 12 09:28:41.802047 containerd[1514]: time="2025-07-12T09:28:41.802007800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jul 12 09:28:41.802088 containerd[1514]: time="2025-07-12T09:28:41.802071760Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jul 12 09:28:41.802176 containerd[1514]: time="2025-07-12T09:28:41.802152600Z" level=info msg="runtime interface created" Jul 12 09:28:41.802176 containerd[1514]: time="2025-07-12T09:28:41.802164480Z" level=info msg="created NRI interface" Jul 12 09:28:41.802240 containerd[1514]: time="2025-07-12T09:28:41.802222680Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jul 12 09:28:41.802260 containerd[1514]: time="2025-07-12T09:28:41.802242280Z" level=info msg="Connect containerd service" Jul 12 09:28:41.802287 containerd[1514]: time="2025-07-12T09:28:41.802277200Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jul 12 09:28:41.803242 containerd[1514]: time="2025-07-12T09:28:41.803214120Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 12 09:28:41.840273 tar[1503]: linux-arm64/LICENSE Jul 12 09:28:41.840273 tar[1503]: linux-arm64/README.md Jul 12 09:28:41.858377 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jul 12 09:28:41.906570 containerd[1514]: time="2025-07-12T09:28:41.906513880Z" level=info msg="Start subscribing containerd event" Jul 12 09:28:41.906735 containerd[1514]: time="2025-07-12T09:28:41.906697960Z" level=info msg="Start recovering state" Jul 12 09:28:41.906875 containerd[1514]: time="2025-07-12T09:28:41.906834200Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 12 09:28:41.906908 containerd[1514]: time="2025-07-12T09:28:41.906895080Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 12 09:28:41.909101 containerd[1514]: time="2025-07-12T09:28:41.909067160Z" level=info msg="Start event monitor" Jul 12 09:28:41.909101 containerd[1514]: time="2025-07-12T09:28:41.909099640Z" level=info msg="Start cni network conf syncer for default" Jul 12 09:28:41.909148 containerd[1514]: time="2025-07-12T09:28:41.909109120Z" level=info msg="Start streaming server" Jul 12 09:28:41.909148 containerd[1514]: time="2025-07-12T09:28:41.909123560Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jul 12 09:28:41.909148 containerd[1514]: time="2025-07-12T09:28:41.909131320Z" level=info msg="runtime interface starting up..." Jul 12 09:28:41.909148 containerd[1514]: time="2025-07-12T09:28:41.909138600Z" level=info msg="starting plugins..." Jul 12 09:28:41.909227 containerd[1514]: time="2025-07-12T09:28:41.909157160Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jul 12 09:28:41.910171 containerd[1514]: time="2025-07-12T09:28:41.909290240Z" level=info msg="containerd successfully booted in 0.132269s" Jul 12 09:28:41.909379 systemd[1]: Started containerd.service - containerd container runtime. Jul 12 09:28:42.112660 sshd_keygen[1501]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 12 09:28:42.131965 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jul 12 09:28:42.134899 systemd[1]: Starting issuegen.service - Generate /run/issue... Jul 12 09:28:42.155757 systemd[1]: issuegen.service: Deactivated successfully. Jul 12 09:28:42.155975 systemd[1]: Finished issuegen.service - Generate /run/issue. Jul 12 09:28:42.159131 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jul 12 09:28:42.177055 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jul 12 09:28:42.179902 systemd[1]: Started getty@tty1.service - Getty on tty1. Jul 12 09:28:42.181721 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Jul 12 09:28:42.182774 systemd[1]: Reached target getty.target - Login Prompts. Jul 12 09:28:42.551048 systemd-networkd[1432]: eth0: Gained IPv6LL Jul 12 09:28:42.553181 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jul 12 09:28:42.554442 systemd[1]: Reached target network-online.target - Network is Online. Jul 12 09:28:42.556394 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jul 12 09:28:42.558342 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:28:42.559984 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jul 12 09:28:42.589101 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jul 12 09:28:42.590346 systemd[1]: coreos-metadata.service: Deactivated successfully. Jul 12 09:28:42.590524 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jul 12 09:28:42.592187 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jul 12 09:28:43.106446 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:28:43.107621 systemd[1]: Reached target multi-user.target - Multi-User System. Jul 12 09:28:43.108561 systemd[1]: Startup finished in 2.030s (kernel) + 5.219s (initrd) + 3.323s (userspace) = 10.573s. Jul 12 09:28:43.109662 (kubelet)[1616]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 12 09:28:43.522312 kubelet[1616]: E0712 09:28:43.522267 1616 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 12 09:28:43.524788 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 12 09:28:43.524941 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 12 09:28:43.525210 systemd[1]: kubelet.service: Consumed 812ms CPU time, 256.7M memory peak. Jul 12 09:28:46.951130 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jul 12 09:28:46.952152 systemd[1]: Started sshd@0-10.0.0.27:22-10.0.0.1:41294.service - OpenSSH per-connection server daemon (10.0.0.1:41294). Jul 12 09:28:47.017354 sshd[1631]: Accepted publickey for core from 10.0.0.1 port 41294 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:47.019269 sshd-session[1631]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:47.024896 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jul 12 09:28:47.025760 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jul 12 09:28:47.032237 systemd-logind[1493]: New session 1 of user core. Jul 12 09:28:47.043165 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jul 12 09:28:47.045621 systemd[1]: Starting user@500.service - User Manager for UID 500... Jul 12 09:28:47.061215 (systemd)[1636]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 12 09:28:47.063565 systemd-logind[1493]: New session c1 of user core. Jul 12 09:28:47.166589 systemd[1636]: Queued start job for default target default.target. Jul 12 09:28:47.173760 systemd[1636]: Created slice app.slice - User Application Slice. Jul 12 09:28:47.173785 systemd[1636]: Reached target paths.target - Paths. Jul 12 09:28:47.173820 systemd[1636]: Reached target timers.target - Timers. Jul 12 09:28:47.175089 systemd[1636]: Starting dbus.socket - D-Bus User Message Bus Socket... Jul 12 09:28:47.183763 systemd[1636]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jul 12 09:28:47.183809 systemd[1636]: Reached target sockets.target - Sockets. Jul 12 09:28:47.183841 systemd[1636]: Reached target basic.target - Basic System. Jul 12 09:28:47.183866 systemd[1636]: Reached target default.target - Main User Target. Jul 12 09:28:47.183891 systemd[1636]: Startup finished in 114ms. Jul 12 09:28:47.184038 systemd[1]: Started user@500.service - User Manager for UID 500. Jul 12 09:28:47.185148 systemd[1]: Started session-1.scope - Session 1 of User core. Jul 12 09:28:47.253758 systemd[1]: Started sshd@1-10.0.0.27:22-10.0.0.1:41310.service - OpenSSH per-connection server daemon (10.0.0.1:41310). Jul 12 09:28:47.309540 sshd[1647]: Accepted publickey for core from 10.0.0.1 port 41310 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:47.310809 sshd-session[1647]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:47.314340 systemd-logind[1493]: New session 2 of user core. Jul 12 09:28:47.330096 systemd[1]: Started session-2.scope - Session 2 of User core. Jul 12 09:28:47.379977 sshd[1650]: Connection closed by 10.0.0.1 port 41310 Jul 12 09:28:47.380145 sshd-session[1647]: pam_unix(sshd:session): session closed for user core Jul 12 09:28:47.389757 systemd[1]: sshd@1-10.0.0.27:22-10.0.0.1:41310.service: Deactivated successfully. Jul 12 09:28:47.391161 systemd[1]: session-2.scope: Deactivated successfully. Jul 12 09:28:47.391775 systemd-logind[1493]: Session 2 logged out. Waiting for processes to exit. Jul 12 09:28:47.393871 systemd[1]: Started sshd@2-10.0.0.27:22-10.0.0.1:41324.service - OpenSSH per-connection server daemon (10.0.0.1:41324). Jul 12 09:28:47.394788 systemd-logind[1493]: Removed session 2. Jul 12 09:28:47.454355 sshd[1656]: Accepted publickey for core from 10.0.0.1 port 41324 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:47.455592 sshd-session[1656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:47.460092 systemd-logind[1493]: New session 3 of user core. Jul 12 09:28:47.476111 systemd[1]: Started session-3.scope - Session 3 of User core. Jul 12 09:28:47.523895 sshd[1659]: Connection closed by 10.0.0.1 port 41324 Jul 12 09:28:47.524191 sshd-session[1656]: pam_unix(sshd:session): session closed for user core Jul 12 09:28:47.536993 systemd[1]: sshd@2-10.0.0.27:22-10.0.0.1:41324.service: Deactivated successfully. Jul 12 09:28:47.538503 systemd[1]: session-3.scope: Deactivated successfully. Jul 12 09:28:47.540538 systemd-logind[1493]: Session 3 logged out. Waiting for processes to exit. Jul 12 09:28:47.542771 systemd[1]: Started sshd@3-10.0.0.27:22-10.0.0.1:41330.service - OpenSSH per-connection server daemon (10.0.0.1:41330). Jul 12 09:28:47.543494 systemd-logind[1493]: Removed session 3. Jul 12 09:28:47.601098 sshd[1665]: Accepted publickey for core from 10.0.0.1 port 41330 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:47.602464 sshd-session[1665]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:47.607262 systemd-logind[1493]: New session 4 of user core. Jul 12 09:28:47.625076 systemd[1]: Started session-4.scope - Session 4 of User core. Jul 12 09:28:47.675652 sshd[1668]: Connection closed by 10.0.0.1 port 41330 Jul 12 09:28:47.675970 sshd-session[1665]: pam_unix(sshd:session): session closed for user core Jul 12 09:28:47.691062 systemd[1]: sshd@3-10.0.0.27:22-10.0.0.1:41330.service: Deactivated successfully. Jul 12 09:28:47.694373 systemd[1]: session-4.scope: Deactivated successfully. Jul 12 09:28:47.695105 systemd-logind[1493]: Session 4 logged out. Waiting for processes to exit. Jul 12 09:28:47.697700 systemd[1]: Started sshd@4-10.0.0.27:22-10.0.0.1:41342.service - OpenSSH per-connection server daemon (10.0.0.1:41342). Jul 12 09:28:47.698415 systemd-logind[1493]: Removed session 4. Jul 12 09:28:47.759640 sshd[1674]: Accepted publickey for core from 10.0.0.1 port 41342 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:47.760848 sshd-session[1674]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:47.764980 systemd-logind[1493]: New session 5 of user core. Jul 12 09:28:47.782087 systemd[1]: Started session-5.scope - Session 5 of User core. Jul 12 09:28:47.838101 sudo[1678]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jul 12 09:28:47.838364 sudo[1678]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 09:28:47.853755 sudo[1678]: pam_unix(sudo:session): session closed for user root Jul 12 09:28:47.855744 sshd[1677]: Connection closed by 10.0.0.1 port 41342 Jul 12 09:28:47.855641 sshd-session[1674]: pam_unix(sshd:session): session closed for user core Jul 12 09:28:47.870503 systemd[1]: sshd@4-10.0.0.27:22-10.0.0.1:41342.service: Deactivated successfully. Jul 12 09:28:47.872023 systemd[1]: session-5.scope: Deactivated successfully. Jul 12 09:28:47.874048 systemd-logind[1493]: Session 5 logged out. Waiting for processes to exit. Jul 12 09:28:47.876411 systemd[1]: Started sshd@5-10.0.0.27:22-10.0.0.1:41356.service - OpenSSH per-connection server daemon (10.0.0.1:41356). Jul 12 09:28:47.876884 systemd-logind[1493]: Removed session 5. Jul 12 09:28:47.926353 sshd[1684]: Accepted publickey for core from 10.0.0.1 port 41356 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:47.927717 sshd-session[1684]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:47.931976 systemd-logind[1493]: New session 6 of user core. Jul 12 09:28:47.940089 systemd[1]: Started session-6.scope - Session 6 of User core. Jul 12 09:28:47.990620 sudo[1689]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jul 12 09:28:47.990867 sudo[1689]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 09:28:48.076058 sudo[1689]: pam_unix(sudo:session): session closed for user root Jul 12 09:28:48.081649 sudo[1688]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jul 12 09:28:48.082225 sudo[1688]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 09:28:48.091942 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 12 09:28:48.132940 augenrules[1711]: No rules Jul 12 09:28:48.134093 systemd[1]: audit-rules.service: Deactivated successfully. Jul 12 09:28:48.136049 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 12 09:28:48.136995 sudo[1688]: pam_unix(sudo:session): session closed for user root Jul 12 09:28:48.138256 sshd[1687]: Connection closed by 10.0.0.1 port 41356 Jul 12 09:28:48.138629 sshd-session[1684]: pam_unix(sshd:session): session closed for user core Jul 12 09:28:48.157926 systemd[1]: sshd@5-10.0.0.27:22-10.0.0.1:41356.service: Deactivated successfully. Jul 12 09:28:48.160319 systemd[1]: session-6.scope: Deactivated successfully. Jul 12 09:28:48.161105 systemd-logind[1493]: Session 6 logged out. Waiting for processes to exit. Jul 12 09:28:48.163131 systemd[1]: Started sshd@6-10.0.0.27:22-10.0.0.1:41372.service - OpenSSH per-connection server daemon (10.0.0.1:41372). Jul 12 09:28:48.163624 systemd-logind[1493]: Removed session 6. Jul 12 09:28:48.208049 sshd[1720]: Accepted publickey for core from 10.0.0.1 port 41372 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:28:48.209313 sshd-session[1720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:28:48.220773 systemd-logind[1493]: New session 7 of user core. Jul 12 09:28:48.235076 systemd[1]: Started session-7.scope - Session 7 of User core. Jul 12 09:28:48.286727 sudo[1724]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 12 09:28:48.287010 sudo[1724]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 09:28:48.647381 systemd[1]: Starting docker.service - Docker Application Container Engine... Jul 12 09:28:48.663232 (dockerd)[1745]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jul 12 09:28:48.907122 dockerd[1745]: time="2025-07-12T09:28:48.907002130Z" level=info msg="Starting up" Jul 12 09:28:48.907894 dockerd[1745]: time="2025-07-12T09:28:48.907870017Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jul 12 09:28:48.917186 dockerd[1745]: time="2025-07-12T09:28:48.917141096Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Jul 12 09:28:48.945565 dockerd[1745]: time="2025-07-12T09:28:48.945522512Z" level=info msg="Loading containers: start." Jul 12 09:28:48.952953 kernel: Initializing XFRM netlink socket Jul 12 09:28:49.166863 systemd-networkd[1432]: docker0: Link UP Jul 12 09:28:49.170232 dockerd[1745]: time="2025-07-12T09:28:49.170184914Z" level=info msg="Loading containers: done." Jul 12 09:28:49.183018 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4248619616-merged.mount: Deactivated successfully. Jul 12 09:28:49.184724 dockerd[1745]: time="2025-07-12T09:28:49.184664298Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 12 09:28:49.184809 dockerd[1745]: time="2025-07-12T09:28:49.184754989Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Jul 12 09:28:49.184873 dockerd[1745]: time="2025-07-12T09:28:49.184848244Z" level=info msg="Initializing buildkit" Jul 12 09:28:49.205216 dockerd[1745]: time="2025-07-12T09:28:49.205170300Z" level=info msg="Completed buildkit initialization" Jul 12 09:28:49.212105 dockerd[1745]: time="2025-07-12T09:28:49.212044744Z" level=info msg="Daemon has completed initialization" Jul 12 09:28:49.212242 dockerd[1745]: time="2025-07-12T09:28:49.212144393Z" level=info msg="API listen on /run/docker.sock" Jul 12 09:28:49.212389 systemd[1]: Started docker.service - Docker Application Container Engine. Jul 12 09:28:49.945009 containerd[1514]: time="2025-07-12T09:28:49.944972345Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\"" Jul 12 09:28:50.601446 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount503590416.mount: Deactivated successfully. Jul 12 09:28:51.787075 containerd[1514]: time="2025-07-12T09:28:51.787027732Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:51.788421 containerd[1514]: time="2025-07-12T09:28:51.788365253Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.10: active requests=0, bytes read=25651795" Jul 12 09:28:51.789195 containerd[1514]: time="2025-07-12T09:28:51.789146252Z" level=info msg="ImageCreate event name:\"sha256:8907c2d36348551c1038e24ef688f6830681069380376707e55518007a20a86c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:51.792541 containerd[1514]: time="2025-07-12T09:28:51.792216605Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:51.793204 containerd[1514]: time="2025-07-12T09:28:51.793050338Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.10\" with image id \"sha256:8907c2d36348551c1038e24ef688f6830681069380376707e55518007a20a86c\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\", size \"25648593\" in 1.84803819s" Jul 12 09:28:51.793204 containerd[1514]: time="2025-07-12T09:28:51.793082961Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\" returns image reference \"sha256:8907c2d36348551c1038e24ef688f6830681069380376707e55518007a20a86c\"" Jul 12 09:28:51.796222 containerd[1514]: time="2025-07-12T09:28:51.796194408Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\"" Jul 12 09:28:53.082944 containerd[1514]: time="2025-07-12T09:28:53.082888450Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:53.083631 containerd[1514]: time="2025-07-12T09:28:53.083599273Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.10: active requests=0, bytes read=22459679" Jul 12 09:28:53.084541 containerd[1514]: time="2025-07-12T09:28:53.084476699Z" level=info msg="ImageCreate event name:\"sha256:0f640d6889416d515a0ac4de1c26f4d80134c47641ff464abc831560a951175f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:53.087672 containerd[1514]: time="2025-07-12T09:28:53.087619305Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:53.088930 containerd[1514]: time="2025-07-12T09:28:53.088874228Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.10\" with image id \"sha256:0f640d6889416d515a0ac4de1c26f4d80134c47641ff464abc831560a951175f\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\", size \"23995467\" in 1.292647973s" Jul 12 09:28:53.089324 containerd[1514]: time="2025-07-12T09:28:53.089282164Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\" returns image reference \"sha256:0f640d6889416d515a0ac4de1c26f4d80134c47641ff464abc831560a951175f\"" Jul 12 09:28:53.090099 containerd[1514]: time="2025-07-12T09:28:53.089731573Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\"" Jul 12 09:28:53.739192 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 12 09:28:53.740509 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:28:53.884316 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:28:53.888128 (kubelet)[2030]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 12 09:28:54.031286 kubelet[2030]: E0712 09:28:54.031151 2030 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 12 09:28:54.034697 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 12 09:28:54.034824 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 12 09:28:54.035216 systemd[1]: kubelet.service: Consumed 147ms CPU time, 105.7M memory peak. Jul 12 09:28:54.402190 containerd[1514]: time="2025-07-12T09:28:54.402051211Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:54.402602 containerd[1514]: time="2025-07-12T09:28:54.402395473Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.10: active requests=0, bytes read=17125068" Jul 12 09:28:54.403361 containerd[1514]: time="2025-07-12T09:28:54.403333553Z" level=info msg="ImageCreate event name:\"sha256:23d79b83d912e2633bcb4f9f7b8b46024893e11d492a4249d8f1f8c9a26b7b2c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:54.405770 containerd[1514]: time="2025-07-12T09:28:54.405732819Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:54.406591 containerd[1514]: time="2025-07-12T09:28:54.406557721Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.10\" with image id \"sha256:23d79b83d912e2633bcb4f9f7b8b46024893e11d492a4249d8f1f8c9a26b7b2c\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\", size \"18660874\" in 1.316796852s" Jul 12 09:28:54.406591 containerd[1514]: time="2025-07-12T09:28:54.406587793Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\" returns image reference \"sha256:23d79b83d912e2633bcb4f9f7b8b46024893e11d492a4249d8f1f8c9a26b7b2c\"" Jul 12 09:28:54.407632 containerd[1514]: time="2025-07-12T09:28:54.407465768Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\"" Jul 12 09:28:55.354581 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2408780355.mount: Deactivated successfully. Jul 12 09:28:55.554990 containerd[1514]: time="2025-07-12T09:28:55.554941896Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:55.555675 containerd[1514]: time="2025-07-12T09:28:55.555648310Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.10: active requests=0, bytes read=26915959" Jul 12 09:28:55.556788 containerd[1514]: time="2025-07-12T09:28:55.556458537Z" level=info msg="ImageCreate event name:\"sha256:dde5ff0da443b455e81aefc7bf6a216fdd659d1cbe13b8e8ac8129c3ecd27f89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:55.558642 containerd[1514]: time="2025-07-12T09:28:55.558607638Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:55.559760 containerd[1514]: time="2025-07-12T09:28:55.559731888Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.10\" with image id \"sha256:dde5ff0da443b455e81aefc7bf6a216fdd659d1cbe13b8e8ac8129c3ecd27f89\", repo tag \"registry.k8s.io/kube-proxy:v1.31.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\", size \"26914976\" in 1.152239339s" Jul 12 09:28:55.559853 containerd[1514]: time="2025-07-12T09:28:55.559837689Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\" returns image reference \"sha256:dde5ff0da443b455e81aefc7bf6a216fdd659d1cbe13b8e8ac8129c3ecd27f89\"" Jul 12 09:28:55.560548 containerd[1514]: time="2025-07-12T09:28:55.560529193Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Jul 12 09:28:56.135594 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2101772175.mount: Deactivated successfully. Jul 12 09:28:56.865556 containerd[1514]: time="2025-07-12T09:28:56.865503182Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:56.869948 containerd[1514]: time="2025-07-12T09:28:56.869280835Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951624" Jul 12 09:28:56.869948 containerd[1514]: time="2025-07-12T09:28:56.869521086Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:56.873255 containerd[1514]: time="2025-07-12T09:28:56.873215697Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:56.874987 containerd[1514]: time="2025-07-12T09:28:56.874947314Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.314318277s" Jul 12 09:28:56.875036 containerd[1514]: time="2025-07-12T09:28:56.874987661Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Jul 12 09:28:56.875513 containerd[1514]: time="2025-07-12T09:28:56.875483682Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jul 12 09:28:57.406970 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2493829666.mount: Deactivated successfully. Jul 12 09:28:57.537639 containerd[1514]: time="2025-07-12T09:28:57.537568539Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 09:28:57.541428 containerd[1514]: time="2025-07-12T09:28:57.541383843Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Jul 12 09:28:57.543254 containerd[1514]: time="2025-07-12T09:28:57.543202447Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 09:28:57.547638 containerd[1514]: time="2025-07-12T09:28:57.547582105Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 09:28:57.548218 containerd[1514]: time="2025-07-12T09:28:57.548179905Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 672.664902ms" Jul 12 09:28:57.548252 containerd[1514]: time="2025-07-12T09:28:57.548214704Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Jul 12 09:28:57.548802 containerd[1514]: time="2025-07-12T09:28:57.548618400Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Jul 12 09:28:58.048297 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3176523536.mount: Deactivated successfully. Jul 12 09:28:59.827592 containerd[1514]: time="2025-07-12T09:28:59.827531444Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:59.828047 containerd[1514]: time="2025-07-12T09:28:59.828013541Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406467" Jul 12 09:28:59.829022 containerd[1514]: time="2025-07-12T09:28:59.828972754Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:59.831896 containerd[1514]: time="2025-07-12T09:28:59.831845051Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:28:59.833238 containerd[1514]: time="2025-07-12T09:28:59.833195881Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 2.28454781s" Jul 12 09:28:59.833238 containerd[1514]: time="2025-07-12T09:28:59.833235103Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Jul 12 09:29:04.239208 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jul 12 09:29:04.240666 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:29:04.367388 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jul 12 09:29:04.367487 systemd[1]: kubelet.service: Failed with result 'signal'. Jul 12 09:29:04.367799 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:29:04.368152 systemd[1]: kubelet.service: Consumed 59ms CPU time, 70.1M memory peak. Jul 12 09:29:04.371699 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:29:04.388841 systemd[1]: Reload requested from client PID 2192 ('systemctl') (unit session-7.scope)... Jul 12 09:29:04.388856 systemd[1]: Reloading... Jul 12 09:29:04.449941 zram_generator::config[2235]: No configuration found. Jul 12 09:29:04.576691 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 09:29:04.660400 systemd[1]: Reloading finished in 271 ms. Jul 12 09:29:04.713448 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:29:04.715620 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:29:04.717325 systemd[1]: kubelet.service: Deactivated successfully. Jul 12 09:29:04.717520 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:29:04.717559 systemd[1]: kubelet.service: Consumed 91ms CPU time, 95.1M memory peak. Jul 12 09:29:04.718964 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:29:04.879386 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:29:04.883677 (kubelet)[2283]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 12 09:29:05.067730 kubelet[2283]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 09:29:05.067730 kubelet[2283]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 12 09:29:05.067730 kubelet[2283]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 09:29:05.067730 kubelet[2283]: I0712 09:29:05.067163 2283 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 12 09:29:05.445844 kubelet[2283]: I0712 09:29:05.445798 2283 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 12 09:29:05.445844 kubelet[2283]: I0712 09:29:05.445833 2283 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 12 09:29:05.446107 kubelet[2283]: I0712 09:29:05.446078 2283 server.go:934] "Client rotation is on, will bootstrap in background" Jul 12 09:29:05.476876 kubelet[2283]: I0712 09:29:05.476832 2283 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 12 09:29:05.477003 kubelet[2283]: E0712 09:29:05.476891 2283 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.27:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.27:6443: connect: connection refused" logger="UnhandledError" Jul 12 09:29:05.491187 kubelet[2283]: I0712 09:29:05.491157 2283 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jul 12 09:29:05.495948 kubelet[2283]: I0712 09:29:05.495024 2283 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 12 09:29:05.495948 kubelet[2283]: I0712 09:29:05.495577 2283 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 12 09:29:05.495948 kubelet[2283]: I0712 09:29:05.495691 2283 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 12 09:29:05.495948 kubelet[2283]: I0712 09:29:05.495713 2283 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 12 09:29:05.496239 kubelet[2283]: I0712 09:29:05.496212 2283 topology_manager.go:138] "Creating topology manager with none policy" Jul 12 09:29:05.496239 kubelet[2283]: I0712 09:29:05.496227 2283 container_manager_linux.go:300] "Creating device plugin manager" Jul 12 09:29:05.496801 kubelet[2283]: I0712 09:29:05.496777 2283 state_mem.go:36] "Initialized new in-memory state store" Jul 12 09:29:05.499584 kubelet[2283]: I0712 09:29:05.499440 2283 kubelet.go:408] "Attempting to sync node with API server" Jul 12 09:29:05.499584 kubelet[2283]: I0712 09:29:05.499472 2283 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 12 09:29:05.499584 kubelet[2283]: I0712 09:29:05.499494 2283 kubelet.go:314] "Adding apiserver pod source" Jul 12 09:29:05.499584 kubelet[2283]: I0712 09:29:05.499504 2283 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 12 09:29:05.500958 kubelet[2283]: W0712 09:29:05.500881 2283 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.27:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.27:6443: connect: connection refused Jul 12 09:29:05.501013 kubelet[2283]: E0712 09:29:05.500968 2283 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.27:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.27:6443: connect: connection refused" logger="UnhandledError" Jul 12 09:29:05.502069 kubelet[2283]: W0712 09:29:05.502025 2283 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.27:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.27:6443: connect: connection refused Jul 12 09:29:05.502112 kubelet[2283]: E0712 09:29:05.502080 2283 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.27:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.27:6443: connect: connection refused" logger="UnhandledError" Jul 12 09:29:05.510499 kubelet[2283]: I0712 09:29:05.510440 2283 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Jul 12 09:29:05.511668 kubelet[2283]: I0712 09:29:05.511646 2283 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 12 09:29:05.511835 kubelet[2283]: W0712 09:29:05.511813 2283 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 12 09:29:05.513151 kubelet[2283]: I0712 09:29:05.513133 2283 server.go:1274] "Started kubelet" Jul 12 09:29:05.513404 kubelet[2283]: I0712 09:29:05.513372 2283 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 12 09:29:05.518170 kubelet[2283]: I0712 09:29:05.515425 2283 server.go:449] "Adding debug handlers to kubelet server" Jul 12 09:29:05.518170 kubelet[2283]: I0712 09:29:05.515844 2283 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 12 09:29:05.518170 kubelet[2283]: I0712 09:29:05.513556 2283 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 12 09:29:05.518170 kubelet[2283]: I0712 09:29:05.517438 2283 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 12 09:29:05.518170 kubelet[2283]: I0712 09:29:05.517590 2283 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 12 09:29:05.520300 kubelet[2283]: E0712 09:29:05.518203 2283 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.27:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.27:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.185176f9dac4ebf0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-07-12 09:29:05.513106416 +0000 UTC m=+0.626565198,LastTimestamp:2025-07-12 09:29:05.513106416 +0000 UTC m=+0.626565198,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jul 12 09:29:05.520300 kubelet[2283]: I0712 09:29:05.519830 2283 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 12 09:29:05.520300 kubelet[2283]: E0712 09:29:05.519861 2283 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 12 09:29:05.520300 kubelet[2283]: I0712 09:29:05.519880 2283 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 12 09:29:05.520300 kubelet[2283]: I0712 09:29:05.519980 2283 reconciler.go:26] "Reconciler: start to sync state" Jul 12 09:29:05.520670 kubelet[2283]: W0712 09:29:05.520634 2283 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.27:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.27:6443: connect: connection refused Jul 12 09:29:05.520759 kubelet[2283]: E0712 09:29:05.520742 2283 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.27:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.27:6443: connect: connection refused" logger="UnhandledError" Jul 12 09:29:05.520805 kubelet[2283]: E0712 09:29:05.520766 2283 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.27:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.27:6443: connect: connection refused" interval="200ms" Jul 12 09:29:05.521143 kubelet[2283]: I0712 09:29:05.521114 2283 factory.go:221] Registration of the systemd container factory successfully Jul 12 09:29:05.521653 kubelet[2283]: I0712 09:29:05.521628 2283 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 12 09:29:05.522998 kubelet[2283]: I0712 09:29:05.522978 2283 factory.go:221] Registration of the containerd container factory successfully Jul 12 09:29:05.523236 kubelet[2283]: E0712 09:29:05.522983 2283 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 12 09:29:05.535669 kubelet[2283]: I0712 09:29:05.535643 2283 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 12 09:29:05.535669 kubelet[2283]: I0712 09:29:05.535663 2283 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 12 09:29:05.535791 kubelet[2283]: I0712 09:29:05.535680 2283 state_mem.go:36] "Initialized new in-memory state store" Jul 12 09:29:05.537377 kubelet[2283]: I0712 09:29:05.537341 2283 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 12 09:29:05.538538 kubelet[2283]: I0712 09:29:05.538515 2283 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 12 09:29:05.538538 kubelet[2283]: I0712 09:29:05.538545 2283 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 12 09:29:05.538630 kubelet[2283]: I0712 09:29:05.538562 2283 kubelet.go:2321] "Starting kubelet main sync loop" Jul 12 09:29:05.538630 kubelet[2283]: E0712 09:29:05.538604 2283 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 12 09:29:05.539131 kubelet[2283]: W0712 09:29:05.539051 2283 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.27:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.27:6443: connect: connection refused Jul 12 09:29:05.539345 kubelet[2283]: E0712 09:29:05.539142 2283 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.27:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.27:6443: connect: connection refused" logger="UnhandledError" Jul 12 09:29:05.539620 kubelet[2283]: I0712 09:29:05.539603 2283 policy_none.go:49] "None policy: Start" Jul 12 09:29:05.540099 kubelet[2283]: I0712 09:29:05.540057 2283 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 12 09:29:05.540099 kubelet[2283]: I0712 09:29:05.540080 2283 state_mem.go:35] "Initializing new in-memory state store" Jul 12 09:29:05.546874 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jul 12 09:29:05.560690 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jul 12 09:29:05.572397 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jul 12 09:29:05.574206 kubelet[2283]: I0712 09:29:05.574172 2283 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 12 09:29:05.574724 kubelet[2283]: I0712 09:29:05.574397 2283 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 12 09:29:05.574724 kubelet[2283]: I0712 09:29:05.574414 2283 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 12 09:29:05.574724 kubelet[2283]: I0712 09:29:05.574664 2283 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 12 09:29:05.575746 kubelet[2283]: E0712 09:29:05.575723 2283 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jul 12 09:29:05.646154 systemd[1]: Created slice kubepods-burstable-pod3290bd95f938f7c87adaa2d6d57f009e.slice - libcontainer container kubepods-burstable-pod3290bd95f938f7c87adaa2d6d57f009e.slice. Jul 12 09:29:05.667785 systemd[1]: Created slice kubepods-burstable-pod3f04709fe51ae4ab5abd58e8da771b74.slice - libcontainer container kubepods-burstable-pod3f04709fe51ae4ab5abd58e8da771b74.slice. Jul 12 09:29:05.676483 kubelet[2283]: I0712 09:29:05.676447 2283 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 12 09:29:05.677099 kubelet[2283]: E0712 09:29:05.677062 2283 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.27:6443/api/v1/nodes\": dial tcp 10.0.0.27:6443: connect: connection refused" node="localhost" Jul 12 09:29:05.677438 systemd[1]: Created slice kubepods-burstable-podb35b56493416c25588cb530e37ffc065.slice - libcontainer container kubepods-burstable-podb35b56493416c25588cb530e37ffc065.slice. Jul 12 09:29:05.721004 kubelet[2283]: I0712 09:29:05.720876 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3290bd95f938f7c87adaa2d6d57f009e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3290bd95f938f7c87adaa2d6d57f009e\") " pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:05.721317 kubelet[2283]: I0712 09:29:05.721046 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:05.721317 kubelet[2283]: I0712 09:29:05.721077 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:05.721317 kubelet[2283]: I0712 09:29:05.721094 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3290bd95f938f7c87adaa2d6d57f009e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3290bd95f938f7c87adaa2d6d57f009e\") " pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:05.721317 kubelet[2283]: I0712 09:29:05.721109 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3290bd95f938f7c87adaa2d6d57f009e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3290bd95f938f7c87adaa2d6d57f009e\") " pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:05.721317 kubelet[2283]: I0712 09:29:05.721125 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:05.721447 kubelet[2283]: I0712 09:29:05.721143 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:05.721447 kubelet[2283]: I0712 09:29:05.721161 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:05.721447 kubelet[2283]: I0712 09:29:05.721177 2283 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b35b56493416c25588cb530e37ffc065-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"b35b56493416c25588cb530e37ffc065\") " pod="kube-system/kube-scheduler-localhost" Jul 12 09:29:05.721447 kubelet[2283]: E0712 09:29:05.721270 2283 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.27:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.27:6443: connect: connection refused" interval="400ms" Jul 12 09:29:05.879011 kubelet[2283]: I0712 09:29:05.878963 2283 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 12 09:29:05.879372 kubelet[2283]: E0712 09:29:05.879336 2283 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.27:6443/api/v1/nodes\": dial tcp 10.0.0.27:6443: connect: connection refused" node="localhost" Jul 12 09:29:05.965823 kubelet[2283]: E0712 09:29:05.965791 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:05.966384 containerd[1514]: time="2025-07-12T09:29:05.966347278Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3290bd95f938f7c87adaa2d6d57f009e,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:05.975985 kubelet[2283]: E0712 09:29:05.975565 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:05.976186 containerd[1514]: time="2025-07-12T09:29:05.976129206Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:3f04709fe51ae4ab5abd58e8da771b74,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:05.979451 kubelet[2283]: E0712 09:29:05.979422 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:05.980395 containerd[1514]: time="2025-07-12T09:29:05.979900920Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:b35b56493416c25588cb530e37ffc065,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:05.987807 containerd[1514]: time="2025-07-12T09:29:05.987774387Z" level=info msg="connecting to shim ec7da6fb5d96b2e2627a320065ff6ee6acb4dc7b75dbab3c262ae1d311123c91" address="unix:///run/containerd/s/b220484acaf81752fc74e1b831f04bc66d7515fbeee8df680d2686ff63316ca3" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:06.001438 containerd[1514]: time="2025-07-12T09:29:06.001394219Z" level=info msg="connecting to shim 821255a99854764b98f07bfebc472d695149574096438a70b0daa56847f909b3" address="unix:///run/containerd/s/73fd13fb744b8a8c90c64599f2cd201ecc67ce750890c1928ab913b03ea4f852" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:06.014423 containerd[1514]: time="2025-07-12T09:29:06.014387280Z" level=info msg="connecting to shim ff767749860742ad7a1cb228ae364fafcfbfeb52b42ee8b51b52d4d8acf8cac2" address="unix:///run/containerd/s/2a16becf7e7fce2da19b0d729ff74c67bf38b008d04ed02a5dc0315c30cb12e3" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:06.021085 systemd[1]: Started cri-containerd-ec7da6fb5d96b2e2627a320065ff6ee6acb4dc7b75dbab3c262ae1d311123c91.scope - libcontainer container ec7da6fb5d96b2e2627a320065ff6ee6acb4dc7b75dbab3c262ae1d311123c91. Jul 12 09:29:06.027263 systemd[1]: Started cri-containerd-821255a99854764b98f07bfebc472d695149574096438a70b0daa56847f909b3.scope - libcontainer container 821255a99854764b98f07bfebc472d695149574096438a70b0daa56847f909b3. Jul 12 09:29:06.045080 systemd[1]: Started cri-containerd-ff767749860742ad7a1cb228ae364fafcfbfeb52b42ee8b51b52d4d8acf8cac2.scope - libcontainer container ff767749860742ad7a1cb228ae364fafcfbfeb52b42ee8b51b52d4d8acf8cac2. Jul 12 09:29:06.062815 containerd[1514]: time="2025-07-12T09:29:06.062676301Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3290bd95f938f7c87adaa2d6d57f009e,Namespace:kube-system,Attempt:0,} returns sandbox id \"ec7da6fb5d96b2e2627a320065ff6ee6acb4dc7b75dbab3c262ae1d311123c91\"" Jul 12 09:29:06.064252 kubelet[2283]: E0712 09:29:06.064219 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:06.066053 containerd[1514]: time="2025-07-12T09:29:06.066015041Z" level=info msg="CreateContainer within sandbox \"ec7da6fb5d96b2e2627a320065ff6ee6acb4dc7b75dbab3c262ae1d311123c91\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 12 09:29:06.072754 containerd[1514]: time="2025-07-12T09:29:06.072714891Z" level=info msg="Container 2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:06.081106 containerd[1514]: time="2025-07-12T09:29:06.081040890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:3f04709fe51ae4ab5abd58e8da771b74,Namespace:kube-system,Attempt:0,} returns sandbox id \"821255a99854764b98f07bfebc472d695149574096438a70b0daa56847f909b3\"" Jul 12 09:29:06.082059 kubelet[2283]: E0712 09:29:06.082023 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:06.084074 containerd[1514]: time="2025-07-12T09:29:06.084033149Z" level=info msg="CreateContainer within sandbox \"821255a99854764b98f07bfebc472d695149574096438a70b0daa56847f909b3\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 12 09:29:06.085134 containerd[1514]: time="2025-07-12T09:29:06.084958789Z" level=info msg="CreateContainer within sandbox \"ec7da6fb5d96b2e2627a320065ff6ee6acb4dc7b75dbab3c262ae1d311123c91\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6\"" Jul 12 09:29:06.085658 containerd[1514]: time="2025-07-12T09:29:06.085633095Z" level=info msg="StartContainer for \"2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6\"" Jul 12 09:29:06.086705 containerd[1514]: time="2025-07-12T09:29:06.086673216Z" level=info msg="connecting to shim 2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6" address="unix:///run/containerd/s/b220484acaf81752fc74e1b831f04bc66d7515fbeee8df680d2686ff63316ca3" protocol=ttrpc version=3 Jul 12 09:29:06.090253 containerd[1514]: time="2025-07-12T09:29:06.090223703Z" level=info msg="Container 24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:06.090590 containerd[1514]: time="2025-07-12T09:29:06.090559958Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:b35b56493416c25588cb530e37ffc065,Namespace:kube-system,Attempt:0,} returns sandbox id \"ff767749860742ad7a1cb228ae364fafcfbfeb52b42ee8b51b52d4d8acf8cac2\"" Jul 12 09:29:06.091213 kubelet[2283]: E0712 09:29:06.091161 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:06.092576 containerd[1514]: time="2025-07-12T09:29:06.092549605Z" level=info msg="CreateContainer within sandbox \"ff767749860742ad7a1cb228ae364fafcfbfeb52b42ee8b51b52d4d8acf8cac2\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 12 09:29:06.099684 containerd[1514]: time="2025-07-12T09:29:06.099156942Z" level=info msg="Container 0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:06.104105 containerd[1514]: time="2025-07-12T09:29:06.104071142Z" level=info msg="CreateContainer within sandbox \"821255a99854764b98f07bfebc472d695149574096438a70b0daa56847f909b3\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d\"" Jul 12 09:29:06.105085 systemd[1]: Started cri-containerd-2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6.scope - libcontainer container 2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6. Jul 12 09:29:06.105596 containerd[1514]: time="2025-07-12T09:29:06.105570028Z" level=info msg="StartContainer for \"24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d\"" Jul 12 09:29:06.107494 containerd[1514]: time="2025-07-12T09:29:06.106964259Z" level=info msg="connecting to shim 24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d" address="unix:///run/containerd/s/73fd13fb744b8a8c90c64599f2cd201ecc67ce750890c1928ab913b03ea4f852" protocol=ttrpc version=3 Jul 12 09:29:06.112716 containerd[1514]: time="2025-07-12T09:29:06.112675516Z" level=info msg="CreateContainer within sandbox \"ff767749860742ad7a1cb228ae364fafcfbfeb52b42ee8b51b52d4d8acf8cac2\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c\"" Jul 12 09:29:06.113143 containerd[1514]: time="2025-07-12T09:29:06.113114509Z" level=info msg="StartContainer for \"0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c\"" Jul 12 09:29:06.114517 containerd[1514]: time="2025-07-12T09:29:06.114456052Z" level=info msg="connecting to shim 0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c" address="unix:///run/containerd/s/2a16becf7e7fce2da19b0d729ff74c67bf38b008d04ed02a5dc0315c30cb12e3" protocol=ttrpc version=3 Jul 12 09:29:06.122819 kubelet[2283]: E0712 09:29:06.122777 2283 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.27:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.27:6443: connect: connection refused" interval="800ms" Jul 12 09:29:06.125296 systemd[1]: Started cri-containerd-24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d.scope - libcontainer container 24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d. Jul 12 09:29:06.128752 systemd[1]: Started cri-containerd-0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c.scope - libcontainer container 0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c. Jul 12 09:29:06.156527 containerd[1514]: time="2025-07-12T09:29:06.156423820Z" level=info msg="StartContainer for \"2f7dc6453bd3427309640bc6c63b6711059ef268ae37a95a9fefc7a0cc08bec6\" returns successfully" Jul 12 09:29:06.190681 containerd[1514]: time="2025-07-12T09:29:06.190645467Z" level=info msg="StartContainer for \"24dd9e60d57ff17be7e847782f6239f096f0521bc0aa3c367a6eedd13501b97d\" returns successfully" Jul 12 09:29:06.196385 containerd[1514]: time="2025-07-12T09:29:06.195488246Z" level=info msg="StartContainer for \"0f925fe050b39809802a54f6bc56efc514dbd06f1302a3f66f7e692bf0b9837c\" returns successfully" Jul 12 09:29:06.283581 kubelet[2283]: I0712 09:29:06.283457 2283 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 12 09:29:06.284027 kubelet[2283]: E0712 09:29:06.283996 2283 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.27:6443/api/v1/nodes\": dial tcp 10.0.0.27:6443: connect: connection refused" node="localhost" Jul 12 09:29:06.550207 kubelet[2283]: E0712 09:29:06.549982 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:06.555899 kubelet[2283]: E0712 09:29:06.555869 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:06.556438 kubelet[2283]: E0712 09:29:06.556421 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:07.085919 kubelet[2283]: I0712 09:29:07.085876 2283 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 12 09:29:07.558707 kubelet[2283]: E0712 09:29:07.558606 2283 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:08.828526 kubelet[2283]: E0712 09:29:08.828481 2283 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jul 12 09:29:08.906148 kubelet[2283]: I0712 09:29:08.906101 2283 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Jul 12 09:29:09.502993 kubelet[2283]: I0712 09:29:09.502953 2283 apiserver.go:52] "Watching apiserver" Jul 12 09:29:09.520984 kubelet[2283]: I0712 09:29:09.520940 2283 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 12 09:29:11.014610 systemd[1]: Reload requested from client PID 2558 ('systemctl') (unit session-7.scope)... Jul 12 09:29:11.014623 systemd[1]: Reloading... Jul 12 09:29:11.068945 zram_generator::config[2601]: No configuration found. Jul 12 09:29:11.138776 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 09:29:11.233285 systemd[1]: Reloading finished in 218 ms. Jul 12 09:29:11.263603 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:29:11.285373 systemd[1]: kubelet.service: Deactivated successfully. Jul 12 09:29:11.285576 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:29:11.285617 systemd[1]: kubelet.service: Consumed 892ms CPU time, 127.8M memory peak. Jul 12 09:29:11.287718 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 09:29:11.421205 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 09:29:11.425327 (kubelet)[2643]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 12 09:29:11.461549 kubelet[2643]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 09:29:11.461549 kubelet[2643]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 12 09:29:11.461549 kubelet[2643]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 09:29:11.462102 kubelet[2643]: I0712 09:29:11.461613 2643 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 12 09:29:11.469068 kubelet[2643]: I0712 09:29:11.469025 2643 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 12 09:29:11.469068 kubelet[2643]: I0712 09:29:11.469056 2643 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 12 09:29:11.469294 kubelet[2643]: I0712 09:29:11.469267 2643 server.go:934] "Client rotation is on, will bootstrap in background" Jul 12 09:29:11.472083 kubelet[2643]: I0712 09:29:11.472058 2643 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jul 12 09:29:11.473893 kubelet[2643]: I0712 09:29:11.473866 2643 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 12 09:29:11.477945 kubelet[2643]: I0712 09:29:11.477778 2643 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jul 12 09:29:11.480246 kubelet[2643]: I0712 09:29:11.480224 2643 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 12 09:29:11.480372 kubelet[2643]: I0712 09:29:11.480358 2643 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 12 09:29:11.480480 kubelet[2643]: I0712 09:29:11.480454 2643 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 12 09:29:11.480646 kubelet[2643]: I0712 09:29:11.480483 2643 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 12 09:29:11.480722 kubelet[2643]: I0712 09:29:11.480653 2643 topology_manager.go:138] "Creating topology manager with none policy" Jul 12 09:29:11.480722 kubelet[2643]: I0712 09:29:11.480661 2643 container_manager_linux.go:300] "Creating device plugin manager" Jul 12 09:29:11.480722 kubelet[2643]: I0712 09:29:11.480693 2643 state_mem.go:36] "Initialized new in-memory state store" Jul 12 09:29:11.480812 kubelet[2643]: I0712 09:29:11.480798 2643 kubelet.go:408] "Attempting to sync node with API server" Jul 12 09:29:11.480838 kubelet[2643]: I0712 09:29:11.480814 2643 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 12 09:29:11.480838 kubelet[2643]: I0712 09:29:11.480837 2643 kubelet.go:314] "Adding apiserver pod source" Jul 12 09:29:11.480887 kubelet[2643]: I0712 09:29:11.480851 2643 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 12 09:29:11.481602 kubelet[2643]: I0712 09:29:11.481575 2643 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Jul 12 09:29:11.484927 kubelet[2643]: I0712 09:29:11.484765 2643 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 12 09:29:11.486419 kubelet[2643]: I0712 09:29:11.486382 2643 server.go:1274] "Started kubelet" Jul 12 09:29:11.489004 kubelet[2643]: I0712 09:29:11.488303 2643 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 12 09:29:11.489366 kubelet[2643]: I0712 09:29:11.489322 2643 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 12 09:29:11.489600 kubelet[2643]: I0712 09:29:11.489583 2643 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 12 09:29:11.490541 kubelet[2643]: I0712 09:29:11.490506 2643 server.go:449] "Adding debug handlers to kubelet server" Jul 12 09:29:11.493276 kubelet[2643]: I0712 09:29:11.493255 2643 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 12 09:29:11.496237 kubelet[2643]: I0712 09:29:11.496211 2643 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 12 09:29:11.498747 kubelet[2643]: E0712 09:29:11.498711 2643 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 12 09:29:11.499124 kubelet[2643]: I0712 09:29:11.499105 2643 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 12 09:29:11.499557 kubelet[2643]: I0712 09:29:11.499537 2643 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 12 09:29:11.499764 kubelet[2643]: I0712 09:29:11.499750 2643 reconciler.go:26] "Reconciler: start to sync state" Jul 12 09:29:11.500936 kubelet[2643]: I0712 09:29:11.500194 2643 factory.go:221] Registration of the systemd container factory successfully Jul 12 09:29:11.500936 kubelet[2643]: I0712 09:29:11.500289 2643 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 12 09:29:11.502002 kubelet[2643]: I0712 09:29:11.501951 2643 factory.go:221] Registration of the containerd container factory successfully Jul 12 09:29:11.511739 kubelet[2643]: I0712 09:29:11.511685 2643 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 12 09:29:11.512864 kubelet[2643]: I0712 09:29:11.512842 2643 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 12 09:29:11.513072 kubelet[2643]: I0712 09:29:11.513056 2643 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 12 09:29:11.513154 kubelet[2643]: I0712 09:29:11.513145 2643 kubelet.go:2321] "Starting kubelet main sync loop" Jul 12 09:29:11.513283 kubelet[2643]: E0712 09:29:11.513265 2643 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 12 09:29:11.537366 kubelet[2643]: I0712 09:29:11.536951 2643 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 12 09:29:11.537366 kubelet[2643]: I0712 09:29:11.536974 2643 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 12 09:29:11.537366 kubelet[2643]: I0712 09:29:11.536998 2643 state_mem.go:36] "Initialized new in-memory state store" Jul 12 09:29:11.537366 kubelet[2643]: I0712 09:29:11.537159 2643 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 12 09:29:11.537366 kubelet[2643]: I0712 09:29:11.537170 2643 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 12 09:29:11.537366 kubelet[2643]: I0712 09:29:11.537186 2643 policy_none.go:49] "None policy: Start" Jul 12 09:29:11.539494 kubelet[2643]: I0712 09:29:11.539477 2643 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 12 09:29:11.539938 kubelet[2643]: I0712 09:29:11.539656 2643 state_mem.go:35] "Initializing new in-memory state store" Jul 12 09:29:11.539938 kubelet[2643]: I0712 09:29:11.539858 2643 state_mem.go:75] "Updated machine memory state" Jul 12 09:29:11.544168 kubelet[2643]: I0712 09:29:11.544144 2643 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 12 09:29:11.544571 kubelet[2643]: I0712 09:29:11.544459 2643 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 12 09:29:11.544685 kubelet[2643]: I0712 09:29:11.544653 2643 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 12 09:29:11.545082 kubelet[2643]: I0712 09:29:11.545063 2643 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 12 09:29:11.651393 kubelet[2643]: I0712 09:29:11.651357 2643 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 12 09:29:11.658112 kubelet[2643]: I0712 09:29:11.658082 2643 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Jul 12 09:29:11.658528 kubelet[2643]: I0712 09:29:11.658208 2643 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Jul 12 09:29:11.701031 kubelet[2643]: I0712 09:29:11.700997 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3290bd95f938f7c87adaa2d6d57f009e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3290bd95f938f7c87adaa2d6d57f009e\") " pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:11.701031 kubelet[2643]: I0712 09:29:11.701034 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:11.701174 kubelet[2643]: I0712 09:29:11.701058 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:11.701174 kubelet[2643]: I0712 09:29:11.701090 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:11.701174 kubelet[2643]: I0712 09:29:11.701113 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b35b56493416c25588cb530e37ffc065-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"b35b56493416c25588cb530e37ffc065\") " pod="kube-system/kube-scheduler-localhost" Jul 12 09:29:11.701174 kubelet[2643]: I0712 09:29:11.701128 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3290bd95f938f7c87adaa2d6d57f009e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3290bd95f938f7c87adaa2d6d57f009e\") " pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:11.701174 kubelet[2643]: I0712 09:29:11.701142 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3290bd95f938f7c87adaa2d6d57f009e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3290bd95f938f7c87adaa2d6d57f009e\") " pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:11.701278 kubelet[2643]: I0712 09:29:11.701156 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:11.701278 kubelet[2643]: I0712 09:29:11.701170 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 12 09:29:11.922179 kubelet[2643]: E0712 09:29:11.922069 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:11.922179 kubelet[2643]: E0712 09:29:11.922085 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:11.922364 kubelet[2643]: E0712 09:29:11.922332 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:12.481377 kubelet[2643]: I0712 09:29:12.481280 2643 apiserver.go:52] "Watching apiserver" Jul 12 09:29:12.500253 kubelet[2643]: I0712 09:29:12.500203 2643 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 12 09:29:12.527404 kubelet[2643]: E0712 09:29:12.527316 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:12.528369 kubelet[2643]: E0712 09:29:12.528299 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:12.534570 kubelet[2643]: E0712 09:29:12.534357 2643 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jul 12 09:29:12.534570 kubelet[2643]: E0712 09:29:12.534528 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:12.553739 kubelet[2643]: I0712 09:29:12.553627 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.5536096590000001 podStartE2EDuration="1.553609659s" podCreationTimestamp="2025-07-12 09:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 09:29:12.547457356 +0000 UTC m=+1.119238108" watchObservedRunningTime="2025-07-12 09:29:12.553609659 +0000 UTC m=+1.125390411" Jul 12 09:29:12.563404 kubelet[2643]: I0712 09:29:12.563344 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.563326629 podStartE2EDuration="1.563326629s" podCreationTimestamp="2025-07-12 09:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 09:29:12.555089061 +0000 UTC m=+1.126869773" watchObservedRunningTime="2025-07-12 09:29:12.563326629 +0000 UTC m=+1.135107381" Jul 12 09:29:12.575232 kubelet[2643]: I0712 09:29:12.575114 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.575093287 podStartE2EDuration="1.575093287s" podCreationTimestamp="2025-07-12 09:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 09:29:12.563606695 +0000 UTC m=+1.135387447" watchObservedRunningTime="2025-07-12 09:29:12.575093287 +0000 UTC m=+1.146874079" Jul 12 09:29:13.529208 kubelet[2643]: E0712 09:29:13.529181 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:14.531131 kubelet[2643]: E0712 09:29:14.531102 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:17.387083 kubelet[2643]: I0712 09:29:17.387048 2643 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 12 09:29:17.387575 kubelet[2643]: I0712 09:29:17.387552 2643 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 12 09:29:17.387626 containerd[1514]: time="2025-07-12T09:29:17.387373744Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 12 09:29:18.136635 systemd[1]: Created slice kubepods-besteffort-pod510a6b8b_a8ab_4aaf_8534_bd220ffcbd5e.slice - libcontainer container kubepods-besteffort-pod510a6b8b_a8ab_4aaf_8534_bd220ffcbd5e.slice. Jul 12 09:29:18.149047 kubelet[2643]: I0712 09:29:18.149002 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e-kube-proxy\") pod \"kube-proxy-ckqkw\" (UID: \"510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e\") " pod="kube-system/kube-proxy-ckqkw" Jul 12 09:29:18.149047 kubelet[2643]: I0712 09:29:18.149045 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e-xtables-lock\") pod \"kube-proxy-ckqkw\" (UID: \"510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e\") " pod="kube-system/kube-proxy-ckqkw" Jul 12 09:29:18.149173 kubelet[2643]: I0712 09:29:18.149061 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e-lib-modules\") pod \"kube-proxy-ckqkw\" (UID: \"510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e\") " pod="kube-system/kube-proxy-ckqkw" Jul 12 09:29:18.149173 kubelet[2643]: I0712 09:29:18.149078 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slpbl\" (UniqueName: \"kubernetes.io/projected/510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e-kube-api-access-slpbl\") pod \"kube-proxy-ckqkw\" (UID: \"510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e\") " pod="kube-system/kube-proxy-ckqkw" Jul 12 09:29:18.155831 kubelet[2643]: E0712 09:29:18.155805 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:18.257110 kubelet[2643]: E0712 09:29:18.257071 2643 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Jul 12 09:29:18.257110 kubelet[2643]: E0712 09:29:18.257102 2643 projected.go:194] Error preparing data for projected volume kube-api-access-slpbl for pod kube-system/kube-proxy-ckqkw: configmap "kube-root-ca.crt" not found Jul 12 09:29:18.257249 kubelet[2643]: E0712 09:29:18.257147 2643 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e-kube-api-access-slpbl podName:510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e nodeName:}" failed. No retries permitted until 2025-07-12 09:29:18.75712883 +0000 UTC m=+7.328909582 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-slpbl" (UniqueName: "kubernetes.io/projected/510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e-kube-api-access-slpbl") pod "kube-proxy-ckqkw" (UID: "510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e") : configmap "kube-root-ca.crt" not found Jul 12 09:29:18.538213 kubelet[2643]: E0712 09:29:18.538178 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:18.563147 systemd[1]: Created slice kubepods-besteffort-pod2b86d547_8daa_4e6f_93e9_08f261b6f180.slice - libcontainer container kubepods-besteffort-pod2b86d547_8daa_4e6f_93e9_08f261b6f180.slice. Jul 12 09:29:18.651106 kubelet[2643]: I0712 09:29:18.651052 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/2b86d547-8daa-4e6f-93e9-08f261b6f180-var-lib-calico\") pod \"tigera-operator-5bf8dfcb4-5tzxk\" (UID: \"2b86d547-8daa-4e6f-93e9-08f261b6f180\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-5tzxk" Jul 12 09:29:18.651106 kubelet[2643]: I0712 09:29:18.651097 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgwtl\" (UniqueName: \"kubernetes.io/projected/2b86d547-8daa-4e6f-93e9-08f261b6f180-kube-api-access-cgwtl\") pod \"tigera-operator-5bf8dfcb4-5tzxk\" (UID: \"2b86d547-8daa-4e6f-93e9-08f261b6f180\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-5tzxk" Jul 12 09:29:18.866767 containerd[1514]: time="2025-07-12T09:29:18.866652721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-5tzxk,Uid:2b86d547-8daa-4e6f-93e9-08f261b6f180,Namespace:tigera-operator,Attempt:0,}" Jul 12 09:29:18.881376 containerd[1514]: time="2025-07-12T09:29:18.881297223Z" level=info msg="connecting to shim d6f0305a3bc0d1a3c6dabe650a222729fdb274ba64f2dd3b46979bb16a04cad6" address="unix:///run/containerd/s/d177237e4ddc41308cb6e464ccfbc998bfb19b52c4ea839a7f39b49de2867fc9" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:18.905091 systemd[1]: Started cri-containerd-d6f0305a3bc0d1a3c6dabe650a222729fdb274ba64f2dd3b46979bb16a04cad6.scope - libcontainer container d6f0305a3bc0d1a3c6dabe650a222729fdb274ba64f2dd3b46979bb16a04cad6. Jul 12 09:29:18.942652 containerd[1514]: time="2025-07-12T09:29:18.942611760Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-5tzxk,Uid:2b86d547-8daa-4e6f-93e9-08f261b6f180,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"d6f0305a3bc0d1a3c6dabe650a222729fdb274ba64f2dd3b46979bb16a04cad6\"" Jul 12 09:29:18.944295 containerd[1514]: time="2025-07-12T09:29:18.944251315Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Jul 12 09:29:19.059900 kubelet[2643]: E0712 09:29:19.059787 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:19.060790 containerd[1514]: time="2025-07-12T09:29:19.060539057Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ckqkw,Uid:510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:19.100740 containerd[1514]: time="2025-07-12T09:29:19.100703780Z" level=info msg="connecting to shim d1c01741107804c01e16927043358547a3eceab9ec7a21859b619aad7eb9bfca" address="unix:///run/containerd/s/28c5809c819b8434f7d2cdac9377d8bcafc36e3ec2524a21e2b6e0675b9c0c25" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:19.123134 systemd[1]: Started cri-containerd-d1c01741107804c01e16927043358547a3eceab9ec7a21859b619aad7eb9bfca.scope - libcontainer container d1c01741107804c01e16927043358547a3eceab9ec7a21859b619aad7eb9bfca. Jul 12 09:29:19.142986 containerd[1514]: time="2025-07-12T09:29:19.142950377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ckqkw,Uid:510a6b8b-a8ab-4aaf-8534-bd220ffcbd5e,Namespace:kube-system,Attempt:0,} returns sandbox id \"d1c01741107804c01e16927043358547a3eceab9ec7a21859b619aad7eb9bfca\"" Jul 12 09:29:19.143507 kubelet[2643]: E0712 09:29:19.143486 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:19.146354 containerd[1514]: time="2025-07-12T09:29:19.146319516Z" level=info msg="CreateContainer within sandbox \"d1c01741107804c01e16927043358547a3eceab9ec7a21859b619aad7eb9bfca\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 12 09:29:19.157075 containerd[1514]: time="2025-07-12T09:29:19.157038603Z" level=info msg="Container 54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:19.163847 containerd[1514]: time="2025-07-12T09:29:19.163804965Z" level=info msg="CreateContainer within sandbox \"d1c01741107804c01e16927043358547a3eceab9ec7a21859b619aad7eb9bfca\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c\"" Jul 12 09:29:19.164460 containerd[1514]: time="2025-07-12T09:29:19.164384230Z" level=info msg="StartContainer for \"54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c\"" Jul 12 09:29:19.165821 containerd[1514]: time="2025-07-12T09:29:19.165795829Z" level=info msg="connecting to shim 54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c" address="unix:///run/containerd/s/28c5809c819b8434f7d2cdac9377d8bcafc36e3ec2524a21e2b6e0675b9c0c25" protocol=ttrpc version=3 Jul 12 09:29:19.187075 systemd[1]: Started cri-containerd-54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c.scope - libcontainer container 54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c. Jul 12 09:29:19.225077 containerd[1514]: time="2025-07-12T09:29:19.224786271Z" level=info msg="StartContainer for \"54334aec28b416796df3557d72034eee174a2d112839d8b0736fa0ab8415f78c\" returns successfully" Jul 12 09:29:19.543990 kubelet[2643]: E0712 09:29:19.542221 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:19.553857 kubelet[2643]: I0712 09:29:19.553654 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-ckqkw" podStartSLOduration=1.553636778 podStartE2EDuration="1.553636778s" podCreationTimestamp="2025-07-12 09:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 09:29:19.552691992 +0000 UTC m=+8.124472744" watchObservedRunningTime="2025-07-12 09:29:19.553636778 +0000 UTC m=+8.125417530" Jul 12 09:29:19.664275 kubelet[2643]: E0712 09:29:19.663907 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:20.484144 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2378539351.mount: Deactivated successfully. Jul 12 09:29:20.548145 kubelet[2643]: E0712 09:29:20.548112 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:20.784389 containerd[1514]: time="2025-07-12T09:29:20.784335848Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:20.784886 containerd[1514]: time="2025-07-12T09:29:20.784840062Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=22150610" Jul 12 09:29:20.785656 containerd[1514]: time="2025-07-12T09:29:20.785623745Z" level=info msg="ImageCreate event name:\"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:20.787534 containerd[1514]: time="2025-07-12T09:29:20.787502545Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:20.788355 containerd[1514]: time="2025-07-12T09:29:20.788319553Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"22146605\" in 1.844037195s" Jul 12 09:29:20.788387 containerd[1514]: time="2025-07-12T09:29:20.788354316Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\"" Jul 12 09:29:20.790280 containerd[1514]: time="2025-07-12T09:29:20.790246758Z" level=info msg="CreateContainer within sandbox \"d6f0305a3bc0d1a3c6dabe650a222729fdb274ba64f2dd3b46979bb16a04cad6\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jul 12 09:29:20.796695 containerd[1514]: time="2025-07-12T09:29:20.796138506Z" level=info msg="Container 0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:20.801214 containerd[1514]: time="2025-07-12T09:29:20.801175883Z" level=info msg="CreateContainer within sandbox \"d6f0305a3bc0d1a3c6dabe650a222729fdb274ba64f2dd3b46979bb16a04cad6\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e\"" Jul 12 09:29:20.801936 containerd[1514]: time="2025-07-12T09:29:20.801784108Z" level=info msg="StartContainer for \"0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e\"" Jul 12 09:29:20.802757 containerd[1514]: time="2025-07-12T09:29:20.802723208Z" level=info msg="connecting to shim 0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e" address="unix:///run/containerd/s/d177237e4ddc41308cb6e464ccfbc998bfb19b52c4ea839a7f39b49de2867fc9" protocol=ttrpc version=3 Jul 12 09:29:20.825117 systemd[1]: Started cri-containerd-0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e.scope - libcontainer container 0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e. Jul 12 09:29:20.852826 containerd[1514]: time="2025-07-12T09:29:20.852789985Z" level=info msg="StartContainer for \"0cebb59e6c23f25bf286cfee38b2171980da0f93961f0d28f161cd8c72667e5e\" returns successfully" Jul 12 09:29:21.558702 kubelet[2643]: I0712 09:29:21.558641 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5bf8dfcb4-5tzxk" podStartSLOduration=1.71334129 podStartE2EDuration="3.558623862s" podCreationTimestamp="2025-07-12 09:29:18 +0000 UTC" firstStartedPulling="2025-07-12 09:29:18.943838786 +0000 UTC m=+7.515619538" lastFinishedPulling="2025-07-12 09:29:20.789121358 +0000 UTC m=+9.360902110" observedRunningTime="2025-07-12 09:29:21.558557455 +0000 UTC m=+10.130338167" watchObservedRunningTime="2025-07-12 09:29:21.558623862 +0000 UTC m=+10.130404614" Jul 12 09:29:23.299065 kubelet[2643]: E0712 09:29:23.298969 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:26.026297 sudo[1724]: pam_unix(sudo:session): session closed for user root Jul 12 09:29:26.027856 sshd[1723]: Connection closed by 10.0.0.1 port 41372 Jul 12 09:29:26.028271 sshd-session[1720]: pam_unix(sshd:session): session closed for user core Jul 12 09:29:26.031990 systemd-logind[1493]: Session 7 logged out. Waiting for processes to exit. Jul 12 09:29:26.032285 systemd[1]: sshd@6-10.0.0.27:22-10.0.0.1:41372.service: Deactivated successfully. Jul 12 09:29:26.033959 systemd[1]: session-7.scope: Deactivated successfully. Jul 12 09:29:26.034162 systemd[1]: session-7.scope: Consumed 6.342s CPU time, 220.9M memory peak. Jul 12 09:29:26.036342 systemd-logind[1493]: Removed session 7. Jul 12 09:29:26.535521 update_engine[1498]: I20250712 09:29:26.535449 1498 update_attempter.cc:509] Updating boot flags... Jul 12 09:29:31.129753 systemd[1]: Created slice kubepods-besteffort-pod8347fd98_bd7c_4069_8018_4befb5973d94.slice - libcontainer container kubepods-besteffort-pod8347fd98_bd7c_4069_8018_4befb5973d94.slice. Jul 12 09:29:31.136761 kubelet[2643]: I0712 09:29:31.136711 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qvfh\" (UniqueName: \"kubernetes.io/projected/8347fd98-bd7c-4069-8018-4befb5973d94-kube-api-access-5qvfh\") pod \"calico-typha-5848b8b897-flm78\" (UID: \"8347fd98-bd7c-4069-8018-4befb5973d94\") " pod="calico-system/calico-typha-5848b8b897-flm78" Jul 12 09:29:31.137049 kubelet[2643]: I0712 09:29:31.136798 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8347fd98-bd7c-4069-8018-4befb5973d94-tigera-ca-bundle\") pod \"calico-typha-5848b8b897-flm78\" (UID: \"8347fd98-bd7c-4069-8018-4befb5973d94\") " pod="calico-system/calico-typha-5848b8b897-flm78" Jul 12 09:29:31.137049 kubelet[2643]: I0712 09:29:31.136827 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/8347fd98-bd7c-4069-8018-4befb5973d94-typha-certs\") pod \"calico-typha-5848b8b897-flm78\" (UID: \"8347fd98-bd7c-4069-8018-4befb5973d94\") " pod="calico-system/calico-typha-5848b8b897-flm78" Jul 12 09:29:31.434058 kubelet[2643]: E0712 09:29:31.433950 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:31.442547 containerd[1514]: time="2025-07-12T09:29:31.442407625Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5848b8b897-flm78,Uid:8347fd98-bd7c-4069-8018-4befb5973d94,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:31.479503 systemd[1]: Created slice kubepods-besteffort-podb56494a8_6a26_46e3_9226_41c8c8b92923.slice - libcontainer container kubepods-besteffort-podb56494a8_6a26_46e3_9226_41c8c8b92923.slice. Jul 12 09:29:31.498007 containerd[1514]: time="2025-07-12T09:29:31.497956245Z" level=info msg="connecting to shim 392388d0f2c44eec68d56389011a912e77f14e52eaae3d29c27aeab352846c7c" address="unix:///run/containerd/s/73e646024df515227abd6ab888034809b848c34b0b4883a2ac7677e82e393bd2" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:31.540527 kubelet[2643]: I0712 09:29:31.540295 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-var-lib-calico\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540527 kubelet[2643]: I0712 09:29:31.540346 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-cni-net-dir\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540527 kubelet[2643]: I0712 09:29:31.540363 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-flexvol-driver-host\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540527 kubelet[2643]: I0712 09:29:31.540381 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/b56494a8-6a26-46e3-9226-41c8c8b92923-node-certs\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540527 kubelet[2643]: I0712 09:29:31.540398 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-var-run-calico\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540756 kubelet[2643]: I0712 09:29:31.540414 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-lib-modules\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540756 kubelet[2643]: I0712 09:29:31.540430 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-xtables-lock\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540756 kubelet[2643]: I0712 09:29:31.540447 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-cni-bin-dir\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540756 kubelet[2643]: I0712 09:29:31.540461 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-policysync\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540756 kubelet[2643]: I0712 09:29:31.540475 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b56494a8-6a26-46e3-9226-41c8c8b92923-tigera-ca-bundle\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540851 kubelet[2643]: I0712 09:29:31.540493 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cns66\" (UniqueName: \"kubernetes.io/projected/b56494a8-6a26-46e3-9226-41c8c8b92923-kube-api-access-cns66\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.540851 kubelet[2643]: I0712 09:29:31.540546 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/b56494a8-6a26-46e3-9226-41c8c8b92923-cni-log-dir\") pod \"calico-node-md287\" (UID: \"b56494a8-6a26-46e3-9226-41c8c8b92923\") " pod="calico-system/calico-node-md287" Jul 12 09:29:31.575128 systemd[1]: Started cri-containerd-392388d0f2c44eec68d56389011a912e77f14e52eaae3d29c27aeab352846c7c.scope - libcontainer container 392388d0f2c44eec68d56389011a912e77f14e52eaae3d29c27aeab352846c7c. Jul 12 09:29:31.619646 containerd[1514]: time="2025-07-12T09:29:31.619540241Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5848b8b897-flm78,Uid:8347fd98-bd7c-4069-8018-4befb5973d94,Namespace:calico-system,Attempt:0,} returns sandbox id \"392388d0f2c44eec68d56389011a912e77f14e52eaae3d29c27aeab352846c7c\"" Jul 12 09:29:31.620382 kubelet[2643]: E0712 09:29:31.620360 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:31.626184 containerd[1514]: time="2025-07-12T09:29:31.626144723Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Jul 12 09:29:31.649773 kubelet[2643]: E0712 09:29:31.649736 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.649773 kubelet[2643]: W0712 09:29:31.649764 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.650185 kubelet[2643]: E0712 09:29:31.649806 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.650185 kubelet[2643]: E0712 09:29:31.650100 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.650185 kubelet[2643]: W0712 09:29:31.650111 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.650185 kubelet[2643]: E0712 09:29:31.650122 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.653570 kubelet[2643]: E0712 09:29:31.653537 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.653570 kubelet[2643]: W0712 09:29:31.653559 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.653694 kubelet[2643]: E0712 09:29:31.653580 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.751544 kubelet[2643]: E0712 09:29:31.750491 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sk54t" podUID="5e5331e4-f4dc-49b3-a2b8-19205dd650aa" Jul 12 09:29:31.784745 containerd[1514]: time="2025-07-12T09:29:31.784047250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-md287,Uid:b56494a8-6a26-46e3-9226-41c8c8b92923,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:31.804851 containerd[1514]: time="2025-07-12T09:29:31.804773271Z" level=info msg="connecting to shim a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825" address="unix:///run/containerd/s/bf12311d7eabc5670af9f6b52eb1569f2907325a63a029a2d417993dbd38ab0e" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:31.825387 kubelet[2643]: E0712 09:29:31.825349 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.825543 kubelet[2643]: W0712 09:29:31.825526 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.825610 kubelet[2643]: E0712 09:29:31.825598 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.825867 kubelet[2643]: E0712 09:29:31.825853 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.825978 kubelet[2643]: W0712 09:29:31.825964 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.826064 kubelet[2643]: E0712 09:29:31.826052 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.826305 kubelet[2643]: E0712 09:29:31.826292 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.826376 kubelet[2643]: W0712 09:29:31.826365 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.826427 kubelet[2643]: E0712 09:29:31.826417 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.827037 kubelet[2643]: E0712 09:29:31.826817 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.827147 kubelet[2643]: W0712 09:29:31.827127 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.827400 kubelet[2643]: E0712 09:29:31.827220 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.827778 kubelet[2643]: E0712 09:29:31.827758 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.827861 kubelet[2643]: W0712 09:29:31.827847 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.827949 kubelet[2643]: E0712 09:29:31.827936 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.828163 systemd[1]: Started cri-containerd-a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825.scope - libcontainer container a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825. Jul 12 09:29:31.828802 kubelet[2643]: E0712 09:29:31.828663 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.828802 kubelet[2643]: W0712 09:29:31.828678 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.828802 kubelet[2643]: E0712 09:29:31.828694 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.829527 kubelet[2643]: E0712 09:29:31.829511 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.829587 kubelet[2643]: W0712 09:29:31.829576 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.829989 kubelet[2643]: E0712 09:29:31.829971 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.830350 kubelet[2643]: E0712 09:29:31.830286 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.830350 kubelet[2643]: W0712 09:29:31.830299 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.830580 kubelet[2643]: E0712 09:29:31.830448 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.831112 kubelet[2643]: E0712 09:29:31.830849 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.831112 kubelet[2643]: W0712 09:29:31.830864 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.831112 kubelet[2643]: E0712 09:29:31.830875 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.831290 kubelet[2643]: E0712 09:29:31.831274 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.831346 kubelet[2643]: W0712 09:29:31.831335 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.831394 kubelet[2643]: E0712 09:29:31.831385 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.831593 kubelet[2643]: E0712 09:29:31.831580 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.831688 kubelet[2643]: W0712 09:29:31.831675 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.831738 kubelet[2643]: E0712 09:29:31.831727 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.832082 kubelet[2643]: E0712 09:29:31.831968 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.832082 kubelet[2643]: W0712 09:29:31.831982 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.832082 kubelet[2643]: E0712 09:29:31.831993 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.832241 kubelet[2643]: E0712 09:29:31.832229 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.832314 kubelet[2643]: W0712 09:29:31.832302 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.832368 kubelet[2643]: E0712 09:29:31.832358 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.832829 kubelet[2643]: E0712 09:29:31.832813 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.832907 kubelet[2643]: W0712 09:29:31.832894 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.833103 kubelet[2643]: E0712 09:29:31.832980 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.833241 kubelet[2643]: E0712 09:29:31.833227 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.833307 kubelet[2643]: W0712 09:29:31.833295 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.833543 kubelet[2643]: E0712 09:29:31.833517 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.833808 kubelet[2643]: E0712 09:29:31.833794 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.834119 kubelet[2643]: W0712 09:29:31.833863 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.834119 kubelet[2643]: E0712 09:29:31.833880 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.835627 kubelet[2643]: E0712 09:29:31.835601 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.835627 kubelet[2643]: W0712 09:29:31.835617 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.835627 kubelet[2643]: E0712 09:29:31.835630 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.836014 kubelet[2643]: E0712 09:29:31.835998 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.836014 kubelet[2643]: W0712 09:29:31.836013 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.836128 kubelet[2643]: E0712 09:29:31.836077 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.836538 kubelet[2643]: E0712 09:29:31.836521 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.836538 kubelet[2643]: W0712 09:29:31.836538 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.836621 kubelet[2643]: E0712 09:29:31.836551 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.836729 kubelet[2643]: E0712 09:29:31.836709 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.836729 kubelet[2643]: W0712 09:29:31.836722 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.836779 kubelet[2643]: E0712 09:29:31.836731 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.842288 kubelet[2643]: E0712 09:29:31.842259 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.842288 kubelet[2643]: W0712 09:29:31.842283 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.842405 kubelet[2643]: E0712 09:29:31.842301 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.842405 kubelet[2643]: I0712 09:29:31.842357 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/5e5331e4-f4dc-49b3-a2b8-19205dd650aa-varrun\") pod \"csi-node-driver-sk54t\" (UID: \"5e5331e4-f4dc-49b3-a2b8-19205dd650aa\") " pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:31.842581 kubelet[2643]: E0712 09:29:31.842553 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.842581 kubelet[2643]: W0712 09:29:31.842580 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.842651 kubelet[2643]: E0712 09:29:31.842595 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.842651 kubelet[2643]: I0712 09:29:31.842611 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e5331e4-f4dc-49b3-a2b8-19205dd650aa-kubelet-dir\") pod \"csi-node-driver-sk54t\" (UID: \"5e5331e4-f4dc-49b3-a2b8-19205dd650aa\") " pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:31.843657 kubelet[2643]: E0712 09:29:31.843625 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.843657 kubelet[2643]: W0712 09:29:31.843645 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.843979 kubelet[2643]: E0712 09:29:31.843960 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.843979 kubelet[2643]: W0712 09:29:31.843978 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.844101 kubelet[2643]: E0712 09:29:31.844085 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.844130 kubelet[2643]: I0712 09:29:31.844114 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e5331e4-f4dc-49b3-a2b8-19205dd650aa-registration-dir\") pod \"csi-node-driver-sk54t\" (UID: \"5e5331e4-f4dc-49b3-a2b8-19205dd650aa\") " pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:31.844430 kubelet[2643]: E0712 09:29:31.844399 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.844430 kubelet[2643]: W0712 09:29:31.844421 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.844489 kubelet[2643]: E0712 09:29:31.844442 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.844489 kubelet[2643]: E0712 09:29:31.844462 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.844800 kubelet[2643]: E0712 09:29:31.844784 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.844800 kubelet[2643]: W0712 09:29:31.844800 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.844857 kubelet[2643]: E0712 09:29:31.844813 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.845147 kubelet[2643]: E0712 09:29:31.845128 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.845188 kubelet[2643]: W0712 09:29:31.845146 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.845211 kubelet[2643]: E0712 09:29:31.845192 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.845232 kubelet[2643]: I0712 09:29:31.845214 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrttl\" (UniqueName: \"kubernetes.io/projected/5e5331e4-f4dc-49b3-a2b8-19205dd650aa-kube-api-access-rrttl\") pod \"csi-node-driver-sk54t\" (UID: \"5e5331e4-f4dc-49b3-a2b8-19205dd650aa\") " pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:31.845581 kubelet[2643]: E0712 09:29:31.845561 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.845581 kubelet[2643]: W0712 09:29:31.845578 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.845716 kubelet[2643]: E0712 09:29:31.845701 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.845925 kubelet[2643]: E0712 09:29:31.845899 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.845957 kubelet[2643]: W0712 09:29:31.845948 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.845977 kubelet[2643]: E0712 09:29:31.845961 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.846165 kubelet[2643]: E0712 09:29:31.846151 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.846187 kubelet[2643]: W0712 09:29:31.846164 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.846187 kubelet[2643]: E0712 09:29:31.846179 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.846322 kubelet[2643]: E0712 09:29:31.846311 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.846344 kubelet[2643]: W0712 09:29:31.846322 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.846344 kubelet[2643]: E0712 09:29:31.846331 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.846504 kubelet[2643]: E0712 09:29:31.846493 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.846529 kubelet[2643]: W0712 09:29:31.846505 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.846529 kubelet[2643]: E0712 09:29:31.846516 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.846571 kubelet[2643]: I0712 09:29:31.846537 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e5331e4-f4dc-49b3-a2b8-19205dd650aa-socket-dir\") pod \"csi-node-driver-sk54t\" (UID: \"5e5331e4-f4dc-49b3-a2b8-19205dd650aa\") " pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:31.846682 kubelet[2643]: E0712 09:29:31.846668 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.846704 kubelet[2643]: W0712 09:29:31.846682 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.846704 kubelet[2643]: E0712 09:29:31.846692 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.847242 kubelet[2643]: E0712 09:29:31.847224 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.847265 kubelet[2643]: W0712 09:29:31.847243 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.847265 kubelet[2643]: E0712 09:29:31.847258 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.847414 kubelet[2643]: E0712 09:29:31.847404 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.847436 kubelet[2643]: W0712 09:29:31.847415 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.847436 kubelet[2643]: E0712 09:29:31.847425 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.866177 containerd[1514]: time="2025-07-12T09:29:31.866110602Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-md287,Uid:b56494a8-6a26-46e3-9226-41c8c8b92923,Namespace:calico-system,Attempt:0,} returns sandbox id \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\"" Jul 12 09:29:31.948182 kubelet[2643]: E0712 09:29:31.948115 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.948182 kubelet[2643]: W0712 09:29:31.948174 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.948351 kubelet[2643]: E0712 09:29:31.948201 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.948974 kubelet[2643]: E0712 09:29:31.948954 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.948974 kubelet[2643]: W0712 09:29:31.948974 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.949875 kubelet[2643]: E0712 09:29:31.949628 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.950517 kubelet[2643]: E0712 09:29:31.950499 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.950661 kubelet[2643]: W0712 09:29:31.950642 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.950691 kubelet[2643]: E0712 09:29:31.950679 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.951118 kubelet[2643]: E0712 09:29:31.951102 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.951118 kubelet[2643]: W0712 09:29:31.951117 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.951240 kubelet[2643]: E0712 09:29:31.951225 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.951340 kubelet[2643]: E0712 09:29:31.951328 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.951365 kubelet[2643]: W0712 09:29:31.951340 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.951388 kubelet[2643]: E0712 09:29:31.951381 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.951562 kubelet[2643]: E0712 09:29:31.951550 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.951588 kubelet[2643]: W0712 09:29:31.951564 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.951632 kubelet[2643]: E0712 09:29:31.951608 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.951761 kubelet[2643]: E0712 09:29:31.951749 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.951782 kubelet[2643]: W0712 09:29:31.951762 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.951861 kubelet[2643]: E0712 09:29:31.951849 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.951997 kubelet[2643]: E0712 09:29:31.951984 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.952022 kubelet[2643]: W0712 09:29:31.951997 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.952042 kubelet[2643]: E0712 09:29:31.952011 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.952245 kubelet[2643]: E0712 09:29:31.952229 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.952277 kubelet[2643]: W0712 09:29:31.952258 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.952299 kubelet[2643]: E0712 09:29:31.952275 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.952483 kubelet[2643]: E0712 09:29:31.952465 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.952508 kubelet[2643]: W0712 09:29:31.952483 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.952508 kubelet[2643]: E0712 09:29:31.952502 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.952657 kubelet[2643]: E0712 09:29:31.952645 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.952684 kubelet[2643]: W0712 09:29:31.952657 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.952684 kubelet[2643]: E0712 09:29:31.952671 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.953937 kubelet[2643]: E0712 09:29:31.953879 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.953937 kubelet[2643]: W0712 09:29:31.953900 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.953985 kubelet[2643]: E0712 09:29:31.953963 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.954212 kubelet[2643]: E0712 09:29:31.954186 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.954212 kubelet[2643]: W0712 09:29:31.954200 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.954255 kubelet[2643]: E0712 09:29:31.954238 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.954417 kubelet[2643]: E0712 09:29:31.954394 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.954417 kubelet[2643]: W0712 09:29:31.954407 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.954460 kubelet[2643]: E0712 09:29:31.954432 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.954549 kubelet[2643]: E0712 09:29:31.954530 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.954549 kubelet[2643]: W0712 09:29:31.954543 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.954681 kubelet[2643]: E0712 09:29:31.954661 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.954681 kubelet[2643]: W0712 09:29:31.954673 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.954854 kubelet[2643]: E0712 09:29:31.954793 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.954854 kubelet[2643]: E0712 09:29:31.954840 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.954902 kubelet[2643]: E0712 09:29:31.954831 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.954902 kubelet[2643]: W0712 09:29:31.954867 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.954902 kubelet[2643]: E0712 09:29:31.954880 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.955566 kubelet[2643]: E0712 09:29:31.955533 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.955566 kubelet[2643]: W0712 09:29:31.955553 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.955622 kubelet[2643]: E0712 09:29:31.955577 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.956110 kubelet[2643]: E0712 09:29:31.956090 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.956110 kubelet[2643]: W0712 09:29:31.956108 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.956172 kubelet[2643]: E0712 09:29:31.956130 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.957043 kubelet[2643]: E0712 09:29:31.957021 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.957043 kubelet[2643]: W0712 09:29:31.957042 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.957115 kubelet[2643]: E0712 09:29:31.957058 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.957290 kubelet[2643]: E0712 09:29:31.957274 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.957290 kubelet[2643]: W0712 09:29:31.957289 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.957366 kubelet[2643]: E0712 09:29:31.957325 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.957555 kubelet[2643]: E0712 09:29:31.957537 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.957585 kubelet[2643]: W0712 09:29:31.957557 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.957585 kubelet[2643]: E0712 09:29:31.957576 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.958148 kubelet[2643]: E0712 09:29:31.958125 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.958177 kubelet[2643]: W0712 09:29:31.958169 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.958210 kubelet[2643]: E0712 09:29:31.958191 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.958597 kubelet[2643]: E0712 09:29:31.958580 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.958632 kubelet[2643]: W0712 09:29:31.958597 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.958908 kubelet[2643]: E0712 09:29:31.958890 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.958908 kubelet[2643]: W0712 09:29:31.958908 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.958990 kubelet[2643]: E0712 09:29:31.958951 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.959106 kubelet[2643]: E0712 09:29:31.958616 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:31.969853 kubelet[2643]: E0712 09:29:31.969815 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:31.969853 kubelet[2643]: W0712 09:29:31.969840 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:31.969853 kubelet[2643]: E0712 09:29:31.969860 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:32.630108 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3773122703.mount: Deactivated successfully. Jul 12 09:29:33.339079 containerd[1514]: time="2025-07-12T09:29:33.339023673Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:33.340051 containerd[1514]: time="2025-07-12T09:29:33.340007288Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=33087207" Jul 12 09:29:33.340974 containerd[1514]: time="2025-07-12T09:29:33.340921019Z" level=info msg="ImageCreate event name:\"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:33.342908 containerd[1514]: time="2025-07-12T09:29:33.342853046Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:33.343385 containerd[1514]: time="2025-07-12T09:29:33.343359634Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"33087061\" in 1.717173788s" Jul 12 09:29:33.343429 containerd[1514]: time="2025-07-12T09:29:33.343386635Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\"" Jul 12 09:29:33.346291 containerd[1514]: time="2025-07-12T09:29:33.346256915Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Jul 12 09:29:33.369551 containerd[1514]: time="2025-07-12T09:29:33.369513725Z" level=info msg="CreateContainer within sandbox \"392388d0f2c44eec68d56389011a912e77f14e52eaae3d29c27aeab352846c7c\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jul 12 09:29:33.377877 containerd[1514]: time="2025-07-12T09:29:33.377829546Z" level=info msg="Container 6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:33.385648 containerd[1514]: time="2025-07-12T09:29:33.385591577Z" level=info msg="CreateContainer within sandbox \"392388d0f2c44eec68d56389011a912e77f14e52eaae3d29c27aeab352846c7c\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb\"" Jul 12 09:29:33.386230 containerd[1514]: time="2025-07-12T09:29:33.386134287Z" level=info msg="StartContainer for \"6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb\"" Jul 12 09:29:33.387467 containerd[1514]: time="2025-07-12T09:29:33.387437839Z" level=info msg="connecting to shim 6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb" address="unix:///run/containerd/s/73e646024df515227abd6ab888034809b848c34b0b4883a2ac7677e82e393bd2" protocol=ttrpc version=3 Jul 12 09:29:33.408120 systemd[1]: Started cri-containerd-6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb.scope - libcontainer container 6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb. Jul 12 09:29:33.447828 containerd[1514]: time="2025-07-12T09:29:33.447791468Z" level=info msg="StartContainer for \"6e93cd7b26d208e068cdc1e0244a3697241f2e3370313825c3e55dd196d9dfdb\" returns successfully" Jul 12 09:29:33.514167 kubelet[2643]: E0712 09:29:33.514107 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sk54t" podUID="5e5331e4-f4dc-49b3-a2b8-19205dd650aa" Jul 12 09:29:33.590288 kubelet[2643]: E0712 09:29:33.590179 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:33.647834 kubelet[2643]: E0712 09:29:33.647798 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.647834 kubelet[2643]: W0712 09:29:33.647822 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.647834 kubelet[2643]: E0712 09:29:33.647842 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.648051 kubelet[2643]: E0712 09:29:33.648026 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.648051 kubelet[2643]: W0712 09:29:33.648035 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.648051 kubelet[2643]: E0712 09:29:33.648045 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.648241 kubelet[2643]: E0712 09:29:33.648212 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.648241 kubelet[2643]: W0712 09:29:33.648230 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.648241 kubelet[2643]: E0712 09:29:33.648240 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.650100 kubelet[2643]: E0712 09:29:33.650031 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.650100 kubelet[2643]: W0712 09:29:33.650050 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.650100 kubelet[2643]: E0712 09:29:33.650075 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.650388 kubelet[2643]: E0712 09:29:33.650371 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.650388 kubelet[2643]: W0712 09:29:33.650385 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.650461 kubelet[2643]: E0712 09:29:33.650398 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.651179 kubelet[2643]: E0712 09:29:33.651158 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.651179 kubelet[2643]: W0712 09:29:33.651180 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.651341 kubelet[2643]: E0712 09:29:33.651195 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.653048 kubelet[2643]: E0712 09:29:33.653007 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.653048 kubelet[2643]: W0712 09:29:33.653042 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.653240 kubelet[2643]: E0712 09:29:33.653057 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.653348 kubelet[2643]: E0712 09:29:33.653332 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.653348 kubelet[2643]: W0712 09:29:33.653346 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.653415 kubelet[2643]: E0712 09:29:33.653357 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.653550 kubelet[2643]: E0712 09:29:33.653536 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.653550 kubelet[2643]: W0712 09:29:33.653548 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.653620 kubelet[2643]: E0712 09:29:33.653560 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.656130 kubelet[2643]: E0712 09:29:33.656103 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.656130 kubelet[2643]: W0712 09:29:33.656123 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.656252 kubelet[2643]: E0712 09:29:33.656140 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.657075 kubelet[2643]: E0712 09:29:33.657054 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.657075 kubelet[2643]: W0712 09:29:33.657071 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.657212 kubelet[2643]: E0712 09:29:33.657084 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.657444 kubelet[2643]: E0712 09:29:33.657430 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.657444 kubelet[2643]: W0712 09:29:33.657442 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.657527 kubelet[2643]: E0712 09:29:33.657452 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.657611 kubelet[2643]: E0712 09:29:33.657599 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.657611 kubelet[2643]: W0712 09:29:33.657609 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.657669 kubelet[2643]: E0712 09:29:33.657617 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.657771 kubelet[2643]: E0712 09:29:33.657759 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.657771 kubelet[2643]: W0712 09:29:33.657769 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.657828 kubelet[2643]: E0712 09:29:33.657777 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.658108 kubelet[2643]: E0712 09:29:33.658082 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.658108 kubelet[2643]: W0712 09:29:33.658104 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.658192 kubelet[2643]: E0712 09:29:33.658115 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.665562 kubelet[2643]: E0712 09:29:33.665419 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.665562 kubelet[2643]: W0712 09:29:33.665443 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.665562 kubelet[2643]: E0712 09:29:33.665461 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.665791 kubelet[2643]: E0712 09:29:33.665779 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.665854 kubelet[2643]: W0712 09:29:33.665843 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.665927 kubelet[2643]: E0712 09:29:33.665904 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.666139 kubelet[2643]: E0712 09:29:33.666120 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.666182 kubelet[2643]: W0712 09:29:33.666139 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.666182 kubelet[2643]: E0712 09:29:33.666159 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.666314 kubelet[2643]: E0712 09:29:33.666302 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.666347 kubelet[2643]: W0712 09:29:33.666315 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.666347 kubelet[2643]: E0712 09:29:33.666331 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.666485 kubelet[2643]: E0712 09:29:33.666474 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.666485 kubelet[2643]: W0712 09:29:33.666484 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.666532 kubelet[2643]: E0712 09:29:33.666498 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.666661 kubelet[2643]: E0712 09:29:33.666650 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.666693 kubelet[2643]: W0712 09:29:33.666662 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.666693 kubelet[2643]: E0712 09:29:33.666676 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.666926 kubelet[2643]: E0712 09:29:33.666904 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.666966 kubelet[2643]: W0712 09:29:33.666926 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.666966 kubelet[2643]: E0712 09:29:33.666944 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.667162 kubelet[2643]: E0712 09:29:33.667149 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.667162 kubelet[2643]: W0712 09:29:33.667162 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.667275 kubelet[2643]: E0712 09:29:33.667239 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.667347 kubelet[2643]: E0712 09:29:33.667332 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.667347 kubelet[2643]: W0712 09:29:33.667344 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.667440 kubelet[2643]: E0712 09:29:33.667424 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.667525 kubelet[2643]: E0712 09:29:33.667514 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.667525 kubelet[2643]: W0712 09:29:33.667524 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.667575 kubelet[2643]: E0712 09:29:33.667540 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.667700 kubelet[2643]: E0712 09:29:33.667689 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.667700 kubelet[2643]: W0712 09:29:33.667699 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.667755 kubelet[2643]: E0712 09:29:33.667715 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.667856 kubelet[2643]: E0712 09:29:33.667845 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.667890 kubelet[2643]: W0712 09:29:33.667856 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.667890 kubelet[2643]: E0712 09:29:33.667867 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.668089 kubelet[2643]: E0712 09:29:33.668072 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.668123 kubelet[2643]: W0712 09:29:33.668090 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.668123 kubelet[2643]: E0712 09:29:33.668109 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.668254 kubelet[2643]: E0712 09:29:33.668242 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.668328 kubelet[2643]: W0712 09:29:33.668254 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.668328 kubelet[2643]: E0712 09:29:33.668267 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.668480 kubelet[2643]: E0712 09:29:33.668467 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.668508 kubelet[2643]: W0712 09:29:33.668480 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.668508 kubelet[2643]: E0712 09:29:33.668496 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.669000 kubelet[2643]: E0712 09:29:33.668980 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.669000 kubelet[2643]: W0712 09:29:33.668997 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.669102 kubelet[2643]: E0712 09:29:33.669023 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.669306 kubelet[2643]: E0712 09:29:33.669289 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.669339 kubelet[2643]: W0712 09:29:33.669306 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.669339 kubelet[2643]: E0712 09:29:33.669318 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:33.670216 kubelet[2643]: E0712 09:29:33.670198 2643 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 12 09:29:33.670268 kubelet[2643]: W0712 09:29:33.670216 2643 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 12 09:29:33.670268 kubelet[2643]: E0712 09:29:33.670230 2643 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 12 09:29:34.457151 containerd[1514]: time="2025-07-12T09:29:34.457073512Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:34.457860 containerd[1514]: time="2025-07-12T09:29:34.457794670Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4266981" Jul 12 09:29:34.458540 containerd[1514]: time="2025-07-12T09:29:34.458507708Z" level=info msg="ImageCreate event name:\"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:34.460810 containerd[1514]: time="2025-07-12T09:29:34.460778308Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:34.461498 containerd[1514]: time="2025-07-12T09:29:34.461315057Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5636182\" in 1.114988178s" Jul 12 09:29:34.461498 containerd[1514]: time="2025-07-12T09:29:34.461344538Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\"" Jul 12 09:29:34.463701 containerd[1514]: time="2025-07-12T09:29:34.463670582Z" level=info msg="CreateContainer within sandbox \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jul 12 09:29:34.511396 containerd[1514]: time="2025-07-12T09:29:34.510442823Z" level=info msg="Container c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:34.518971 containerd[1514]: time="2025-07-12T09:29:34.518898752Z" level=info msg="CreateContainer within sandbox \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\"" Jul 12 09:29:34.519487 containerd[1514]: time="2025-07-12T09:29:34.519440900Z" level=info msg="StartContainer for \"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\"" Jul 12 09:29:34.520810 containerd[1514]: time="2025-07-12T09:29:34.520783212Z" level=info msg="connecting to shim c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5" address="unix:///run/containerd/s/bf12311d7eabc5670af9f6b52eb1569f2907325a63a029a2d417993dbd38ab0e" protocol=ttrpc version=3 Jul 12 09:29:34.555157 systemd[1]: Started cri-containerd-c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5.scope - libcontainer container c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5. Jul 12 09:29:34.597351 kubelet[2643]: I0712 09:29:34.597230 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:34.597661 containerd[1514]: time="2025-07-12T09:29:34.597285750Z" level=info msg="StartContainer for \"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\" returns successfully" Jul 12 09:29:34.598493 kubelet[2643]: E0712 09:29:34.598428 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:34.635899 systemd[1]: cri-containerd-c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5.scope: Deactivated successfully. Jul 12 09:29:34.661844 containerd[1514]: time="2025-07-12T09:29:34.661669725Z" level=info msg="received exit event container_id:\"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\" id:\"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\" pid:3341 exited_at:{seconds:1752312574 nanos:655210102}" Jul 12 09:29:34.661844 containerd[1514]: time="2025-07-12T09:29:34.661776611Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\" id:\"c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5\" pid:3341 exited_at:{seconds:1752312574 nanos:655210102}" Jul 12 09:29:34.701136 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c0de3cfd94b6bfb690b5b94b17896cef03c6d4f0231a4bc45cd2448531fb79e5-rootfs.mount: Deactivated successfully. Jul 12 09:29:35.514284 kubelet[2643]: E0712 09:29:35.514224 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sk54t" podUID="5e5331e4-f4dc-49b3-a2b8-19205dd650aa" Jul 12 09:29:35.601940 containerd[1514]: time="2025-07-12T09:29:35.601745986Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Jul 12 09:29:35.619711 kubelet[2643]: I0712 09:29:35.619627 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5848b8b897-flm78" podStartSLOduration=2.899257887 podStartE2EDuration="4.619608013s" podCreationTimestamp="2025-07-12 09:29:31 +0000 UTC" firstStartedPulling="2025-07-12 09:29:31.625739459 +0000 UTC m=+20.197520211" lastFinishedPulling="2025-07-12 09:29:33.346089585 +0000 UTC m=+21.917870337" observedRunningTime="2025-07-12 09:29:33.613265328 +0000 UTC m=+22.185046080" watchObservedRunningTime="2025-07-12 09:29:35.619608013 +0000 UTC m=+24.191388805" Jul 12 09:29:37.514937 kubelet[2643]: E0712 09:29:37.514044 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sk54t" podUID="5e5331e4-f4dc-49b3-a2b8-19205dd650aa" Jul 12 09:29:38.835689 containerd[1514]: time="2025-07-12T09:29:38.835645183Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:38.838649 containerd[1514]: time="2025-07-12T09:29:38.838600755Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=65888320" Jul 12 09:29:38.839441 containerd[1514]: time="2025-07-12T09:29:38.839408191Z" level=info msg="ImageCreate event name:\"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:38.841574 containerd[1514]: time="2025-07-12T09:29:38.841537406Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:38.842079 containerd[1514]: time="2025-07-12T09:29:38.842046509Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"67257561\" in 3.240265162s" Jul 12 09:29:38.842112 containerd[1514]: time="2025-07-12T09:29:38.842080831Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\"" Jul 12 09:29:38.845299 containerd[1514]: time="2025-07-12T09:29:38.845266733Z" level=info msg="CreateContainer within sandbox \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jul 12 09:29:38.854963 containerd[1514]: time="2025-07-12T09:29:38.854617192Z" level=info msg="Container 698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:38.862917 containerd[1514]: time="2025-07-12T09:29:38.862871441Z" level=info msg="CreateContainer within sandbox \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\"" Jul 12 09:29:38.863383 containerd[1514]: time="2025-07-12T09:29:38.863360463Z" level=info msg="StartContainer for \"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\"" Jul 12 09:29:38.866056 containerd[1514]: time="2025-07-12T09:29:38.865998021Z" level=info msg="connecting to shim 698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2" address="unix:///run/containerd/s/bf12311d7eabc5670af9f6b52eb1569f2907325a63a029a2d417993dbd38ab0e" protocol=ttrpc version=3 Jul 12 09:29:38.899132 systemd[1]: Started cri-containerd-698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2.scope - libcontainer container 698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2. Jul 12 09:29:38.945062 containerd[1514]: time="2025-07-12T09:29:38.945007237Z" level=info msg="StartContainer for \"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\" returns successfully" Jul 12 09:29:39.478759 systemd[1]: cri-containerd-698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2.scope: Deactivated successfully. Jul 12 09:29:39.479057 systemd[1]: cri-containerd-698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2.scope: Consumed 451ms CPU time, 176.2M memory peak, 2.7M read from disk, 165.8M written to disk. Jul 12 09:29:39.481604 containerd[1514]: time="2025-07-12T09:29:39.481552360Z" level=info msg="TaskExit event in podsandbox handler container_id:\"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\" id:\"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\" pid:3402 exited_at:{seconds:1752312579 nanos:480886012}" Jul 12 09:29:39.481763 containerd[1514]: time="2025-07-12T09:29:39.481716527Z" level=info msg="received exit event container_id:\"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\" id:\"698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2\" pid:3402 exited_at:{seconds:1752312579 nanos:480886012}" Jul 12 09:29:39.500836 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-698b07148b6b7eb84fefa0242704a18e91187d7a8e0149f2980758cade762ba2-rootfs.mount: Deactivated successfully. Jul 12 09:29:39.515322 kubelet[2643]: E0712 09:29:39.515263 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sk54t" podUID="5e5331e4-f4dc-49b3-a2b8-19205dd650aa" Jul 12 09:29:39.576706 kubelet[2643]: I0712 09:29:39.576658 2643 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Jul 12 09:29:39.647031 systemd[1]: Created slice kubepods-besteffort-pod58e64786_df0a_4048_91ff_e601fc952e42.slice - libcontainer container kubepods-besteffort-pod58e64786_df0a_4048_91ff_e601fc952e42.slice. Jul 12 09:29:39.654280 systemd[1]: Created slice kubepods-burstable-podf2948faa_6bb4_4b0e_bf03_6e794e9ce052.slice - libcontainer container kubepods-burstable-podf2948faa_6bb4_4b0e_bf03_6e794e9ce052.slice. Jul 12 09:29:39.663990 systemd[1]: Created slice kubepods-burstable-podc3a983bd_242c_4076_abf5_8ee86b62b108.slice - libcontainer container kubepods-burstable-podc3a983bd_242c_4076_abf5_8ee86b62b108.slice. Jul 12 09:29:39.677072 systemd[1]: Created slice kubepods-besteffort-pod4aa0af8e_b7e4_46e7_a2a1_12cadfbfe817.slice - libcontainer container kubepods-besteffort-pod4aa0af8e_b7e4_46e7_a2a1_12cadfbfe817.slice. Jul 12 09:29:39.693895 systemd[1]: Created slice kubepods-besteffort-poda67ae9fe_f378_4098_b84d_38bd8456798a.slice - libcontainer container kubepods-besteffort-poda67ae9fe_f378_4098_b84d_38bd8456798a.slice. Jul 12 09:29:39.699584 systemd[1]: Created slice kubepods-besteffort-pod056e934c_fa80_4772_a8bc_0ff57d009fc2.slice - libcontainer container kubepods-besteffort-pod056e934c_fa80_4772_a8bc_0ff57d009fc2.slice. Jul 12 09:29:39.705360 systemd[1]: Created slice kubepods-besteffort-podf0b9a0d1_4bba_408e_a8de_97ec9d5156e7.slice - libcontainer container kubepods-besteffort-podf0b9a0d1_4bba_408e_a8de_97ec9d5156e7.slice. Jul 12 09:29:39.808495 kubelet[2643]: I0712 09:29:39.808334 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817-tigera-ca-bundle\") pod \"calico-kube-controllers-557fb495b6-kkxx2\" (UID: \"4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817\") " pod="calico-system/calico-kube-controllers-557fb495b6-kkxx2" Jul 12 09:29:39.808495 kubelet[2643]: I0712 09:29:39.808387 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/056e934c-fa80-4772-a8bc-0ff57d009fc2-calico-apiserver-certs\") pod \"calico-apiserver-5c7b85d947-fjdcg\" (UID: \"056e934c-fa80-4772-a8bc-0ff57d009fc2\") " pod="calico-apiserver/calico-apiserver-5c7b85d947-fjdcg" Jul 12 09:29:39.808495 kubelet[2643]: I0712 09:29:39.808409 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72x56\" (UniqueName: \"kubernetes.io/projected/c3a983bd-242c-4076-abf5-8ee86b62b108-kube-api-access-72x56\") pod \"coredns-7c65d6cfc9-4d9r4\" (UID: \"c3a983bd-242c-4076-abf5-8ee86b62b108\") " pod="kube-system/coredns-7c65d6cfc9-4d9r4" Jul 12 09:29:39.808495 kubelet[2643]: I0712 09:29:39.808426 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pv4h\" (UniqueName: \"kubernetes.io/projected/58e64786-df0a-4048-91ff-e601fc952e42-kube-api-access-9pv4h\") pod \"calico-apiserver-5c7b85d947-pcqlb\" (UID: \"58e64786-df0a-4048-91ff-e601fc952e42\") " pod="calico-apiserver/calico-apiserver-5c7b85d947-pcqlb" Jul 12 09:29:39.808495 kubelet[2643]: I0712 09:29:39.808448 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67ae9fe-f378-4098-b84d-38bd8456798a-goldmane-ca-bundle\") pod \"goldmane-58fd7646b9-hc8ld\" (UID: \"a67ae9fe-f378-4098-b84d-38bd8456798a\") " pod="calico-system/goldmane-58fd7646b9-hc8ld" Jul 12 09:29:39.808721 kubelet[2643]: I0712 09:29:39.808466 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2948faa-6bb4-4b0e-bf03-6e794e9ce052-config-volume\") pod \"coredns-7c65d6cfc9-cf9ps\" (UID: \"f2948faa-6bb4-4b0e-bf03-6e794e9ce052\") " pod="kube-system/coredns-7c65d6cfc9-cf9ps" Jul 12 09:29:39.808903 kubelet[2643]: I0712 09:29:39.808874 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwgsg\" (UniqueName: \"kubernetes.io/projected/056e934c-fa80-4772-a8bc-0ff57d009fc2-kube-api-access-wwgsg\") pod \"calico-apiserver-5c7b85d947-fjdcg\" (UID: \"056e934c-fa80-4772-a8bc-0ff57d009fc2\") " pod="calico-apiserver/calico-apiserver-5c7b85d947-fjdcg" Jul 12 09:29:39.809021 kubelet[2643]: I0712 09:29:39.809003 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/a67ae9fe-f378-4098-b84d-38bd8456798a-goldmane-key-pair\") pod \"goldmane-58fd7646b9-hc8ld\" (UID: \"a67ae9fe-f378-4098-b84d-38bd8456798a\") " pod="calico-system/goldmane-58fd7646b9-hc8ld" Jul 12 09:29:39.809067 kubelet[2643]: I0712 09:29:39.809047 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/58e64786-df0a-4048-91ff-e601fc952e42-calico-apiserver-certs\") pod \"calico-apiserver-5c7b85d947-pcqlb\" (UID: \"58e64786-df0a-4048-91ff-e601fc952e42\") " pod="calico-apiserver/calico-apiserver-5c7b85d947-pcqlb" Jul 12 09:29:39.809096 kubelet[2643]: I0712 09:29:39.809072 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4c2j\" (UniqueName: \"kubernetes.io/projected/f2948faa-6bb4-4b0e-bf03-6e794e9ce052-kube-api-access-l4c2j\") pod \"coredns-7c65d6cfc9-cf9ps\" (UID: \"f2948faa-6bb4-4b0e-bf03-6e794e9ce052\") " pod="kube-system/coredns-7c65d6cfc9-cf9ps" Jul 12 09:29:39.809096 kubelet[2643]: I0712 09:29:39.809091 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-backend-key-pair\") pod \"whisker-77cdb8f6c6-g5cxl\" (UID: \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\") " pod="calico-system/whisker-77cdb8f6c6-g5cxl" Jul 12 09:29:39.809157 kubelet[2643]: I0712 09:29:39.809110 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhvbw\" (UniqueName: \"kubernetes.io/projected/4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817-kube-api-access-zhvbw\") pod \"calico-kube-controllers-557fb495b6-kkxx2\" (UID: \"4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817\") " pod="calico-system/calico-kube-controllers-557fb495b6-kkxx2" Jul 12 09:29:39.809157 kubelet[2643]: I0712 09:29:39.809126 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67ae9fe-f378-4098-b84d-38bd8456798a-config\") pod \"goldmane-58fd7646b9-hc8ld\" (UID: \"a67ae9fe-f378-4098-b84d-38bd8456798a\") " pod="calico-system/goldmane-58fd7646b9-hc8ld" Jul 12 09:29:39.809157 kubelet[2643]: I0712 09:29:39.809144 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xmx2\" (UniqueName: \"kubernetes.io/projected/a67ae9fe-f378-4098-b84d-38bd8456798a-kube-api-access-8xmx2\") pod \"goldmane-58fd7646b9-hc8ld\" (UID: \"a67ae9fe-f378-4098-b84d-38bd8456798a\") " pod="calico-system/goldmane-58fd7646b9-hc8ld" Jul 12 09:29:39.809237 kubelet[2643]: I0712 09:29:39.809162 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-ca-bundle\") pod \"whisker-77cdb8f6c6-g5cxl\" (UID: \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\") " pod="calico-system/whisker-77cdb8f6c6-g5cxl" Jul 12 09:29:39.809237 kubelet[2643]: I0712 09:29:39.809179 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctqmw\" (UniqueName: \"kubernetes.io/projected/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-kube-api-access-ctqmw\") pod \"whisker-77cdb8f6c6-g5cxl\" (UID: \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\") " pod="calico-system/whisker-77cdb8f6c6-g5cxl" Jul 12 09:29:39.809237 kubelet[2643]: I0712 09:29:39.809211 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3a983bd-242c-4076-abf5-8ee86b62b108-config-volume\") pod \"coredns-7c65d6cfc9-4d9r4\" (UID: \"c3a983bd-242c-4076-abf5-8ee86b62b108\") " pod="kube-system/coredns-7c65d6cfc9-4d9r4" Jul 12 09:29:39.952213 containerd[1514]: time="2025-07-12T09:29:39.951935061Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-pcqlb,Uid:58e64786-df0a-4048-91ff-e601fc952e42,Namespace:calico-apiserver,Attempt:0,}" Jul 12 09:29:39.960765 kubelet[2643]: E0712 09:29:39.960720 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:39.964628 containerd[1514]: time="2025-07-12T09:29:39.964001780Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-cf9ps,Uid:f2948faa-6bb4-4b0e-bf03-6e794e9ce052,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:39.976630 kubelet[2643]: E0712 09:29:39.973625 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:39.976727 containerd[1514]: time="2025-07-12T09:29:39.974268501Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-4d9r4,Uid:c3a983bd-242c-4076-abf5-8ee86b62b108,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:39.984463 containerd[1514]: time="2025-07-12T09:29:39.983721788Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-557fb495b6-kkxx2,Uid:4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:40.002531 containerd[1514]: time="2025-07-12T09:29:40.002490312Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-hc8ld,Uid:a67ae9fe-f378-4098-b84d-38bd8456798a,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:40.004544 containerd[1514]: time="2025-07-12T09:29:40.004313267Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-fjdcg,Uid:056e934c-fa80-4772-a8bc-0ff57d009fc2,Namespace:calico-apiserver,Attempt:0,}" Jul 12 09:29:40.008880 containerd[1514]: time="2025-07-12T09:29:40.008839814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77cdb8f6c6-g5cxl,Uid:f0b9a0d1-4bba-408e-a8de-97ec9d5156e7,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:40.397476 containerd[1514]: time="2025-07-12T09:29:40.397406877Z" level=error msg="Failed to destroy network for sandbox \"a64e5d1a82f536fa59584acc0a4dc8f3dd7d21ed2eb80935e04a917b9ca74cc6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.400527 containerd[1514]: time="2025-07-12T09:29:40.400387640Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-hc8ld,Uid:a67ae9fe-f378-4098-b84d-38bd8456798a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a64e5d1a82f536fa59584acc0a4dc8f3dd7d21ed2eb80935e04a917b9ca74cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.401078 containerd[1514]: time="2025-07-12T09:29:40.400985665Z" level=error msg="Failed to destroy network for sandbox \"954a7f85c257c8b560d88412d86a305d36288180f1b5ef6bd2ba465bfe3e7618\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.401953 containerd[1514]: time="2025-07-12T09:29:40.401884062Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-4d9r4,Uid:c3a983bd-242c-4076-abf5-8ee86b62b108,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"954a7f85c257c8b560d88412d86a305d36288180f1b5ef6bd2ba465bfe3e7618\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.402206 kubelet[2643]: E0712 09:29:40.402160 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a64e5d1a82f536fa59584acc0a4dc8f3dd7d21ed2eb80935e04a917b9ca74cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.402668 kubelet[2643]: E0712 09:29:40.402234 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a64e5d1a82f536fa59584acc0a4dc8f3dd7d21ed2eb80935e04a917b9ca74cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-hc8ld" Jul 12 09:29:40.402668 kubelet[2643]: E0712 09:29:40.402277 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a64e5d1a82f536fa59584acc0a4dc8f3dd7d21ed2eb80935e04a917b9ca74cc6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-hc8ld" Jul 12 09:29:40.402668 kubelet[2643]: E0712 09:29:40.402327 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-58fd7646b9-hc8ld_calico-system(a67ae9fe-f378-4098-b84d-38bd8456798a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-58fd7646b9-hc8ld_calico-system(a67ae9fe-f378-4098-b84d-38bd8456798a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a64e5d1a82f536fa59584acc0a4dc8f3dd7d21ed2eb80935e04a917b9ca74cc6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-hc8ld" podUID="a67ae9fe-f378-4098-b84d-38bd8456798a" Jul 12 09:29:40.402791 kubelet[2643]: E0712 09:29:40.402581 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"954a7f85c257c8b560d88412d86a305d36288180f1b5ef6bd2ba465bfe3e7618\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.402791 kubelet[2643]: E0712 09:29:40.402609 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"954a7f85c257c8b560d88412d86a305d36288180f1b5ef6bd2ba465bfe3e7618\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-4d9r4" Jul 12 09:29:40.402791 kubelet[2643]: E0712 09:29:40.402627 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"954a7f85c257c8b560d88412d86a305d36288180f1b5ef6bd2ba465bfe3e7618\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-4d9r4" Jul 12 09:29:40.402854 kubelet[2643]: E0712 09:29:40.402655 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-4d9r4_kube-system(c3a983bd-242c-4076-abf5-8ee86b62b108)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-4d9r4_kube-system(c3a983bd-242c-4076-abf5-8ee86b62b108)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"954a7f85c257c8b560d88412d86a305d36288180f1b5ef6bd2ba465bfe3e7618\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-4d9r4" podUID="c3a983bd-242c-4076-abf5-8ee86b62b108" Jul 12 09:29:40.407207 containerd[1514]: time="2025-07-12T09:29:40.407157720Z" level=error msg="Failed to destroy network for sandbox \"498b1b018807ff3b662b3d282dee853b70520ce48fc390e136ce1f66c42ae3c3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.408690 containerd[1514]: time="2025-07-12T09:29:40.408657782Z" level=error msg="Failed to destroy network for sandbox \"7324f1f0dcd4dee67bd565c997f65fe98bd6688e9a6a1749f589c6b38efe67ed\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.411202 containerd[1514]: time="2025-07-12T09:29:40.411156885Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77cdb8f6c6-g5cxl,Uid:f0b9a0d1-4bba-408e-a8de-97ec9d5156e7,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"498b1b018807ff3b662b3d282dee853b70520ce48fc390e136ce1f66c42ae3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.411424 kubelet[2643]: E0712 09:29:40.411363 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"498b1b018807ff3b662b3d282dee853b70520ce48fc390e136ce1f66c42ae3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.411502 kubelet[2643]: E0712 09:29:40.411444 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"498b1b018807ff3b662b3d282dee853b70520ce48fc390e136ce1f66c42ae3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77cdb8f6c6-g5cxl" Jul 12 09:29:40.411502 kubelet[2643]: E0712 09:29:40.411465 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"498b1b018807ff3b662b3d282dee853b70520ce48fc390e136ce1f66c42ae3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77cdb8f6c6-g5cxl" Jul 12 09:29:40.411553 kubelet[2643]: E0712 09:29:40.411506 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-77cdb8f6c6-g5cxl_calico-system(f0b9a0d1-4bba-408e-a8de-97ec9d5156e7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-77cdb8f6c6-g5cxl_calico-system(f0b9a0d1-4bba-408e-a8de-97ec9d5156e7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"498b1b018807ff3b662b3d282dee853b70520ce48fc390e136ce1f66c42ae3c3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-77cdb8f6c6-g5cxl" podUID="f0b9a0d1-4bba-408e-a8de-97ec9d5156e7" Jul 12 09:29:40.412271 containerd[1514]: time="2025-07-12T09:29:40.412140486Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-fjdcg,Uid:056e934c-fa80-4772-a8bc-0ff57d009fc2,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7324f1f0dcd4dee67bd565c997f65fe98bd6688e9a6a1749f589c6b38efe67ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.412578 kubelet[2643]: E0712 09:29:40.412541 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7324f1f0dcd4dee67bd565c997f65fe98bd6688e9a6a1749f589c6b38efe67ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.412653 kubelet[2643]: E0712 09:29:40.412589 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7324f1f0dcd4dee67bd565c997f65fe98bd6688e9a6a1749f589c6b38efe67ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7b85d947-fjdcg" Jul 12 09:29:40.412653 kubelet[2643]: E0712 09:29:40.412606 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7324f1f0dcd4dee67bd565c997f65fe98bd6688e9a6a1749f589c6b38efe67ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7b85d947-fjdcg" Jul 12 09:29:40.412653 kubelet[2643]: E0712 09:29:40.412635 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5c7b85d947-fjdcg_calico-apiserver(056e934c-fa80-4772-a8bc-0ff57d009fc2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5c7b85d947-fjdcg_calico-apiserver(056e934c-fa80-4772-a8bc-0ff57d009fc2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7324f1f0dcd4dee67bd565c997f65fe98bd6688e9a6a1749f589c6b38efe67ed\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5c7b85d947-fjdcg" podUID="056e934c-fa80-4772-a8bc-0ff57d009fc2" Jul 12 09:29:40.420824 containerd[1514]: time="2025-07-12T09:29:40.420776043Z" level=error msg="Failed to destroy network for sandbox \"f2ebd01978223216d3607692d8c8c704c5debe172bbbee154e83d38cb0d3bc88\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.421662 containerd[1514]: time="2025-07-12T09:29:40.421619998Z" level=error msg="Failed to destroy network for sandbox \"3f0d6fbce0215d253518a8eb5e322686e591660d65bb42a5d1e9627cf6e618a9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.421940 containerd[1514]: time="2025-07-12T09:29:40.421859728Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-557fb495b6-kkxx2,Uid:4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ebd01978223216d3607692d8c8c704c5debe172bbbee154e83d38cb0d3bc88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.422198 kubelet[2643]: E0712 09:29:40.422165 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ebd01978223216d3607692d8c8c704c5debe172bbbee154e83d38cb0d3bc88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.422263 kubelet[2643]: E0712 09:29:40.422215 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ebd01978223216d3607692d8c8c704c5debe172bbbee154e83d38cb0d3bc88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-557fb495b6-kkxx2" Jul 12 09:29:40.422263 kubelet[2643]: E0712 09:29:40.422233 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ebd01978223216d3607692d8c8c704c5debe172bbbee154e83d38cb0d3bc88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-557fb495b6-kkxx2" Jul 12 09:29:40.422582 containerd[1514]: time="2025-07-12T09:29:40.422535516Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-cf9ps,Uid:f2948faa-6bb4-4b0e-bf03-6e794e9ce052,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f0d6fbce0215d253518a8eb5e322686e591660d65bb42a5d1e9627cf6e618a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.422643 kubelet[2643]: E0712 09:29:40.422392 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-557fb495b6-kkxx2_calico-system(4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-557fb495b6-kkxx2_calico-system(4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f2ebd01978223216d3607692d8c8c704c5debe172bbbee154e83d38cb0d3bc88\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-557fb495b6-kkxx2" podUID="4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817" Jul 12 09:29:40.422891 kubelet[2643]: E0712 09:29:40.422814 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f0d6fbce0215d253518a8eb5e322686e591660d65bb42a5d1e9627cf6e618a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.422891 kubelet[2643]: E0712 09:29:40.422846 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f0d6fbce0215d253518a8eb5e322686e591660d65bb42a5d1e9627cf6e618a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-cf9ps" Jul 12 09:29:40.422891 kubelet[2643]: E0712 09:29:40.422863 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f0d6fbce0215d253518a8eb5e322686e591660d65bb42a5d1e9627cf6e618a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-cf9ps" Jul 12 09:29:40.423062 kubelet[2643]: E0712 09:29:40.422886 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-cf9ps_kube-system(f2948faa-6bb4-4b0e-bf03-6e794e9ce052)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-cf9ps_kube-system(f2948faa-6bb4-4b0e-bf03-6e794e9ce052)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f0d6fbce0215d253518a8eb5e322686e591660d65bb42a5d1e9627cf6e618a9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-cf9ps" podUID="f2948faa-6bb4-4b0e-bf03-6e794e9ce052" Jul 12 09:29:40.426549 containerd[1514]: time="2025-07-12T09:29:40.426514000Z" level=error msg="Failed to destroy network for sandbox \"ad4ed3bc30cf4fa7836e19cb8294c0436b7e4b0ddacd22ea8ffeff981682724d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.427348 containerd[1514]: time="2025-07-12T09:29:40.427277912Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-pcqlb,Uid:58e64786-df0a-4048-91ff-e601fc952e42,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ad4ed3bc30cf4fa7836e19cb8294c0436b7e4b0ddacd22ea8ffeff981682724d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.427531 kubelet[2643]: E0712 09:29:40.427506 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ad4ed3bc30cf4fa7836e19cb8294c0436b7e4b0ddacd22ea8ffeff981682724d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:40.427582 kubelet[2643]: E0712 09:29:40.427566 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ad4ed3bc30cf4fa7836e19cb8294c0436b7e4b0ddacd22ea8ffeff981682724d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7b85d947-pcqlb" Jul 12 09:29:40.427615 kubelet[2643]: E0712 09:29:40.427584 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ad4ed3bc30cf4fa7836e19cb8294c0436b7e4b0ddacd22ea8ffeff981682724d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7b85d947-pcqlb" Jul 12 09:29:40.427653 kubelet[2643]: E0712 09:29:40.427631 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5c7b85d947-pcqlb_calico-apiserver(58e64786-df0a-4048-91ff-e601fc952e42)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5c7b85d947-pcqlb_calico-apiserver(58e64786-df0a-4048-91ff-e601fc952e42)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ad4ed3bc30cf4fa7836e19cb8294c0436b7e4b0ddacd22ea8ffeff981682724d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5c7b85d947-pcqlb" podUID="58e64786-df0a-4048-91ff-e601fc952e42" Jul 12 09:29:40.618169 containerd[1514]: time="2025-07-12T09:29:40.618074879Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Jul 12 09:29:41.519562 systemd[1]: Created slice kubepods-besteffort-pod5e5331e4_f4dc_49b3_a2b8_19205dd650aa.slice - libcontainer container kubepods-besteffort-pod5e5331e4_f4dc_49b3_a2b8_19205dd650aa.slice. Jul 12 09:29:41.521881 containerd[1514]: time="2025-07-12T09:29:41.521807952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sk54t,Uid:5e5331e4-f4dc-49b3-a2b8-19205dd650aa,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:41.568509 containerd[1514]: time="2025-07-12T09:29:41.568444807Z" level=error msg="Failed to destroy network for sandbox \"d0be6227f64ae8de378e8d19e7dfe950f584c71a1eac8ff61fcdad59ff230cbb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:41.571694 containerd[1514]: time="2025-07-12T09:29:41.571590533Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sk54t,Uid:5e5331e4-f4dc-49b3-a2b8-19205dd650aa,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0be6227f64ae8de378e8d19e7dfe950f584c71a1eac8ff61fcdad59ff230cbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:41.571881 kubelet[2643]: E0712 09:29:41.571835 2643 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0be6227f64ae8de378e8d19e7dfe950f584c71a1eac8ff61fcdad59ff230cbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 12 09:29:41.572265 kubelet[2643]: E0712 09:29:41.571893 2643 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0be6227f64ae8de378e8d19e7dfe950f584c71a1eac8ff61fcdad59ff230cbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:41.572265 kubelet[2643]: E0712 09:29:41.571930 2643 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0be6227f64ae8de378e8d19e7dfe950f584c71a1eac8ff61fcdad59ff230cbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sk54t" Jul 12 09:29:41.572265 kubelet[2643]: E0712 09:29:41.571973 2643 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-sk54t_calico-system(5e5331e4-f4dc-49b3-a2b8-19205dd650aa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-sk54t_calico-system(5e5331e4-f4dc-49b3-a2b8-19205dd650aa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d0be6227f64ae8de378e8d19e7dfe950f584c71a1eac8ff61fcdad59ff230cbb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-sk54t" podUID="5e5331e4-f4dc-49b3-a2b8-19205dd650aa" Jul 12 09:29:41.572900 systemd[1]: run-netns-cni\x2ddd1c4f93\x2dde39\x2d2ce2\x2d50ca\x2d29fe74e085a6.mount: Deactivated successfully. Jul 12 09:29:44.544304 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2845667313.mount: Deactivated successfully. Jul 12 09:29:44.598776 containerd[1514]: time="2025-07-12T09:29:44.598713878Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:44.599528 containerd[1514]: time="2025-07-12T09:29:44.599496226Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=152544909" Jul 12 09:29:44.600571 containerd[1514]: time="2025-07-12T09:29:44.600525663Z" level=info msg="ImageCreate event name:\"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:44.602347 containerd[1514]: time="2025-07-12T09:29:44.602308846Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:44.602984 containerd[1514]: time="2025-07-12T09:29:44.602951869Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"152544771\" in 3.984837269s" Jul 12 09:29:44.603019 containerd[1514]: time="2025-07-12T09:29:44.602986831Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\"" Jul 12 09:29:44.613744 containerd[1514]: time="2025-07-12T09:29:44.613698213Z" level=info msg="CreateContainer within sandbox \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jul 12 09:29:44.631178 containerd[1514]: time="2025-07-12T09:29:44.631119635Z" level=info msg="Container bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:44.641372 containerd[1514]: time="2025-07-12T09:29:44.641317559Z" level=info msg="CreateContainer within sandbox \"a6f69f861ed456f98debbdefdcc6389bdd4feedbcd37405ffe650cf110a4e825\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5\"" Jul 12 09:29:44.642948 containerd[1514]: time="2025-07-12T09:29:44.642170270Z" level=info msg="StartContainer for \"bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5\"" Jul 12 09:29:44.648075 containerd[1514]: time="2025-07-12T09:29:44.648017518Z" level=info msg="connecting to shim bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5" address="unix:///run/containerd/s/bf12311d7eabc5670af9f6b52eb1569f2907325a63a029a2d417993dbd38ab0e" protocol=ttrpc version=3 Jul 12 09:29:44.671131 systemd[1]: Started cri-containerd-bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5.scope - libcontainer container bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5. Jul 12 09:29:44.793256 containerd[1514]: time="2025-07-12T09:29:44.793204822Z" level=info msg="StartContainer for \"bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5\" returns successfully" Jul 12 09:29:44.919942 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jul 12 09:29:44.920051 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jul 12 09:29:45.156029 kubelet[2643]: I0712 09:29:45.155983 2643 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-backend-key-pair\") pod \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\" (UID: \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\") " Jul 12 09:29:45.156390 kubelet[2643]: I0712 09:29:45.156041 2643 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctqmw\" (UniqueName: \"kubernetes.io/projected/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-kube-api-access-ctqmw\") pod \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\" (UID: \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\") " Jul 12 09:29:45.156390 kubelet[2643]: I0712 09:29:45.156069 2643 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-ca-bundle\") pod \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\" (UID: \"f0b9a0d1-4bba-408e-a8de-97ec9d5156e7\") " Jul 12 09:29:45.157635 kubelet[2643]: I0712 09:29:45.157511 2643 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "f0b9a0d1-4bba-408e-a8de-97ec9d5156e7" (UID: "f0b9a0d1-4bba-408e-a8de-97ec9d5156e7"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 12 09:29:45.160665 kubelet[2643]: I0712 09:29:45.160622 2643 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-kube-api-access-ctqmw" (OuterVolumeSpecName: "kube-api-access-ctqmw") pod "f0b9a0d1-4bba-408e-a8de-97ec9d5156e7" (UID: "f0b9a0d1-4bba-408e-a8de-97ec9d5156e7"). InnerVolumeSpecName "kube-api-access-ctqmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 12 09:29:45.167668 kubelet[2643]: I0712 09:29:45.167618 2643 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "f0b9a0d1-4bba-408e-a8de-97ec9d5156e7" (UID: "f0b9a0d1-4bba-408e-a8de-97ec9d5156e7"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 12 09:29:45.256645 kubelet[2643]: I0712 09:29:45.256600 2643 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jul 12 09:29:45.256645 kubelet[2643]: I0712 09:29:45.256636 2643 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Jul 12 09:29:45.256645 kubelet[2643]: I0712 09:29:45.256647 2643 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctqmw\" (UniqueName: \"kubernetes.io/projected/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7-kube-api-access-ctqmw\") on node \"localhost\" DevicePath \"\"" Jul 12 09:29:45.520125 systemd[1]: Removed slice kubepods-besteffort-podf0b9a0d1_4bba_408e_a8de_97ec9d5156e7.slice - libcontainer container kubepods-besteffort-podf0b9a0d1_4bba_408e_a8de_97ec9d5156e7.slice. Jul 12 09:29:45.544947 systemd[1]: var-lib-kubelet-pods-f0b9a0d1\x2d4bba\x2d408e\x2da8de\x2d97ec9d5156e7-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dctqmw.mount: Deactivated successfully. Jul 12 09:29:45.545044 systemd[1]: var-lib-kubelet-pods-f0b9a0d1\x2d4bba\x2d408e\x2da8de\x2d97ec9d5156e7-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jul 12 09:29:45.669641 kubelet[2643]: I0712 09:29:45.669577 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-md287" podStartSLOduration=1.93348879 podStartE2EDuration="14.669560112s" podCreationTimestamp="2025-07-12 09:29:31 +0000 UTC" firstStartedPulling="2025-07-12 09:29:31.867638735 +0000 UTC m=+20.439419487" lastFinishedPulling="2025-07-12 09:29:44.603710057 +0000 UTC m=+33.175490809" observedRunningTime="2025-07-12 09:29:45.650506334 +0000 UTC m=+34.222287126" watchObservedRunningTime="2025-07-12 09:29:45.669560112 +0000 UTC m=+34.241340824" Jul 12 09:29:45.722822 systemd[1]: Created slice kubepods-besteffort-podd8532931_900c_41f2_8df2_615d97ab2e12.slice - libcontainer container kubepods-besteffort-podd8532931_900c_41f2_8df2_615d97ab2e12.slice. Jul 12 09:29:45.759466 kubelet[2643]: I0712 09:29:45.759416 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8532931-900c-41f2-8df2-615d97ab2e12-whisker-ca-bundle\") pod \"whisker-578c565cd4-5nf77\" (UID: \"d8532931-900c-41f2-8df2-615d97ab2e12\") " pod="calico-system/whisker-578c565cd4-5nf77" Jul 12 09:29:45.759466 kubelet[2643]: I0712 09:29:45.759465 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d8532931-900c-41f2-8df2-615d97ab2e12-whisker-backend-key-pair\") pod \"whisker-578c565cd4-5nf77\" (UID: \"d8532931-900c-41f2-8df2-615d97ab2e12\") " pod="calico-system/whisker-578c565cd4-5nf77" Jul 12 09:29:45.759613 kubelet[2643]: I0712 09:29:45.759485 2643 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-298f5\" (UniqueName: \"kubernetes.io/projected/d8532931-900c-41f2-8df2-615d97ab2e12-kube-api-access-298f5\") pod \"whisker-578c565cd4-5nf77\" (UID: \"d8532931-900c-41f2-8df2-615d97ab2e12\") " pod="calico-system/whisker-578c565cd4-5nf77" Jul 12 09:29:45.847128 kubelet[2643]: I0712 09:29:45.846735 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:45.847339 kubelet[2643]: E0712 09:29:45.847295 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:46.043710 containerd[1514]: time="2025-07-12T09:29:46.043643459Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-578c565cd4-5nf77,Uid:d8532931-900c-41f2-8df2-615d97ab2e12,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:46.258569 systemd-networkd[1432]: cali55952cbdb9a: Link UP Jul 12 09:29:46.260525 systemd-networkd[1432]: cali55952cbdb9a: Gained carrier Jul 12 09:29:46.276890 containerd[1514]: 2025-07-12 09:29:46.063 [INFO][3784] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jul 12 09:29:46.276890 containerd[1514]: 2025-07-12 09:29:46.111 [INFO][3784] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--578c565cd4--5nf77-eth0 whisker-578c565cd4- calico-system d8532931-900c-41f2-8df2-615d97ab2e12 917 0 2025-07-12 09:29:45 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:578c565cd4 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-578c565cd4-5nf77 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali55952cbdb9a [] [] }} ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-" Jul 12 09:29:46.276890 containerd[1514]: 2025-07-12 09:29:46.112 [INFO][3784] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.276890 containerd[1514]: 2025-07-12 09:29:46.197 [INFO][3798] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" HandleID="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Workload="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.197 [INFO][3798] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" HandleID="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Workload="localhost-k8s-whisker--578c565cd4--5nf77-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000536800), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-578c565cd4-5nf77", "timestamp":"2025-07-12 09:29:46.197744044 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.197 [INFO][3798] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.198 [INFO][3798] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.199 [INFO][3798] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.209 [INFO][3798] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" host="localhost" Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.217 [INFO][3798] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.223 [INFO][3798] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.225 [INFO][3798] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.227 [INFO][3798] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:46.277112 containerd[1514]: 2025-07-12 09:29:46.227 [INFO][3798] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" host="localhost" Jul 12 09:29:46.277304 containerd[1514]: 2025-07-12 09:29:46.229 [INFO][3798] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4 Jul 12 09:29:46.277304 containerd[1514]: 2025-07-12 09:29:46.232 [INFO][3798] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" host="localhost" Jul 12 09:29:46.277304 containerd[1514]: 2025-07-12 09:29:46.240 [INFO][3798] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" host="localhost" Jul 12 09:29:46.277304 containerd[1514]: 2025-07-12 09:29:46.240 [INFO][3798] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" host="localhost" Jul 12 09:29:46.277304 containerd[1514]: 2025-07-12 09:29:46.240 [INFO][3798] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:46.277304 containerd[1514]: 2025-07-12 09:29:46.240 [INFO][3798] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" HandleID="k8s-pod-network.314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Workload="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.277413 containerd[1514]: 2025-07-12 09:29:46.245 [INFO][3784] cni-plugin/k8s.go 418: Populated endpoint ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--578c565cd4--5nf77-eth0", GenerateName:"whisker-578c565cd4-", Namespace:"calico-system", SelfLink:"", UID:"d8532931-900c-41f2-8df2-615d97ab2e12", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"578c565cd4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-578c565cd4-5nf77", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali55952cbdb9a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:46.277413 containerd[1514]: 2025-07-12 09:29:46.245 [INFO][3784] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.277477 containerd[1514]: 2025-07-12 09:29:46.246 [INFO][3784] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali55952cbdb9a ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.277477 containerd[1514]: 2025-07-12 09:29:46.258 [INFO][3784] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.277523 containerd[1514]: 2025-07-12 09:29:46.259 [INFO][3784] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--578c565cd4--5nf77-eth0", GenerateName:"whisker-578c565cd4-", Namespace:"calico-system", SelfLink:"", UID:"d8532931-900c-41f2-8df2-615d97ab2e12", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"578c565cd4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4", Pod:"whisker-578c565cd4-5nf77", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali55952cbdb9a", MAC:"c2:0b:d3:ed:20:03", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:46.277618 containerd[1514]: 2025-07-12 09:29:46.273 [INFO][3784] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" Namespace="calico-system" Pod="whisker-578c565cd4-5nf77" WorkloadEndpoint="localhost-k8s-whisker--578c565cd4--5nf77-eth0" Jul 12 09:29:46.340934 containerd[1514]: time="2025-07-12T09:29:46.340497170Z" level=info msg="connecting to shim 314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4" address="unix:///run/containerd/s/f221c1c6722dc10620fc5e5eb88f69315601864181b1170eb169bb39feed4208" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:46.391292 systemd[1]: Started cri-containerd-314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4.scope - libcontainer container 314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4. Jul 12 09:29:46.424272 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:46.465858 containerd[1514]: time="2025-07-12T09:29:46.465817594Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-578c565cd4-5nf77,Uid:d8532931-900c-41f2-8df2-615d97ab2e12,Namespace:calico-system,Attempt:0,} returns sandbox id \"314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4\"" Jul 12 09:29:46.482365 containerd[1514]: time="2025-07-12T09:29:46.482318385Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Jul 12 09:29:46.635145 kubelet[2643]: I0712 09:29:46.635049 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:46.635443 kubelet[2643]: E0712 09:29:46.635359 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:46.669603 systemd-networkd[1432]: vxlan.calico: Link UP Jul 12 09:29:46.669610 systemd-networkd[1432]: vxlan.calico: Gained carrier Jul 12 09:29:47.472366 containerd[1514]: time="2025-07-12T09:29:47.472295023Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4605614" Jul 12 09:29:47.474499 containerd[1514]: time="2025-07-12T09:29:47.474416891Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:47.475642 containerd[1514]: time="2025-07-12T09:29:47.475612090Z" level=info msg="ImageCreate event name:\"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:47.476317 containerd[1514]: time="2025-07-12T09:29:47.476292432Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:47.477624 containerd[1514]: time="2025-07-12T09:29:47.477572673Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"5974847\" in 995.203966ms" Jul 12 09:29:47.477624 containerd[1514]: time="2025-07-12T09:29:47.477609955Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\"" Jul 12 09:29:47.494034 containerd[1514]: time="2025-07-12T09:29:47.493996724Z" level=info msg="CreateContainer within sandbox \"314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Jul 12 09:29:47.503770 containerd[1514]: time="2025-07-12T09:29:47.503565114Z" level=info msg="Container 05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:47.509857 containerd[1514]: time="2025-07-12T09:29:47.509825156Z" level=info msg="CreateContainer within sandbox \"314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8\"" Jul 12 09:29:47.510320 containerd[1514]: time="2025-07-12T09:29:47.510295252Z" level=info msg="StartContainer for \"05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8\"" Jul 12 09:29:47.511280 containerd[1514]: time="2025-07-12T09:29:47.511256323Z" level=info msg="connecting to shim 05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8" address="unix:///run/containerd/s/f221c1c6722dc10620fc5e5eb88f69315601864181b1170eb169bb39feed4208" protocol=ttrpc version=3 Jul 12 09:29:47.518499 kubelet[2643]: I0712 09:29:47.518470 2643 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0b9a0d1-4bba-408e-a8de-97ec9d5156e7" path="/var/lib/kubelet/pods/f0b9a0d1-4bba-408e-a8de-97ec9d5156e7/volumes" Jul 12 09:29:47.536065 systemd[1]: Started cri-containerd-05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8.scope - libcontainer container 05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8. Jul 12 09:29:47.571518 containerd[1514]: time="2025-07-12T09:29:47.571465430Z" level=info msg="StartContainer for \"05c35b7f715e26dcf80ce37de772b1999139fc7a55697b0cac374a815c8baca8\" returns successfully" Jul 12 09:29:47.574081 containerd[1514]: time="2025-07-12T09:29:47.574018352Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Jul 12 09:29:48.087056 systemd-networkd[1432]: cali55952cbdb9a: Gained IPv6LL Jul 12 09:29:48.434997 kubelet[2643]: I0712 09:29:48.434875 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:48.550277 containerd[1514]: time="2025-07-12T09:29:48.550232665Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5\" id:\"225c0ef6283e7851704bb0227aff69166fa4ee3c8770fb1f99680891fa51462a\" pid:4114 exited_at:{seconds:1752312588 nanos:549930936}" Jul 12 09:29:48.656041 containerd[1514]: time="2025-07-12T09:29:48.655999941Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5\" id:\"b40094357176f86e27c3fdaca5f544368ecbcd09f4664cc4de4094bf048e85e8\" pid:4140 exited_at:{seconds:1752312588 nanos:655696292}" Jul 12 09:29:48.664537 systemd-networkd[1432]: vxlan.calico: Gained IPv6LL Jul 12 09:29:48.903477 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount365104449.mount: Deactivated successfully. Jul 12 09:29:48.918634 containerd[1514]: time="2025-07-12T09:29:48.918401288Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:48.919360 containerd[1514]: time="2025-07-12T09:29:48.919325877Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=30814581" Jul 12 09:29:48.921944 containerd[1514]: time="2025-07-12T09:29:48.921850116Z" level=info msg="ImageCreate event name:\"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:48.923778 containerd[1514]: time="2025-07-12T09:29:48.923751296Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:48.924524 containerd[1514]: time="2025-07-12T09:29:48.924489119Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"30814411\" in 1.350439485s" Jul 12 09:29:48.924524 containerd[1514]: time="2025-07-12T09:29:48.924522760Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\"" Jul 12 09:29:48.927491 containerd[1514]: time="2025-07-12T09:29:48.927463212Z" level=info msg="CreateContainer within sandbox \"314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Jul 12 09:29:48.934941 containerd[1514]: time="2025-07-12T09:29:48.934326147Z" level=info msg="Container 8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:48.946167 containerd[1514]: time="2025-07-12T09:29:48.946130317Z" level=info msg="CreateContainer within sandbox \"314adfe8a53ef033ef70759a57e634c2ae22f112d5ce6214c4dfe4dd868bc6d4\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785\"" Jul 12 09:29:48.954402 containerd[1514]: time="2025-07-12T09:29:48.954370096Z" level=info msg="StartContainer for \"8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785\"" Jul 12 09:29:48.955650 containerd[1514]: time="2025-07-12T09:29:48.955625975Z" level=info msg="connecting to shim 8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785" address="unix:///run/containerd/s/f221c1c6722dc10620fc5e5eb88f69315601864181b1170eb169bb39feed4208" protocol=ttrpc version=3 Jul 12 09:29:48.994345 systemd[1]: Started cri-containerd-8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785.scope - libcontainer container 8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785. Jul 12 09:29:49.028041 containerd[1514]: time="2025-07-12T09:29:49.027996579Z" level=info msg="StartContainer for \"8e7e5d84e34e30ddd3331353621e686ad05c0a612fa81c2a7b4d3cbbc13f6785\" returns successfully" Jul 12 09:29:49.655931 kubelet[2643]: I0712 09:29:49.655840 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-578c565cd4-5nf77" podStartSLOduration=2.197564013 podStartE2EDuration="4.655822722s" podCreationTimestamp="2025-07-12 09:29:45 +0000 UTC" firstStartedPulling="2025-07-12 09:29:46.467000594 +0000 UTC m=+35.038781346" lastFinishedPulling="2025-07-12 09:29:48.925259303 +0000 UTC m=+37.497040055" observedRunningTime="2025-07-12 09:29:49.654963296 +0000 UTC m=+38.226744048" watchObservedRunningTime="2025-07-12 09:29:49.655822722 +0000 UTC m=+38.227603514" Jul 12 09:29:51.146121 systemd[1]: Started sshd@7-10.0.0.27:22-10.0.0.1:40858.service - OpenSSH per-connection server daemon (10.0.0.1:40858). Jul 12 09:29:51.205085 sshd[4204]: Accepted publickey for core from 10.0.0.1 port 40858 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:29:51.206379 sshd-session[4204]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:29:51.210694 systemd-logind[1493]: New session 8 of user core. Jul 12 09:29:51.229051 systemd[1]: Started session-8.scope - Session 8 of User core. Jul 12 09:29:51.398068 sshd[4207]: Connection closed by 10.0.0.1 port 40858 Jul 12 09:29:51.398707 sshd-session[4204]: pam_unix(sshd:session): session closed for user core Jul 12 09:29:51.402406 systemd[1]: sshd@7-10.0.0.27:22-10.0.0.1:40858.service: Deactivated successfully. Jul 12 09:29:51.404192 systemd[1]: session-8.scope: Deactivated successfully. Jul 12 09:29:51.405088 systemd-logind[1493]: Session 8 logged out. Waiting for processes to exit. Jul 12 09:29:51.406607 systemd-logind[1493]: Removed session 8. Jul 12 09:29:51.515640 containerd[1514]: time="2025-07-12T09:29:51.515375936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-fjdcg,Uid:056e934c-fa80-4772-a8bc-0ff57d009fc2,Namespace:calico-apiserver,Attempt:0,}" Jul 12 09:29:51.622622 systemd-networkd[1432]: cali6c665289248: Link UP Jul 12 09:29:51.623580 systemd-networkd[1432]: cali6c665289248: Gained carrier Jul 12 09:29:51.637700 containerd[1514]: 2025-07-12 09:29:51.555 [INFO][4228] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0 calico-apiserver-5c7b85d947- calico-apiserver 056e934c-fa80-4772-a8bc-0ff57d009fc2 851 0 2025-07-12 09:29:26 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5c7b85d947 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-5c7b85d947-fjdcg eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali6c665289248 [] [] }} ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-" Jul 12 09:29:51.637700 containerd[1514]: 2025-07-12 09:29:51.555 [INFO][4228] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.637700 containerd[1514]: 2025-07-12 09:29:51.582 [INFO][4242] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" HandleID="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Workload="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.582 [INFO][4242] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" HandleID="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Workload="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002d53c0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-5c7b85d947-fjdcg", "timestamp":"2025-07-12 09:29:51.582754234 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.582 [INFO][4242] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.583 [INFO][4242] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.584 [INFO][4242] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.592 [INFO][4242] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" host="localhost" Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.598 [INFO][4242] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.602 [INFO][4242] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.604 [INFO][4242] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.606 [INFO][4242] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:51.638060 containerd[1514]: 2025-07-12 09:29:51.606 [INFO][4242] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" host="localhost" Jul 12 09:29:51.638374 containerd[1514]: 2025-07-12 09:29:51.607 [INFO][4242] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb Jul 12 09:29:51.638374 containerd[1514]: 2025-07-12 09:29:51.611 [INFO][4242] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" host="localhost" Jul 12 09:29:51.638374 containerd[1514]: 2025-07-12 09:29:51.616 [INFO][4242] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" host="localhost" Jul 12 09:29:51.638374 containerd[1514]: 2025-07-12 09:29:51.616 [INFO][4242] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" host="localhost" Jul 12 09:29:51.638374 containerd[1514]: 2025-07-12 09:29:51.616 [INFO][4242] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:51.638374 containerd[1514]: 2025-07-12 09:29:51.616 [INFO][4242] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" HandleID="k8s-pod-network.5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Workload="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.638558 containerd[1514]: 2025-07-12 09:29:51.619 [INFO][4228] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0", GenerateName:"calico-apiserver-5c7b85d947-", Namespace:"calico-apiserver", SelfLink:"", UID:"056e934c-fa80-4772-a8bc-0ff57d009fc2", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7b85d947", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-5c7b85d947-fjdcg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6c665289248", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:51.638619 containerd[1514]: 2025-07-12 09:29:51.619 [INFO][4228] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.638619 containerd[1514]: 2025-07-12 09:29:51.619 [INFO][4228] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6c665289248 ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.638619 containerd[1514]: 2025-07-12 09:29:51.623 [INFO][4228] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.638698 containerd[1514]: 2025-07-12 09:29:51.624 [INFO][4228] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0", GenerateName:"calico-apiserver-5c7b85d947-", Namespace:"calico-apiserver", SelfLink:"", UID:"056e934c-fa80-4772-a8bc-0ff57d009fc2", ResourceVersion:"851", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7b85d947", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb", Pod:"calico-apiserver-5c7b85d947-fjdcg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6c665289248", MAC:"f2:4e:1d:7b:a0:49", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:51.638748 containerd[1514]: 2025-07-12 09:29:51.633 [INFO][4228] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-fjdcg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--fjdcg-eth0" Jul 12 09:29:51.659654 containerd[1514]: time="2025-07-12T09:29:51.659512200Z" level=info msg="connecting to shim 5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb" address="unix:///run/containerd/s/8d25e2863aad0e4655f955d1f1fd626d04e195181672ce3e41230e8269fec74a" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:51.698130 systemd[1]: Started cri-containerd-5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb.scope - libcontainer container 5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb. Jul 12 09:29:51.709026 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:51.745078 containerd[1514]: time="2025-07-12T09:29:51.745040099Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-fjdcg,Uid:056e934c-fa80-4772-a8bc-0ff57d009fc2,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb\"" Jul 12 09:29:51.746586 containerd[1514]: time="2025-07-12T09:29:51.746563503Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Jul 12 09:29:52.514069 kubelet[2643]: E0712 09:29:52.514012 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:52.515309 containerd[1514]: time="2025-07-12T09:29:52.514656116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-pcqlb,Uid:58e64786-df0a-4048-91ff-e601fc952e42,Namespace:calico-apiserver,Attempt:0,}" Jul 12 09:29:52.515309 containerd[1514]: time="2025-07-12T09:29:52.514906483Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-cf9ps,Uid:f2948faa-6bb4-4b0e-bf03-6e794e9ce052,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:52.515309 containerd[1514]: time="2025-07-12T09:29:52.515157890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-557fb495b6-kkxx2,Uid:4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:52.645102 systemd-networkd[1432]: cali1c0ab8c5379: Link UP Jul 12 09:29:52.646246 systemd-networkd[1432]: cali1c0ab8c5379: Gained carrier Jul 12 09:29:52.659382 containerd[1514]: 2025-07-12 09:29:52.565 [INFO][4323] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0 coredns-7c65d6cfc9- kube-system f2948faa-6bb4-4b0e-bf03-6e794e9ce052 852 0 2025-07-12 09:29:18 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7c65d6cfc9-cf9ps eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali1c0ab8c5379 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-" Jul 12 09:29:52.659382 containerd[1514]: 2025-07-12 09:29:52.565 [INFO][4323] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.659382 containerd[1514]: 2025-07-12 09:29:52.593 [INFO][4358] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" HandleID="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Workload="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.593 [INFO][4358] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" HandleID="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Workload="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002e5200), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7c65d6cfc9-cf9ps", "timestamp":"2025-07-12 09:29:52.593716408 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.593 [INFO][4358] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.593 [INFO][4358] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.593 [INFO][4358] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.603 [INFO][4358] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" host="localhost" Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.610 [INFO][4358] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.620 [INFO][4358] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.622 [INFO][4358] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.623 [INFO][4358] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:52.659789 containerd[1514]: 2025-07-12 09:29:52.623 [INFO][4358] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" host="localhost" Jul 12 09:29:52.660035 containerd[1514]: 2025-07-12 09:29:52.625 [INFO][4358] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0 Jul 12 09:29:52.660035 containerd[1514]: 2025-07-12 09:29:52.628 [INFO][4358] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" host="localhost" Jul 12 09:29:52.660035 containerd[1514]: 2025-07-12 09:29:52.633 [INFO][4358] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" host="localhost" Jul 12 09:29:52.660035 containerd[1514]: 2025-07-12 09:29:52.633 [INFO][4358] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" host="localhost" Jul 12 09:29:52.660035 containerd[1514]: 2025-07-12 09:29:52.633 [INFO][4358] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:52.660035 containerd[1514]: 2025-07-12 09:29:52.633 [INFO][4358] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" HandleID="k8s-pod-network.dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Workload="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.660162 containerd[1514]: 2025-07-12 09:29:52.636 [INFO][4323] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"f2948faa-6bb4-4b0e-bf03-6e794e9ce052", ResourceVersion:"852", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7c65d6cfc9-cf9ps", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1c0ab8c5379", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:52.660245 containerd[1514]: 2025-07-12 09:29:52.637 [INFO][4323] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.660245 containerd[1514]: 2025-07-12 09:29:52.637 [INFO][4323] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1c0ab8c5379 ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.660245 containerd[1514]: 2025-07-12 09:29:52.646 [INFO][4323] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.660448 containerd[1514]: 2025-07-12 09:29:52.646 [INFO][4323] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"f2948faa-6bb4-4b0e-bf03-6e794e9ce052", ResourceVersion:"852", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0", Pod:"coredns-7c65d6cfc9-cf9ps", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali1c0ab8c5379", MAC:"1a:82:48:c6:cf:a0", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:52.660448 containerd[1514]: 2025-07-12 09:29:52.656 [INFO][4323] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" Namespace="kube-system" Pod="coredns-7c65d6cfc9-cf9ps" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--cf9ps-eth0" Jul 12 09:29:52.705373 containerd[1514]: time="2025-07-12T09:29:52.704082938Z" level=info msg="connecting to shim dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0" address="unix:///run/containerd/s/54a43cf017bb39931e45431a2edf6b11910ce6b9f80d4fd1541845dd264d6437" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:52.736506 systemd[1]: Started cri-containerd-dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0.scope - libcontainer container dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0. Jul 12 09:29:52.779172 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:52.793570 systemd-networkd[1432]: cali40ce50392ea: Link UP Jul 12 09:29:52.797009 systemd-networkd[1432]: cali40ce50392ea: Gained carrier Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.567 [INFO][4310] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0 calico-apiserver-5c7b85d947- calico-apiserver 58e64786-df0a-4048-91ff-e601fc952e42 853 0 2025-07-12 09:29:26 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5c7b85d947 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-5c7b85d947-pcqlb eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali40ce50392ea [] [] }} ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.568 [INFO][4310] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.595 [INFO][4356] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" HandleID="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Workload="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.595 [INFO][4356] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" HandleID="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Workload="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c3160), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-5c7b85d947-pcqlb", "timestamp":"2025-07-12 09:29:52.595247771 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.595 [INFO][4356] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.633 [INFO][4356] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.633 [INFO][4356] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.706 [INFO][4356] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.712 [INFO][4356] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.724 [INFO][4356] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.726 [INFO][4356] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.730 [INFO][4356] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.730 [INFO][4356] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.732 [INFO][4356] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.737 [INFO][4356] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.761 [INFO][4356] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.761 [INFO][4356] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" host="localhost" Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.761 [INFO][4356] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:52.819431 containerd[1514]: 2025-07-12 09:29:52.761 [INFO][4356] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" HandleID="k8s-pod-network.efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Workload="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.820123 containerd[1514]: 2025-07-12 09:29:52.771 [INFO][4310] cni-plugin/k8s.go 418: Populated endpoint ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0", GenerateName:"calico-apiserver-5c7b85d947-", Namespace:"calico-apiserver", SelfLink:"", UID:"58e64786-df0a-4048-91ff-e601fc952e42", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7b85d947", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-5c7b85d947-pcqlb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali40ce50392ea", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:52.820123 containerd[1514]: 2025-07-12 09:29:52.771 [INFO][4310] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.820123 containerd[1514]: 2025-07-12 09:29:52.771 [INFO][4310] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali40ce50392ea ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.820123 containerd[1514]: 2025-07-12 09:29:52.782 [INFO][4310] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.820123 containerd[1514]: 2025-07-12 09:29:52.783 [INFO][4310] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0", GenerateName:"calico-apiserver-5c7b85d947-", Namespace:"calico-apiserver", SelfLink:"", UID:"58e64786-df0a-4048-91ff-e601fc952e42", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7b85d947", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae", Pod:"calico-apiserver-5c7b85d947-pcqlb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali40ce50392ea", MAC:"d2:d9:a4:b9:fe:2b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:52.820123 containerd[1514]: 2025-07-12 09:29:52.808 [INFO][4310] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" Namespace="calico-apiserver" Pod="calico-apiserver-5c7b85d947-pcqlb" WorkloadEndpoint="localhost-k8s-calico--apiserver--5c7b85d947--pcqlb-eth0" Jul 12 09:29:52.820123 containerd[1514]: time="2025-07-12T09:29:52.819751295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-cf9ps,Uid:f2948faa-6bb4-4b0e-bf03-6e794e9ce052,Namespace:kube-system,Attempt:0,} returns sandbox id \"dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0\"" Jul 12 09:29:52.822939 kubelet[2643]: E0712 09:29:52.821811 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:52.835020 containerd[1514]: time="2025-07-12T09:29:52.834976321Z" level=info msg="CreateContainer within sandbox \"dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 12 09:29:52.862945 containerd[1514]: time="2025-07-12T09:29:52.860677561Z" level=info msg="Container c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:52.863248 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1713002431.mount: Deactivated successfully. Jul 12 09:29:52.875525 containerd[1514]: time="2025-07-12T09:29:52.875447494Z" level=info msg="CreateContainer within sandbox \"dfe8e74ae548e193539ea9fdd93101b9037b61868cbac180aebcbf95190f02d0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0\"" Jul 12 09:29:52.880306 containerd[1514]: time="2025-07-12T09:29:52.880262109Z" level=info msg="StartContainer for \"c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0\"" Jul 12 09:29:52.881111 containerd[1514]: time="2025-07-12T09:29:52.881073692Z" level=info msg="connecting to shim c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0" address="unix:///run/containerd/s/54a43cf017bb39931e45431a2edf6b11910ce6b9f80d4fd1541845dd264d6437" protocol=ttrpc version=3 Jul 12 09:29:52.905205 systemd-networkd[1432]: calibfdb33375f1: Link UP Jul 12 09:29:52.907136 containerd[1514]: time="2025-07-12T09:29:52.906854693Z" level=info msg="connecting to shim efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae" address="unix:///run/containerd/s/6948c6a89f0ee165235fcbe3db5430dc1052e622017495dd182631093f79af74" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:52.908066 systemd-networkd[1432]: calibfdb33375f1: Gained carrier Jul 12 09:29:52.924099 systemd[1]: Started cri-containerd-c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0.scope - libcontainer container c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0. Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.568 [INFO][4313] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0 calico-kube-controllers-557fb495b6- calico-system 4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817 846 0 2025-07-12 09:29:31 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:557fb495b6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-557fb495b6-kkxx2 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calibfdb33375f1 [] [] }} ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.568 [INFO][4313] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.597 [INFO][4369] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" HandleID="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Workload="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.597 [INFO][4369] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" HandleID="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Workload="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001af450), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-557fb495b6-kkxx2", "timestamp":"2025-07-12 09:29:52.597679839 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.597 [INFO][4369] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.761 [INFO][4369] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.761 [INFO][4369] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.822 [INFO][4369] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.837 [INFO][4369] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.850 [INFO][4369] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.856 [INFO][4369] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.862 [INFO][4369] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.864 [INFO][4369] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.869 [INFO][4369] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.873 [INFO][4369] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.891 [INFO][4369] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.891 [INFO][4369] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" host="localhost" Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.891 [INFO][4369] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:52.931237 containerd[1514]: 2025-07-12 09:29:52.891 [INFO][4369] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" HandleID="k8s-pod-network.ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Workload="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.931769 containerd[1514]: 2025-07-12 09:29:52.899 [INFO][4313] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0", GenerateName:"calico-kube-controllers-557fb495b6-", Namespace:"calico-system", SelfLink:"", UID:"4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817", ResourceVersion:"846", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"557fb495b6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-557fb495b6-kkxx2", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calibfdb33375f1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:52.931769 containerd[1514]: 2025-07-12 09:29:52.899 [INFO][4313] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.931769 containerd[1514]: 2025-07-12 09:29:52.899 [INFO][4313] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibfdb33375f1 ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.931769 containerd[1514]: 2025-07-12 09:29:52.910 [INFO][4313] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.931769 containerd[1514]: 2025-07-12 09:29:52.911 [INFO][4313] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0", GenerateName:"calico-kube-controllers-557fb495b6-", Namespace:"calico-system", SelfLink:"", UID:"4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817", ResourceVersion:"846", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"557fb495b6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b", Pod:"calico-kube-controllers-557fb495b6-kkxx2", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calibfdb33375f1", MAC:"7e:d1:01:26:98:be", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:52.931769 containerd[1514]: 2025-07-12 09:29:52.921 [INFO][4313] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" Namespace="calico-system" Pod="calico-kube-controllers-557fb495b6-kkxx2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--557fb495b6--kkxx2-eth0" Jul 12 09:29:52.940166 systemd[1]: Started cri-containerd-efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae.scope - libcontainer container efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae. Jul 12 09:29:52.954312 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:52.979866 containerd[1514]: time="2025-07-12T09:29:52.979825936Z" level=info msg="StartContainer for \"c8aaf5a91d8d7ce384c64d34c49d3dbe8301c4a8a5ad5a8884535ee9ef36ade0\" returns successfully" Jul 12 09:29:52.986868 containerd[1514]: time="2025-07-12T09:29:52.984947319Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7b85d947-pcqlb,Uid:58e64786-df0a-4048-91ff-e601fc952e42,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae\"" Jul 12 09:29:53.005414 containerd[1514]: time="2025-07-12T09:29:53.003387274Z" level=info msg="connecting to shim ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b" address="unix:///run/containerd/s/67b601478d024282e9e1a8444fea93e00973295226fc8bab1449c31851e1ca56" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:53.036853 systemd[1]: Started cri-containerd-ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b.scope - libcontainer container ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b. Jul 12 09:29:53.063014 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:53.097506 containerd[1514]: time="2025-07-12T09:29:53.097455920Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-557fb495b6-kkxx2,Uid:4aa0af8e-b7e4-46e7-a2a1-12cadfbfe817,Namespace:calico-system,Attempt:0,} returns sandbox id \"ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b\"" Jul 12 09:29:53.207039 systemd-networkd[1432]: cali6c665289248: Gained IPv6LL Jul 12 09:29:53.516668 containerd[1514]: time="2025-07-12T09:29:53.516619113Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:53.517394 containerd[1514]: time="2025-07-12T09:29:53.517168728Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=44517149" Jul 12 09:29:53.518378 containerd[1514]: time="2025-07-12T09:29:53.518341520Z" level=info msg="ImageCreate event name:\"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:53.521452 containerd[1514]: time="2025-07-12T09:29:53.521407204Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:53.521825 containerd[1514]: time="2025-07-12T09:29:53.521794214Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"45886406\" in 1.77520543s" Jul 12 09:29:53.521870 containerd[1514]: time="2025-07-12T09:29:53.521823055Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\"" Jul 12 09:29:53.523968 containerd[1514]: time="2025-07-12T09:29:53.523934593Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Jul 12 09:29:53.525616 containerd[1514]: time="2025-07-12T09:29:53.525411513Z" level=info msg="CreateContainer within sandbox \"5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jul 12 09:29:53.535957 containerd[1514]: time="2025-07-12T09:29:53.535320543Z" level=info msg="Container 80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:53.541534 containerd[1514]: time="2025-07-12T09:29:53.541479551Z" level=info msg="CreateContainer within sandbox \"5e6adf9bd2cedc4d303ff83514dc6dac4ef4b9861f91626050b54690335a45fb\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73\"" Jul 12 09:29:53.543052 containerd[1514]: time="2025-07-12T09:29:53.543017273Z" level=info msg="StartContainer for \"80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73\"" Jul 12 09:29:53.544625 containerd[1514]: time="2025-07-12T09:29:53.544586236Z" level=info msg="connecting to shim 80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73" address="unix:///run/containerd/s/8d25e2863aad0e4655f955d1f1fd626d04e195181672ce3e41230e8269fec74a" protocol=ttrpc version=3 Jul 12 09:29:53.570087 systemd[1]: Started cri-containerd-80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73.scope - libcontainer container 80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73. Jul 12 09:29:53.612104 containerd[1514]: time="2025-07-12T09:29:53.612054876Z" level=info msg="StartContainer for \"80bbc4d814311e3df114e313d6febaee3f0f517b042b347dab60a1530756df73\" returns successfully" Jul 12 09:29:53.660382 kubelet[2643]: E0712 09:29:53.660346 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:53.673957 kubelet[2643]: I0712 09:29:53.673878 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-cf9ps" podStartSLOduration=35.673846322 podStartE2EDuration="35.673846322s" podCreationTimestamp="2025-07-12 09:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 09:29:53.671407655 +0000 UTC m=+42.243188447" watchObservedRunningTime="2025-07-12 09:29:53.673846322 +0000 UTC m=+42.245627074" Jul 12 09:29:53.694621 kubelet[2643]: I0712 09:29:53.694340 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-5c7b85d947-fjdcg" podStartSLOduration=25.917390803 podStartE2EDuration="27.69432312s" podCreationTimestamp="2025-07-12 09:29:26 +0000 UTC" firstStartedPulling="2025-07-12 09:29:51.746298016 +0000 UTC m=+40.318078728" lastFinishedPulling="2025-07-12 09:29:53.523230293 +0000 UTC m=+42.095011045" observedRunningTime="2025-07-12 09:29:53.69431596 +0000 UTC m=+42.266096712" watchObservedRunningTime="2025-07-12 09:29:53.69432312 +0000 UTC m=+42.266103872" Jul 12 09:29:53.796115 containerd[1514]: time="2025-07-12T09:29:53.796007214Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:53.797141 containerd[1514]: time="2025-07-12T09:29:53.797098164Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=77" Jul 12 09:29:53.798604 containerd[1514]: time="2025-07-12T09:29:53.798491562Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"45886406\" in 274.522488ms" Jul 12 09:29:53.798604 containerd[1514]: time="2025-07-12T09:29:53.798522642Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\"" Jul 12 09:29:53.799704 containerd[1514]: time="2025-07-12T09:29:53.799514549Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Jul 12 09:29:53.802563 containerd[1514]: time="2025-07-12T09:29:53.802088140Z" level=info msg="CreateContainer within sandbox \"efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jul 12 09:29:53.814120 containerd[1514]: time="2025-07-12T09:29:53.814085867Z" level=info msg="Container 2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:53.820877 containerd[1514]: time="2025-07-12T09:29:53.820844771Z" level=info msg="CreateContainer within sandbox \"efe1bcc64a1d45ae51514370f6d82a69ee6a7f53f906c0bf00892195070c3cae\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f\"" Jul 12 09:29:53.821378 containerd[1514]: time="2025-07-12T09:29:53.821353985Z" level=info msg="StartContainer for \"2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f\"" Jul 12 09:29:53.822543 containerd[1514]: time="2025-07-12T09:29:53.822486176Z" level=info msg="connecting to shim 2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f" address="unix:///run/containerd/s/6948c6a89f0ee165235fcbe3db5430dc1052e622017495dd182631093f79af74" protocol=ttrpc version=3 Jul 12 09:29:53.845138 systemd[1]: Started cri-containerd-2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f.scope - libcontainer container 2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f. Jul 12 09:29:53.880973 containerd[1514]: time="2025-07-12T09:29:53.880012905Z" level=info msg="StartContainer for \"2a6b9e7aa3b38040cab355e7357354592613c9c53b0a57a662642798fb53b91f\" returns successfully" Jul 12 09:29:53.975070 systemd-networkd[1432]: cali1c0ab8c5379: Gained IPv6LL Jul 12 09:29:54.515672 containerd[1514]: time="2025-07-12T09:29:54.515112925Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sk54t,Uid:5e5331e4-f4dc-49b3-a2b8-19205dd650aa,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:54.553183 systemd-networkd[1432]: cali40ce50392ea: Gained IPv6LL Jul 12 09:29:54.646928 systemd-networkd[1432]: calie0d5028cd72: Link UP Jul 12 09:29:54.647216 systemd-networkd[1432]: calie0d5028cd72: Gained carrier Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.560 [INFO][4667] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--sk54t-eth0 csi-node-driver- calico-system 5e5331e4-f4dc-49b3-a2b8-19205dd650aa 745 0 2025-07-12 09:29:31 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:57bd658777 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-sk54t eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calie0d5028cd72 [] [] }} ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.560 [INFO][4667] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.586 [INFO][4682] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" HandleID="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Workload="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.586 [INFO][4682] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" HandleID="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Workload="localhost-k8s-csi--node--driver--sk54t-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000140ce0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-sk54t", "timestamp":"2025-07-12 09:29:54.58633282 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.586 [INFO][4682] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.586 [INFO][4682] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.586 [INFO][4682] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.610 [INFO][4682] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.615 [INFO][4682] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.620 [INFO][4682] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.622 [INFO][4682] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.625 [INFO][4682] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.625 [INFO][4682] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.627 [INFO][4682] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893 Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.633 [INFO][4682] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.640 [INFO][4682] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.640 [INFO][4682] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" host="localhost" Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.640 [INFO][4682] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:54.661538 containerd[1514]: 2025-07-12 09:29:54.640 [INFO][4682] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" HandleID="k8s-pod-network.710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Workload="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.663759 containerd[1514]: 2025-07-12 09:29:54.643 [INFO][4667] cni-plugin/k8s.go 418: Populated endpoint ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--sk54t-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"5e5331e4-f4dc-49b3-a2b8-19205dd650aa", ResourceVersion:"745", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-sk54t", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie0d5028cd72", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:54.663759 containerd[1514]: 2025-07-12 09:29:54.643 [INFO][4667] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.663759 containerd[1514]: 2025-07-12 09:29:54.643 [INFO][4667] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie0d5028cd72 ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.663759 containerd[1514]: 2025-07-12 09:29:54.645 [INFO][4667] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.663759 containerd[1514]: 2025-07-12 09:29:54.645 [INFO][4667] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--sk54t-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"5e5331e4-f4dc-49b3-a2b8-19205dd650aa", ResourceVersion:"745", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893", Pod:"csi-node-driver-sk54t", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie0d5028cd72", MAC:"aa:ad:dd:72:c4:da", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:54.663759 containerd[1514]: 2025-07-12 09:29:54.657 [INFO][4667] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" Namespace="calico-system" Pod="csi-node-driver-sk54t" WorkloadEndpoint="localhost-k8s-csi--node--driver--sk54t-eth0" Jul 12 09:29:54.669425 kubelet[2643]: I0712 09:29:54.669398 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:54.670721 kubelet[2643]: E0712 09:29:54.670683 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:54.680268 systemd-networkd[1432]: calibfdb33375f1: Gained IPv6LL Jul 12 09:29:54.682925 kubelet[2643]: I0712 09:29:54.682836 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-5c7b85d947-pcqlb" podStartSLOduration=27.87662231 podStartE2EDuration="28.682819107s" podCreationTimestamp="2025-07-12 09:29:26 +0000 UTC" firstStartedPulling="2025-07-12 09:29:52.993184149 +0000 UTC m=+41.564964901" lastFinishedPulling="2025-07-12 09:29:53.799380946 +0000 UTC m=+42.371161698" observedRunningTime="2025-07-12 09:29:54.682602182 +0000 UTC m=+43.254383054" watchObservedRunningTime="2025-07-12 09:29:54.682819107 +0000 UTC m=+43.254599859" Jul 12 09:29:54.701922 containerd[1514]: time="2025-07-12T09:29:54.701834373Z" level=info msg="connecting to shim 710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893" address="unix:///run/containerd/s/1fd9c9843da86057fbd05d97d4232b85cdb855f96249f88c3c771d2a859c9989" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:54.735236 systemd[1]: Started cri-containerd-710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893.scope - libcontainer container 710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893. Jul 12 09:29:54.747346 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:54.764712 containerd[1514]: time="2025-07-12T09:29:54.764677125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sk54t,Uid:5e5331e4-f4dc-49b3-a2b8-19205dd650aa,Namespace:calico-system,Attempt:0,} returns sandbox id \"710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893\"" Jul 12 09:29:55.515233 kubelet[2643]: E0712 09:29:55.514557 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:55.515436 containerd[1514]: time="2025-07-12T09:29:55.515032729Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-4d9r4,Uid:c3a983bd-242c-4076-abf5-8ee86b62b108,Namespace:kube-system,Attempt:0,}" Jul 12 09:29:55.516958 containerd[1514]: time="2025-07-12T09:29:55.516138998Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-hc8ld,Uid:a67ae9fe-f378-4098-b84d-38bd8456798a,Namespace:calico-system,Attempt:0,}" Jul 12 09:29:55.674758 kubelet[2643]: I0712 09:29:55.674730 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:55.675895 kubelet[2643]: E0712 09:29:55.675033 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:55.708842 systemd-networkd[1432]: cali90a616df187: Link UP Jul 12 09:29:55.709062 systemd-networkd[1432]: cali90a616df187: Gained carrier Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.600 [INFO][4753] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0 coredns-7c65d6cfc9- kube-system c3a983bd-242c-4076-abf5-8ee86b62b108 850 0 2025-07-12 09:29:18 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7c65d6cfc9-4d9r4 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali90a616df187 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.601 [INFO][4753] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.648 [INFO][4789] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" HandleID="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Workload="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.648 [INFO][4789] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" HandleID="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Workload="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c3240), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7c65d6cfc9-4d9r4", "timestamp":"2025-07-12 09:29:55.64824563 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.648 [INFO][4789] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.648 [INFO][4789] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.648 [INFO][4789] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.661 [INFO][4789] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.669 [INFO][4789] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.674 [INFO][4789] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.677 [INFO][4789] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.680 [INFO][4789] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.680 [INFO][4789] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.682 [INFO][4789] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.690 [INFO][4789] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.698 [INFO][4789] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.698 [INFO][4789] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" host="localhost" Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.698 [INFO][4789] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:55.723697 containerd[1514]: 2025-07-12 09:29:55.699 [INFO][4789] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" HandleID="k8s-pod-network.cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Workload="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.724644 containerd[1514]: 2025-07-12 09:29:55.703 [INFO][4753] cni-plugin/k8s.go 418: Populated endpoint ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"c3a983bd-242c-4076-abf5-8ee86b62b108", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7c65d6cfc9-4d9r4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali90a616df187", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:55.724644 containerd[1514]: 2025-07-12 09:29:55.703 [INFO][4753] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.724644 containerd[1514]: 2025-07-12 09:29:55.703 [INFO][4753] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali90a616df187 ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.724644 containerd[1514]: 2025-07-12 09:29:55.710 [INFO][4753] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.724644 containerd[1514]: 2025-07-12 09:29:55.711 [INFO][4753] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"c3a983bd-242c-4076-abf5-8ee86b62b108", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a", Pod:"coredns-7c65d6cfc9-4d9r4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali90a616df187", MAC:"1e:11:37:c0:67:19", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:55.724644 containerd[1514]: 2025-07-12 09:29:55.720 [INFO][4753] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" Namespace="kube-system" Pod="coredns-7c65d6cfc9-4d9r4" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--4d9r4-eth0" Jul 12 09:29:55.751952 containerd[1514]: time="2025-07-12T09:29:55.751901203Z" level=info msg="connecting to shim cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a" address="unix:///run/containerd/s/0d152b97868a82a26d3a8348926e12730c8a9ce6fb3055767e3eefdb0103a962" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:55.790113 systemd[1]: Started cri-containerd-cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a.scope - libcontainer container cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a. Jul 12 09:29:55.806467 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:55.823692 systemd-networkd[1432]: cali69d2b4d843d: Link UP Jul 12 09:29:55.825699 systemd-networkd[1432]: cali69d2b4d843d: Gained carrier Jul 12 09:29:55.846427 containerd[1514]: time="2025-07-12T09:29:55.846364057Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-4d9r4,Uid:c3a983bd-242c-4076-abf5-8ee86b62b108,Namespace:kube-system,Attempt:0,} returns sandbox id \"cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a\"" Jul 12 09:29:55.848032 kubelet[2643]: E0712 09:29:55.847985 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:55.852821 containerd[1514]: time="2025-07-12T09:29:55.852258770Z" level=info msg="CreateContainer within sandbox \"cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.596 [INFO][4765] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0 goldmane-58fd7646b9- calico-system a67ae9fe-f378-4098-b84d-38bd8456798a 849 0 2025-07-12 09:29:31 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:58fd7646b9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-58fd7646b9-hc8ld eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali69d2b4d843d [] [] }} ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.597 [INFO][4765] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.659 [INFO][4783] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" HandleID="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Workload="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.660 [INFO][4783] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" HandleID="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Workload="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40005b51d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-58fd7646b9-hc8ld", "timestamp":"2025-07-12 09:29:55.659650326 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.660 [INFO][4783] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.699 [INFO][4783] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.699 [INFO][4783] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.762 [INFO][4783] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.771 [INFO][4783] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.778 [INFO][4783] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.783 [INFO][4783] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.786 [INFO][4783] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.786 [INFO][4783] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.789 [INFO][4783] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.795 [INFO][4783] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.816 [INFO][4783] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.816 [INFO][4783] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" host="localhost" Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.816 [INFO][4783] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 12 09:29:55.856572 containerd[1514]: 2025-07-12 09:29:55.816 [INFO][4783] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" HandleID="k8s-pod-network.aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Workload="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.857895 containerd[1514]: 2025-07-12 09:29:55.819 [INFO][4765] cni-plugin/k8s.go 418: Populated endpoint ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"a67ae9fe-f378-4098-b84d-38bd8456798a", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-58fd7646b9-hc8ld", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali69d2b4d843d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:55.857895 containerd[1514]: 2025-07-12 09:29:55.820 [INFO][4765] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.857895 containerd[1514]: 2025-07-12 09:29:55.820 [INFO][4765] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali69d2b4d843d ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.857895 containerd[1514]: 2025-07-12 09:29:55.827 [INFO][4765] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.857895 containerd[1514]: 2025-07-12 09:29:55.830 [INFO][4765] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"a67ae9fe-f378-4098-b84d-38bd8456798a", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.July, 12, 9, 29, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea", Pod:"goldmane-58fd7646b9-hc8ld", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali69d2b4d843d", MAC:"fe:3f:5f:d5:7c:eb", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 12 09:29:55.857895 containerd[1514]: 2025-07-12 09:29:55.851 [INFO][4765] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" Namespace="calico-system" Pod="goldmane-58fd7646b9-hc8ld" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--hc8ld-eth0" Jul 12 09:29:55.874141 containerd[1514]: time="2025-07-12T09:29:55.874094337Z" level=info msg="Container 62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:55.903192 containerd[1514]: time="2025-07-12T09:29:55.902459954Z" level=info msg="CreateContainer within sandbox \"cc43bd805274e81ec4cfa3864489c1a6c9228897ce2a10b2182d32995704ff5a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3\"" Jul 12 09:29:55.906222 containerd[1514]: time="2025-07-12T09:29:55.905244387Z" level=info msg="StartContainer for \"62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3\"" Jul 12 09:29:55.907536 containerd[1514]: time="2025-07-12T09:29:55.907510646Z" level=info msg="connecting to shim 62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3" address="unix:///run/containerd/s/0d152b97868a82a26d3a8348926e12730c8a9ce6fb3055767e3eefdb0103a962" protocol=ttrpc version=3 Jul 12 09:29:55.943252 systemd[1]: Started cri-containerd-62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3.scope - libcontainer container 62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3. Jul 12 09:29:55.947611 containerd[1514]: time="2025-07-12T09:29:55.947329600Z" level=info msg="connecting to shim aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea" address="unix:///run/containerd/s/d9db48a8c070fc677ad14f5b702216c8afb20c8b6aed3623fdccb4b617e1913e" namespace=k8s.io protocol=ttrpc version=3 Jul 12 09:29:55.970240 systemd[1]: Started cri-containerd-aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea.scope - libcontainer container aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea. Jul 12 09:29:55.988583 containerd[1514]: time="2025-07-12T09:29:55.988538031Z" level=info msg="StartContainer for \"62c83e7cb62ee42fbe6bbc0dfb4bc9a58259d6ec92e8881a2a4a4d7e1e4ff7f3\" returns successfully" Jul 12 09:29:55.991243 systemd-resolved[1356]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 12 09:29:56.029136 containerd[1514]: time="2025-07-12T09:29:56.029083988Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-hc8ld,Uid:a67ae9fe-f378-4098-b84d-38bd8456798a,Namespace:calico-system,Attempt:0,} returns sandbox id \"aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea\"" Jul 12 09:29:56.203612 containerd[1514]: time="2025-07-12T09:29:56.203502897Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:56.204314 containerd[1514]: time="2025-07-12T09:29:56.204285837Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=48128336" Jul 12 09:29:56.205214 containerd[1514]: time="2025-07-12T09:29:56.205179620Z" level=info msg="ImageCreate event name:\"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:56.207950 containerd[1514]: time="2025-07-12T09:29:56.207464238Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:56.208043 containerd[1514]: time="2025-07-12T09:29:56.208012412Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"49497545\" in 2.408470501s" Jul 12 09:29:56.208076 containerd[1514]: time="2025-07-12T09:29:56.208048173Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\"" Jul 12 09:29:56.209120 containerd[1514]: time="2025-07-12T09:29:56.209096239Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Jul 12 09:29:56.215162 containerd[1514]: time="2025-07-12T09:29:56.215132792Z" level=info msg="CreateContainer within sandbox \"ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jul 12 09:29:56.221744 containerd[1514]: time="2025-07-12T09:29:56.221706239Z" level=info msg="Container 4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:56.227982 containerd[1514]: time="2025-07-12T09:29:56.227948518Z" level=info msg="CreateContainer within sandbox \"ede1a3d9f04eb01fcd8ce34413bb11ce040eda7338a3be24b8de6c4abed0b14b\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd\"" Jul 12 09:29:56.228834 containerd[1514]: time="2025-07-12T09:29:56.228450811Z" level=info msg="StartContainer for \"4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd\"" Jul 12 09:29:56.230027 containerd[1514]: time="2025-07-12T09:29:56.229964809Z" level=info msg="connecting to shim 4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd" address="unix:///run/containerd/s/67b601478d024282e9e1a8444fea93e00973295226fc8bab1449c31851e1ca56" protocol=ttrpc version=3 Jul 12 09:29:56.253105 systemd[1]: Started cri-containerd-4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd.scope - libcontainer container 4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd. Jul 12 09:29:56.293174 containerd[1514]: time="2025-07-12T09:29:56.293124733Z" level=info msg="StartContainer for \"4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd\" returns successfully" Jul 12 09:29:56.408123 systemd-networkd[1432]: calie0d5028cd72: Gained IPv6LL Jul 12 09:29:56.415391 systemd[1]: Started sshd@8-10.0.0.27:22-10.0.0.1:59944.service - OpenSSH per-connection server daemon (10.0.0.1:59944). Jul 12 09:29:56.486818 sshd[4992]: Accepted publickey for core from 10.0.0.1 port 59944 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:29:56.488416 sshd-session[4992]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:29:56.492240 systemd-logind[1493]: New session 9 of user core. Jul 12 09:29:56.502074 systemd[1]: Started session-9.scope - Session 9 of User core. Jul 12 09:29:56.548134 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3307813466.mount: Deactivated successfully. Jul 12 09:29:56.685710 kubelet[2643]: E0712 09:29:56.685676 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:56.714616 kubelet[2643]: I0712 09:29:56.714360 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-557fb495b6-kkxx2" podStartSLOduration=22.605030053 podStartE2EDuration="25.714338469s" podCreationTimestamp="2025-07-12 09:29:31 +0000 UTC" firstStartedPulling="2025-07-12 09:29:53.099620459 +0000 UTC m=+41.671401211" lastFinishedPulling="2025-07-12 09:29:56.208928915 +0000 UTC m=+44.780709627" observedRunningTime="2025-07-12 09:29:56.696228169 +0000 UTC m=+45.268008921" watchObservedRunningTime="2025-07-12 09:29:56.714338469 +0000 UTC m=+45.286119221" Jul 12 09:29:56.714949 kubelet[2643]: I0712 09:29:56.714901 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-4d9r4" podStartSLOduration=38.714892123 podStartE2EDuration="38.714892123s" podCreationTimestamp="2025-07-12 09:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 09:29:56.713779374 +0000 UTC m=+45.285560126" watchObservedRunningTime="2025-07-12 09:29:56.714892123 +0000 UTC m=+45.286672875" Jul 12 09:29:56.759908 sshd[4997]: Connection closed by 10.0.0.1 port 59944 Jul 12 09:29:56.760229 sshd-session[4992]: pam_unix(sshd:session): session closed for user core Jul 12 09:29:56.763822 systemd[1]: sshd@8-10.0.0.27:22-10.0.0.1:59944.service: Deactivated successfully. Jul 12 09:29:56.765604 systemd[1]: session-9.scope: Deactivated successfully. Jul 12 09:29:56.766688 systemd-logind[1493]: Session 9 logged out. Waiting for processes to exit. Jul 12 09:29:56.767906 systemd-logind[1493]: Removed session 9. Jul 12 09:29:57.304173 systemd-networkd[1432]: cali90a616df187: Gained IPv6LL Jul 12 09:29:57.432048 systemd-networkd[1432]: cali69d2b4d843d: Gained IPv6LL Jul 12 09:29:57.612558 containerd[1514]: time="2025-07-12T09:29:57.612433297Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:57.613270 containerd[1514]: time="2025-07-12T09:29:57.613226316Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8225702" Jul 12 09:29:57.613891 containerd[1514]: time="2025-07-12T09:29:57.613863812Z" level=info msg="ImageCreate event name:\"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:57.615830 containerd[1514]: time="2025-07-12T09:29:57.615781500Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:57.616600 containerd[1514]: time="2025-07-12T09:29:57.616566519Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"9594943\" in 1.407357557s" Jul 12 09:29:57.616646 containerd[1514]: time="2025-07-12T09:29:57.616600200Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\"" Jul 12 09:29:57.618194 containerd[1514]: time="2025-07-12T09:29:57.618144959Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Jul 12 09:29:57.618960 containerd[1514]: time="2025-07-12T09:29:57.618867217Z" level=info msg="CreateContainer within sandbox \"710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Jul 12 09:29:57.634256 containerd[1514]: time="2025-07-12T09:29:57.634202958Z" level=info msg="Container 31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:57.638205 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount927693913.mount: Deactivated successfully. Jul 12 09:29:57.643411 containerd[1514]: time="2025-07-12T09:29:57.643368465Z" level=info msg="CreateContainer within sandbox \"710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23\"" Jul 12 09:29:57.643975 containerd[1514]: time="2025-07-12T09:29:57.643935519Z" level=info msg="StartContainer for \"31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23\"" Jul 12 09:29:57.645653 containerd[1514]: time="2025-07-12T09:29:57.645621041Z" level=info msg="connecting to shim 31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23" address="unix:///run/containerd/s/1fd9c9843da86057fbd05d97d4232b85cdb855f96249f88c3c771d2a859c9989" protocol=ttrpc version=3 Jul 12 09:29:57.668089 systemd[1]: Started cri-containerd-31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23.scope - libcontainer container 31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23. Jul 12 09:29:57.692227 kubelet[2643]: I0712 09:29:57.692164 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:29:57.692584 kubelet[2643]: E0712 09:29:57.692497 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:57.703823 containerd[1514]: time="2025-07-12T09:29:57.703776126Z" level=info msg="StartContainer for \"31efc0b9f3b3330cfe73280ca0c13d1492ff60e77bc6693381bbb384df20da23\" returns successfully" Jul 12 09:29:58.704091 kubelet[2643]: E0712 09:29:58.703820 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:29:59.169502 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount290691292.mount: Deactivated successfully. Jul 12 09:29:59.584428 containerd[1514]: time="2025-07-12T09:29:59.584384783Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:59.585029 containerd[1514]: time="2025-07-12T09:29:59.585001677Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=61838790" Jul 12 09:29:59.585747 containerd[1514]: time="2025-07-12T09:29:59.585724095Z" level=info msg="ImageCreate event name:\"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:59.588245 containerd[1514]: time="2025-07-12T09:29:59.587957788Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:29:59.588651 containerd[1514]: time="2025-07-12T09:29:59.588626644Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"61838636\" in 1.969826189s" Jul 12 09:29:59.588943 containerd[1514]: time="2025-07-12T09:29:59.588897010Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\"" Jul 12 09:29:59.590605 containerd[1514]: time="2025-07-12T09:29:59.590566210Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Jul 12 09:29:59.592959 containerd[1514]: time="2025-07-12T09:29:59.592700461Z" level=info msg="CreateContainer within sandbox \"aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Jul 12 09:29:59.600759 containerd[1514]: time="2025-07-12T09:29:59.599698028Z" level=info msg="Container 187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:29:59.608455 containerd[1514]: time="2025-07-12T09:29:59.608410315Z" level=info msg="CreateContainer within sandbox \"aff312651b86248dca60ca8104b4992e9659b056235c60e944122680ab3d16ea\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472\"" Jul 12 09:29:59.609099 containerd[1514]: time="2025-07-12T09:29:59.608967409Z" level=info msg="StartContainer for \"187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472\"" Jul 12 09:29:59.610134 containerd[1514]: time="2025-07-12T09:29:59.610105036Z" level=info msg="connecting to shim 187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472" address="unix:///run/containerd/s/d9db48a8c070fc677ad14f5b702216c8afb20c8b6aed3623fdccb4b617e1913e" protocol=ttrpc version=3 Jul 12 09:29:59.636140 systemd[1]: Started cri-containerd-187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472.scope - libcontainer container 187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472. Jul 12 09:29:59.678649 containerd[1514]: time="2025-07-12T09:29:59.678611909Z" level=info msg="StartContainer for \"187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472\" returns successfully" Jul 12 09:29:59.760972 kubelet[2643]: I0712 09:29:59.760807 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-58fd7646b9-hc8ld" podStartSLOduration=25.201630064 podStartE2EDuration="28.760787508s" podCreationTimestamp="2025-07-12 09:29:31 +0000 UTC" firstStartedPulling="2025-07-12 09:29:56.030971916 +0000 UTC m=+44.602752668" lastFinishedPulling="2025-07-12 09:29:59.59012936 +0000 UTC m=+48.161910112" observedRunningTime="2025-07-12 09:29:59.758799461 +0000 UTC m=+48.330580213" watchObservedRunningTime="2025-07-12 09:29:59.760787508 +0000 UTC m=+48.332568260" Jul 12 09:30:00.700902 containerd[1514]: time="2025-07-12T09:30:00.700150946Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:30:00.700902 containerd[1514]: time="2025-07-12T09:30:00.700818521Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=13754366" Jul 12 09:30:00.701786 containerd[1514]: time="2025-07-12T09:30:00.701759783Z" level=info msg="ImageCreate event name:\"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:30:00.704606 containerd[1514]: time="2025-07-12T09:30:00.704553569Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 09:30:00.705238 containerd[1514]: time="2025-07-12T09:30:00.705078981Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"15123559\" in 1.114371848s" Jul 12 09:30:00.705238 containerd[1514]: time="2025-07-12T09:30:00.705110462Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\"" Jul 12 09:30:00.708626 containerd[1514]: time="2025-07-12T09:30:00.708372418Z" level=info msg="CreateContainer within sandbox \"710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Jul 12 09:30:00.729071 containerd[1514]: time="2025-07-12T09:30:00.729036101Z" level=info msg="Container 6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c: CDI devices from CRI Config.CDIDevices: []" Jul 12 09:30:00.739306 containerd[1514]: time="2025-07-12T09:30:00.739261780Z" level=info msg="CreateContainer within sandbox \"710451804b85b741a12fcb004281b8919abac756576eedf2d23d5e28c69fe893\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c\"" Jul 12 09:30:00.740162 containerd[1514]: time="2025-07-12T09:30:00.740132161Z" level=info msg="StartContainer for \"6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c\"" Jul 12 09:30:00.748074 containerd[1514]: time="2025-07-12T09:30:00.748006945Z" level=info msg="connecting to shim 6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c" address="unix:///run/containerd/s/1fd9c9843da86057fbd05d97d4232b85cdb855f96249f88c3c771d2a859c9989" protocol=ttrpc version=3 Jul 12 09:30:00.772146 systemd[1]: Started cri-containerd-6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c.scope - libcontainer container 6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c. Jul 12 09:30:00.826188 containerd[1514]: time="2025-07-12T09:30:00.825312713Z" level=info msg="StartContainer for \"6add32d9d94a24ff795fb022939c7af2cfc0b93e9eca698ce964df1f90a9a10c\" returns successfully" Jul 12 09:30:00.903718 containerd[1514]: time="2025-07-12T09:30:00.903673905Z" level=info msg="TaskExit event in podsandbox handler container_id:\"187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472\" id:\"7e1a278ecea732afe67eaae47db908b3fe68b31b4df0a1dceb60d9bb9dcc546f\" pid:5118 exited_at:{seconds:1752312600 nanos:903266816}" Jul 12 09:30:01.591732 kubelet[2643]: I0712 09:30:01.591631 2643 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Jul 12 09:30:01.591732 kubelet[2643]: I0712 09:30:01.591681 2643 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Jul 12 09:30:01.774364 systemd[1]: Started sshd@9-10.0.0.27:22-10.0.0.1:59948.service - OpenSSH per-connection server daemon (10.0.0.1:59948). Jul 12 09:30:01.851346 sshd[5175]: Accepted publickey for core from 10.0.0.1 port 59948 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:01.852867 sshd-session[5175]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:01.859179 systemd-logind[1493]: New session 10 of user core. Jul 12 09:30:01.873245 systemd[1]: Started session-10.scope - Session 10 of User core. Jul 12 09:30:02.103057 sshd[5178]: Connection closed by 10.0.0.1 port 59948 Jul 12 09:30:02.103315 sshd-session[5175]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:02.111605 systemd[1]: sshd@9-10.0.0.27:22-10.0.0.1:59948.service: Deactivated successfully. Jul 12 09:30:02.114185 systemd[1]: session-10.scope: Deactivated successfully. Jul 12 09:30:02.115236 systemd-logind[1493]: Session 10 logged out. Waiting for processes to exit. Jul 12 09:30:02.117696 systemd[1]: Started sshd@10-10.0.0.27:22-10.0.0.1:59950.service - OpenSSH per-connection server daemon (10.0.0.1:59950). Jul 12 09:30:02.119085 systemd-logind[1493]: Removed session 10. Jul 12 09:30:02.184123 sshd[5192]: Accepted publickey for core from 10.0.0.1 port 59950 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:02.185454 sshd-session[5192]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:02.189552 systemd-logind[1493]: New session 11 of user core. Jul 12 09:30:02.196090 systemd[1]: Started session-11.scope - Session 11 of User core. Jul 12 09:30:02.407455 sshd[5195]: Connection closed by 10.0.0.1 port 59950 Jul 12 09:30:02.406861 sshd-session[5192]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:02.422040 systemd[1]: sshd@10-10.0.0.27:22-10.0.0.1:59950.service: Deactivated successfully. Jul 12 09:30:02.426515 systemd[1]: session-11.scope: Deactivated successfully. Jul 12 09:30:02.430754 systemd-logind[1493]: Session 11 logged out. Waiting for processes to exit. Jul 12 09:30:02.436722 systemd[1]: Started sshd@11-10.0.0.27:22-10.0.0.1:59960.service - OpenSSH per-connection server daemon (10.0.0.1:59960). Jul 12 09:30:02.437448 systemd-logind[1493]: Removed session 11. Jul 12 09:30:02.493465 sshd[5206]: Accepted publickey for core from 10.0.0.1 port 59960 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:02.495175 sshd-session[5206]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:02.499307 systemd-logind[1493]: New session 12 of user core. Jul 12 09:30:02.509106 systemd[1]: Started session-12.scope - Session 12 of User core. Jul 12 09:30:02.671281 sshd[5209]: Connection closed by 10.0.0.1 port 59960 Jul 12 09:30:02.671738 sshd-session[5206]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:02.675722 systemd-logind[1493]: Session 12 logged out. Waiting for processes to exit. Jul 12 09:30:02.676057 systemd[1]: sshd@11-10.0.0.27:22-10.0.0.1:59960.service: Deactivated successfully. Jul 12 09:30:02.677682 systemd[1]: session-12.scope: Deactivated successfully. Jul 12 09:30:02.679624 systemd-logind[1493]: Removed session 12. Jul 12 09:30:06.179597 kubelet[2643]: I0712 09:30:06.179455 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:30:06.233738 containerd[1514]: time="2025-07-12T09:30:06.233690411Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd\" id:\"9c5b1e967d1e1c7d4e517bf5d2f4311a9ee90ae8ba601b20559027e347fcae4a\" pid:5239 exited_at:{seconds:1752312606 nanos:228399179}" Jul 12 09:30:06.248721 kubelet[2643]: I0712 09:30:06.248652 2643 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-sk54t" podStartSLOduration=29.30839536 podStartE2EDuration="35.248635528s" podCreationTimestamp="2025-07-12 09:29:31 +0000 UTC" firstStartedPulling="2025-07-12 09:29:54.766024561 +0000 UTC m=+43.337805313" lastFinishedPulling="2025-07-12 09:30:00.706264729 +0000 UTC m=+49.278045481" observedRunningTime="2025-07-12 09:30:01.734455902 +0000 UTC m=+50.306236654" watchObservedRunningTime="2025-07-12 09:30:06.248635528 +0000 UTC m=+54.820416280" Jul 12 09:30:06.277733 containerd[1514]: time="2025-07-12T09:30:06.277594982Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd\" id:\"f2f997dc938227e394ee8a56d5597b8c807e5795a7b05b18650d8a00a8224a86\" pid:5261 exited_at:{seconds:1752312606 nanos:277226654}" Jul 12 09:30:07.693098 systemd[1]: Started sshd@12-10.0.0.27:22-10.0.0.1:33172.service - OpenSSH per-connection server daemon (10.0.0.1:33172). Jul 12 09:30:07.756050 sshd[5280]: Accepted publickey for core from 10.0.0.1 port 33172 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:07.757398 sshd-session[5280]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:07.761727 systemd-logind[1493]: New session 13 of user core. Jul 12 09:30:07.774067 systemd[1]: Started session-13.scope - Session 13 of User core. Jul 12 09:30:07.941768 sshd[5283]: Connection closed by 10.0.0.1 port 33172 Jul 12 09:30:07.942247 sshd-session[5280]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:07.945694 systemd[1]: sshd@12-10.0.0.27:22-10.0.0.1:33172.service: Deactivated successfully. Jul 12 09:30:07.947546 systemd[1]: session-13.scope: Deactivated successfully. Jul 12 09:30:07.948387 systemd-logind[1493]: Session 13 logged out. Waiting for processes to exit. Jul 12 09:30:07.949437 systemd-logind[1493]: Removed session 13. Jul 12 09:30:12.963091 systemd[1]: Started sshd@13-10.0.0.27:22-10.0.0.1:40966.service - OpenSSH per-connection server daemon (10.0.0.1:40966). Jul 12 09:30:13.025949 sshd[5300]: Accepted publickey for core from 10.0.0.1 port 40966 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:13.027615 sshd-session[5300]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:13.031819 systemd-logind[1493]: New session 14 of user core. Jul 12 09:30:13.043093 systemd[1]: Started session-14.scope - Session 14 of User core. Jul 12 09:30:13.189004 sshd[5303]: Connection closed by 10.0.0.1 port 40966 Jul 12 09:30:13.189322 sshd-session[5300]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:13.192651 systemd[1]: sshd@13-10.0.0.27:22-10.0.0.1:40966.service: Deactivated successfully. Jul 12 09:30:13.195310 systemd[1]: session-14.scope: Deactivated successfully. Jul 12 09:30:13.196871 systemd-logind[1493]: Session 14 logged out. Waiting for processes to exit. Jul 12 09:30:13.198286 systemd-logind[1493]: Removed session 14. Jul 12 09:30:18.208056 systemd[1]: Started sshd@14-10.0.0.27:22-10.0.0.1:40968.service - OpenSSH per-connection server daemon (10.0.0.1:40968). Jul 12 09:30:18.292428 sshd[5319]: Accepted publickey for core from 10.0.0.1 port 40968 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:18.295761 sshd-session[5319]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:18.301518 systemd-logind[1493]: New session 15 of user core. Jul 12 09:30:18.312520 systemd[1]: Started session-15.scope - Session 15 of User core. Jul 12 09:30:18.496897 containerd[1514]: time="2025-07-12T09:30:18.496854224Z" level=info msg="TaskExit event in podsandbox handler container_id:\"187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472\" id:\"1b98e6477b47dfd0785174a0281e68e0aed50700375b455197fd717f55ae8976\" pid:5345 exited_at:{seconds:1752312618 nanos:496560186}" Jul 12 09:30:18.552515 sshd[5322]: Connection closed by 10.0.0.1 port 40968 Jul 12 09:30:18.552836 sshd-session[5319]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:18.557606 systemd[1]: sshd@14-10.0.0.27:22-10.0.0.1:40968.service: Deactivated successfully. Jul 12 09:30:18.561601 systemd[1]: session-15.scope: Deactivated successfully. Jul 12 09:30:18.565062 systemd-logind[1493]: Session 15 logged out. Waiting for processes to exit. Jul 12 09:30:18.567106 systemd-logind[1493]: Removed session 15. Jul 12 09:30:18.571212 containerd[1514]: time="2025-07-12T09:30:18.571032487Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bcc20263742bc6ac9dcd90d124ef9c434460e29576d2575e09769ec8531bafd5\" id:\"75659ccc192481ec3c506e67cf73569f11bbaa1c64ab35e4a586c161f15c0460\" pid:5368 exited_at:{seconds:1752312618 nanos:570592890}" Jul 12 09:30:23.564775 systemd[1]: Started sshd@15-10.0.0.27:22-10.0.0.1:54980.service - OpenSSH per-connection server daemon (10.0.0.1:54980). Jul 12 09:30:23.621925 sshd[5391]: Accepted publickey for core from 10.0.0.1 port 54980 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:23.622774 sshd-session[5391]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:23.629118 systemd-logind[1493]: New session 16 of user core. Jul 12 09:30:23.638130 systemd[1]: Started session-16.scope - Session 16 of User core. Jul 12 09:30:23.874804 sshd[5394]: Connection closed by 10.0.0.1 port 54980 Jul 12 09:30:23.875699 sshd-session[5391]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:23.885884 systemd[1]: sshd@15-10.0.0.27:22-10.0.0.1:54980.service: Deactivated successfully. Jul 12 09:30:23.888391 systemd[1]: session-16.scope: Deactivated successfully. Jul 12 09:30:23.889489 systemd-logind[1493]: Session 16 logged out. Waiting for processes to exit. Jul 12 09:30:23.892229 systemd[1]: Started sshd@16-10.0.0.27:22-10.0.0.1:54986.service - OpenSSH per-connection server daemon (10.0.0.1:54986). Jul 12 09:30:23.895517 systemd-logind[1493]: Removed session 16. Jul 12 09:30:23.961396 sshd[5408]: Accepted publickey for core from 10.0.0.1 port 54986 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:23.962837 sshd-session[5408]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:23.969493 systemd-logind[1493]: New session 17 of user core. Jul 12 09:30:23.975694 systemd[1]: Started session-17.scope - Session 17 of User core. Jul 12 09:30:24.128157 kubelet[2643]: I0712 09:30:24.128035 2643 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 12 09:30:24.210214 sshd[5411]: Connection closed by 10.0.0.1 port 54986 Jul 12 09:30:24.211017 sshd-session[5408]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:24.219451 systemd[1]: sshd@16-10.0.0.27:22-10.0.0.1:54986.service: Deactivated successfully. Jul 12 09:30:24.221503 systemd[1]: session-17.scope: Deactivated successfully. Jul 12 09:30:24.222813 systemd-logind[1493]: Session 17 logged out. Waiting for processes to exit. Jul 12 09:30:24.226011 systemd-logind[1493]: Removed session 17. Jul 12 09:30:24.228074 systemd[1]: Started sshd@17-10.0.0.27:22-10.0.0.1:55000.service - OpenSSH per-connection server daemon (10.0.0.1:55000). Jul 12 09:30:24.288575 sshd[5424]: Accepted publickey for core from 10.0.0.1 port 55000 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:24.289965 sshd-session[5424]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:24.294011 systemd-logind[1493]: New session 18 of user core. Jul 12 09:30:24.304103 systemd[1]: Started session-18.scope - Session 18 of User core. Jul 12 09:30:26.179508 sshd[5427]: Connection closed by 10.0.0.1 port 55000 Jul 12 09:30:26.180503 sshd-session[5424]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:26.197532 systemd[1]: Started sshd@18-10.0.0.27:22-10.0.0.1:55012.service - OpenSSH per-connection server daemon (10.0.0.1:55012). Jul 12 09:30:26.200308 systemd[1]: sshd@17-10.0.0.27:22-10.0.0.1:55000.service: Deactivated successfully. Jul 12 09:30:26.210274 systemd[1]: session-18.scope: Deactivated successfully. Jul 12 09:30:26.210888 systemd[1]: session-18.scope: Consumed 550ms CPU time, 71.7M memory peak. Jul 12 09:30:26.214423 systemd-logind[1493]: Session 18 logged out. Waiting for processes to exit. Jul 12 09:30:26.216876 systemd-logind[1493]: Removed session 18. Jul 12 09:30:26.269439 sshd[5448]: Accepted publickey for core from 10.0.0.1 port 55012 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:26.270661 sshd-session[5448]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:26.278714 systemd-logind[1493]: New session 19 of user core. Jul 12 09:30:26.287107 systemd[1]: Started session-19.scope - Session 19 of User core. Jul 12 09:30:26.694446 sshd[5455]: Connection closed by 10.0.0.1 port 55012 Jul 12 09:30:26.694802 sshd-session[5448]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:26.707648 systemd[1]: sshd@18-10.0.0.27:22-10.0.0.1:55012.service: Deactivated successfully. Jul 12 09:30:26.711388 systemd[1]: session-19.scope: Deactivated successfully. Jul 12 09:30:26.712817 systemd-logind[1493]: Session 19 logged out. Waiting for processes to exit. Jul 12 09:30:26.720727 systemd[1]: Started sshd@19-10.0.0.27:22-10.0.0.1:55022.service - OpenSSH per-connection server daemon (10.0.0.1:55022). Jul 12 09:30:26.722557 systemd-logind[1493]: Removed session 19. Jul 12 09:30:26.773815 sshd[5466]: Accepted publickey for core from 10.0.0.1 port 55022 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:26.776262 sshd-session[5466]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:26.784129 systemd-logind[1493]: New session 20 of user core. Jul 12 09:30:26.799110 systemd[1]: Started session-20.scope - Session 20 of User core. Jul 12 09:30:26.942934 sshd[5469]: Connection closed by 10.0.0.1 port 55022 Jul 12 09:30:26.944086 sshd-session[5466]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:26.949342 systemd[1]: sshd@19-10.0.0.27:22-10.0.0.1:55022.service: Deactivated successfully. Jul 12 09:30:26.951448 systemd[1]: session-20.scope: Deactivated successfully. Jul 12 09:30:26.952319 systemd-logind[1493]: Session 20 logged out. Waiting for processes to exit. Jul 12 09:30:26.954117 systemd-logind[1493]: Removed session 20. Jul 12 09:30:27.515173 kubelet[2643]: E0712 09:30:27.515121 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:30:27.515543 kubelet[2643]: E0712 09:30:27.515412 2643 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jul 12 09:30:31.956854 systemd[1]: Started sshd@20-10.0.0.27:22-10.0.0.1:55030.service - OpenSSH per-connection server daemon (10.0.0.1:55030). Jul 12 09:30:32.013572 sshd[5491]: Accepted publickey for core from 10.0.0.1 port 55030 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:32.014828 sshd-session[5491]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:32.018703 systemd-logind[1493]: New session 21 of user core. Jul 12 09:30:32.025067 systemd[1]: Started session-21.scope - Session 21 of User core. Jul 12 09:30:32.179761 sshd[5494]: Connection closed by 10.0.0.1 port 55030 Jul 12 09:30:32.180095 sshd-session[5491]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:32.183598 systemd[1]: sshd@20-10.0.0.27:22-10.0.0.1:55030.service: Deactivated successfully. Jul 12 09:30:32.185341 systemd[1]: session-21.scope: Deactivated successfully. Jul 12 09:30:32.187728 systemd-logind[1493]: Session 21 logged out. Waiting for processes to exit. Jul 12 09:30:32.190690 systemd-logind[1493]: Removed session 21. Jul 12 09:30:36.210868 containerd[1514]: time="2025-07-12T09:30:36.210802515Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4a97ca27c6b932e5193609ab554c586933b9d0f89b0ff3e36fe850e806cbbcdd\" id:\"105566e67895198fb37e77b6118f6255668a472fa07a2bbacc4e5b493969649b\" pid:5522 exited_at:{seconds:1752312636 nanos:210299594}" Jul 12 09:30:37.193940 systemd[1]: Started sshd@21-10.0.0.27:22-10.0.0.1:55714.service - OpenSSH per-connection server daemon (10.0.0.1:55714). Jul 12 09:30:37.244489 sshd[5533]: Accepted publickey for core from 10.0.0.1 port 55714 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:37.245640 sshd-session[5533]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:37.249378 systemd-logind[1493]: New session 22 of user core. Jul 12 09:30:37.260099 systemd[1]: Started session-22.scope - Session 22 of User core. Jul 12 09:30:37.379539 sshd[5536]: Connection closed by 10.0.0.1 port 55714 Jul 12 09:30:37.379878 sshd-session[5533]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:37.383311 systemd[1]: sshd@21-10.0.0.27:22-10.0.0.1:55714.service: Deactivated successfully. Jul 12 09:30:37.385203 systemd[1]: session-22.scope: Deactivated successfully. Jul 12 09:30:37.385931 systemd-logind[1493]: Session 22 logged out. Waiting for processes to exit. Jul 12 09:30:37.387503 systemd-logind[1493]: Removed session 22. Jul 12 09:30:42.396202 systemd[1]: Started sshd@22-10.0.0.27:22-10.0.0.1:55716.service - OpenSSH per-connection server daemon (10.0.0.1:55716). Jul 12 09:30:42.446723 sshd[5550]: Accepted publickey for core from 10.0.0.1 port 55716 ssh2: RSA SHA256:fhp558siaf39QLJw5fsAHbaRafIwNXdVZ+VoGPeGhpE Jul 12 09:30:42.447521 sshd-session[5550]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 09:30:42.451973 systemd-logind[1493]: New session 23 of user core. Jul 12 09:30:42.457153 systemd[1]: Started session-23.scope - Session 23 of User core. Jul 12 09:30:42.578499 sshd[5553]: Connection closed by 10.0.0.1 port 55716 Jul 12 09:30:42.578260 sshd-session[5550]: pam_unix(sshd:session): session closed for user core Jul 12 09:30:42.582124 systemd-logind[1493]: Session 23 logged out. Waiting for processes to exit. Jul 12 09:30:42.582268 systemd[1]: sshd@22-10.0.0.27:22-10.0.0.1:55716.service: Deactivated successfully. Jul 12 09:30:42.584015 systemd[1]: session-23.scope: Deactivated successfully. Jul 12 09:30:42.587382 systemd-logind[1493]: Removed session 23. Jul 12 09:30:43.444830 containerd[1514]: time="2025-07-12T09:30:43.444746790Z" level=info msg="TaskExit event in podsandbox handler container_id:\"187af369804da91a7cae4dc69417f09f649bcf4c68cbba20946a7e11caf12472\" id:\"64d64158922ab02a58d773451573deabe15705b6fa68288dae2b38fae0344165\" pid:5578 exited_at:{seconds:1752312643 nanos:444305788}"