Sep 11 05:13:54.871685 kernel: Linux version 6.12.46-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu Sep 11 03:19:46 -00 2025 Sep 11 05:13:54.871713 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=51f9391273345312ef1dd27c8e30359e1ffc3b3227e5ed433bbb9ca7d53ecc1c Sep 11 05:13:54.871732 kernel: BIOS-provided physical RAM map: Sep 11 05:13:54.871739 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 11 05:13:54.871746 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 11 05:13:54.871753 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 11 05:13:54.871760 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Sep 11 05:13:54.871767 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 11 05:13:54.871777 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Sep 11 05:13:54.871785 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Sep 11 05:13:54.871792 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Sep 11 05:13:54.871799 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Sep 11 05:13:54.871805 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Sep 11 05:13:54.871812 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Sep 11 05:13:54.871820 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Sep 11 05:13:54.871830 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 11 05:13:54.871840 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Sep 11 05:13:54.871847 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Sep 11 05:13:54.871854 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Sep 11 05:13:54.871861 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Sep 11 05:13:54.871870 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Sep 11 05:13:54.871879 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 11 05:13:54.871887 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 11 05:13:54.871896 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 11 05:13:54.871904 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Sep 11 05:13:54.871916 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 11 05:13:54.871924 kernel: NX (Execute Disable) protection: active Sep 11 05:13:54.871933 kernel: APIC: Static calls initialized Sep 11 05:13:54.871958 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Sep 11 05:13:54.871966 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Sep 11 05:13:54.871973 kernel: extended physical RAM map: Sep 11 05:13:54.871981 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 11 05:13:54.871990 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 11 05:13:54.871999 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 11 05:13:54.872009 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Sep 11 05:13:54.872018 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 11 05:13:54.872028 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Sep 11 05:13:54.872036 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Sep 11 05:13:54.872043 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Sep 11 05:13:54.872050 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Sep 11 05:13:54.872061 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Sep 11 05:13:54.872068 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Sep 11 05:13:54.872078 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Sep 11 05:13:54.872085 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Sep 11 05:13:54.872093 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Sep 11 05:13:54.872100 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Sep 11 05:13:54.872108 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Sep 11 05:13:54.872115 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 11 05:13:54.872122 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Sep 11 05:13:54.872130 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Sep 11 05:13:54.872137 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Sep 11 05:13:54.872145 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Sep 11 05:13:54.872154 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Sep 11 05:13:54.872162 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 11 05:13:54.872169 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 11 05:13:54.872176 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 11 05:13:54.872183 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Sep 11 05:13:54.872191 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 11 05:13:54.872201 kernel: efi: EFI v2.7 by EDK II Sep 11 05:13:54.872209 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Sep 11 05:13:54.872216 kernel: random: crng init done Sep 11 05:13:54.872226 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Sep 11 05:13:54.872234 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Sep 11 05:13:54.872246 kernel: secureboot: Secure boot disabled Sep 11 05:13:54.872253 kernel: SMBIOS 2.8 present. Sep 11 05:13:54.872261 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Sep 11 05:13:54.872268 kernel: DMI: Memory slots populated: 1/1 Sep 11 05:13:54.872276 kernel: Hypervisor detected: KVM Sep 11 05:13:54.872283 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 11 05:13:54.872290 kernel: kvm-clock: using sched offset of 5474500528 cycles Sep 11 05:13:54.872298 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 11 05:13:54.872306 kernel: tsc: Detected 2794.748 MHz processor Sep 11 05:13:54.872314 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 11 05:13:54.872322 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 11 05:13:54.872331 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Sep 11 05:13:54.872339 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 11 05:13:54.872347 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 11 05:13:54.872354 kernel: Using GB pages for direct mapping Sep 11 05:13:54.872362 kernel: ACPI: Early table checksum verification disabled Sep 11 05:13:54.872370 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Sep 11 05:13:54.872377 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Sep 11 05:13:54.872385 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 05:13:54.872393 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 05:13:54.872402 kernel: ACPI: FACS 0x000000009CBDD000 000040 Sep 11 05:13:54.872410 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 05:13:54.872417 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 05:13:54.872425 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 05:13:54.872433 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 05:13:54.872440 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Sep 11 05:13:54.872448 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Sep 11 05:13:54.872455 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Sep 11 05:13:54.872463 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Sep 11 05:13:54.872473 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Sep 11 05:13:54.872480 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Sep 11 05:13:54.872488 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Sep 11 05:13:54.872495 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Sep 11 05:13:54.872503 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Sep 11 05:13:54.872510 kernel: No NUMA configuration found Sep 11 05:13:54.872518 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Sep 11 05:13:54.872525 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Sep 11 05:13:54.872533 kernel: Zone ranges: Sep 11 05:13:54.872543 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 11 05:13:54.872550 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Sep 11 05:13:54.872558 kernel: Normal empty Sep 11 05:13:54.872565 kernel: Device empty Sep 11 05:13:54.872573 kernel: Movable zone start for each node Sep 11 05:13:54.872580 kernel: Early memory node ranges Sep 11 05:13:54.872587 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 11 05:13:54.872595 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Sep 11 05:13:54.872605 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Sep 11 05:13:54.872614 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Sep 11 05:13:54.872622 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Sep 11 05:13:54.872629 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Sep 11 05:13:54.872636 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Sep 11 05:13:54.872644 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Sep 11 05:13:54.872651 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Sep 11 05:13:54.872659 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 11 05:13:54.872669 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 11 05:13:54.872686 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Sep 11 05:13:54.872693 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 11 05:13:54.872701 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Sep 11 05:13:54.872709 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Sep 11 05:13:54.872717 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Sep 11 05:13:54.872735 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Sep 11 05:13:54.872743 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Sep 11 05:13:54.872751 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 11 05:13:54.872759 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 11 05:13:54.872767 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 11 05:13:54.872778 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 11 05:13:54.872785 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 11 05:13:54.872793 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 11 05:13:54.872801 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 11 05:13:54.872809 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 11 05:13:54.872817 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 11 05:13:54.872825 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 11 05:13:54.872833 kernel: TSC deadline timer available Sep 11 05:13:54.872843 kernel: CPU topo: Max. logical packages: 1 Sep 11 05:13:54.872851 kernel: CPU topo: Max. logical dies: 1 Sep 11 05:13:54.872859 kernel: CPU topo: Max. dies per package: 1 Sep 11 05:13:54.872866 kernel: CPU topo: Max. threads per core: 1 Sep 11 05:13:54.872874 kernel: CPU topo: Num. cores per package: 4 Sep 11 05:13:54.872882 kernel: CPU topo: Num. threads per package: 4 Sep 11 05:13:54.872890 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Sep 11 05:13:54.872897 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 11 05:13:54.872905 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 11 05:13:54.872913 kernel: kvm-guest: setup PV sched yield Sep 11 05:13:54.872923 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Sep 11 05:13:54.872931 kernel: Booting paravirtualized kernel on KVM Sep 11 05:13:54.872950 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 11 05:13:54.872959 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 11 05:13:54.872967 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Sep 11 05:13:54.872975 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Sep 11 05:13:54.872983 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 11 05:13:54.872990 kernel: kvm-guest: PV spinlocks enabled Sep 11 05:13:54.872998 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 11 05:13:54.873011 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=51f9391273345312ef1dd27c8e30359e1ffc3b3227e5ed433bbb9ca7d53ecc1c Sep 11 05:13:54.873022 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 11 05:13:54.873030 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 11 05:13:54.873038 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 11 05:13:54.873045 kernel: Fallback order for Node 0: 0 Sep 11 05:13:54.873054 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Sep 11 05:13:54.873061 kernel: Policy zone: DMA32 Sep 11 05:13:54.873069 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 11 05:13:54.873079 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 11 05:13:54.873087 kernel: ftrace: allocating 40106 entries in 157 pages Sep 11 05:13:54.873095 kernel: ftrace: allocated 157 pages with 5 groups Sep 11 05:13:54.873103 kernel: Dynamic Preempt: voluntary Sep 11 05:13:54.873111 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 11 05:13:54.873119 kernel: rcu: RCU event tracing is enabled. Sep 11 05:13:54.873128 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 11 05:13:54.873136 kernel: Trampoline variant of Tasks RCU enabled. Sep 11 05:13:54.873144 kernel: Rude variant of Tasks RCU enabled. Sep 11 05:13:54.873154 kernel: Tracing variant of Tasks RCU enabled. Sep 11 05:13:54.873162 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 11 05:13:54.873173 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 11 05:13:54.873181 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 05:13:54.873189 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 05:13:54.873197 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 05:13:54.873205 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 11 05:13:54.873212 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 11 05:13:54.873220 kernel: Console: colour dummy device 80x25 Sep 11 05:13:54.873231 kernel: printk: legacy console [ttyS0] enabled Sep 11 05:13:54.873239 kernel: ACPI: Core revision 20240827 Sep 11 05:13:54.873247 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 11 05:13:54.873254 kernel: APIC: Switch to symmetric I/O mode setup Sep 11 05:13:54.873262 kernel: x2apic enabled Sep 11 05:13:54.873270 kernel: APIC: Switched APIC routing to: physical x2apic Sep 11 05:13:54.873278 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 11 05:13:54.873286 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 11 05:13:54.873294 kernel: kvm-guest: setup PV IPIs Sep 11 05:13:54.873304 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 11 05:13:54.873313 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Sep 11 05:13:54.873328 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Sep 11 05:13:54.873340 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 11 05:13:54.873350 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 11 05:13:54.873360 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 11 05:13:54.873370 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 11 05:13:54.873380 kernel: Spectre V2 : Mitigation: Retpolines Sep 11 05:13:54.873390 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 11 05:13:54.873404 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 11 05:13:54.873411 kernel: active return thunk: retbleed_return_thunk Sep 11 05:13:54.873419 kernel: RETBleed: Mitigation: untrained return thunk Sep 11 05:13:54.873431 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 11 05:13:54.873439 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 11 05:13:54.873447 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 11 05:13:54.873456 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 11 05:13:54.873464 kernel: active return thunk: srso_return_thunk Sep 11 05:13:54.873474 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 11 05:13:54.873482 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 11 05:13:54.873490 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 11 05:13:54.873498 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 11 05:13:54.873506 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 11 05:13:54.873514 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 11 05:13:54.873522 kernel: Freeing SMP alternatives memory: 32K Sep 11 05:13:54.873529 kernel: pid_max: default: 32768 minimum: 301 Sep 11 05:13:54.873537 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 11 05:13:54.873547 kernel: landlock: Up and running. Sep 11 05:13:54.873555 kernel: SELinux: Initializing. Sep 11 05:13:54.873563 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 11 05:13:54.873570 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 11 05:13:54.873578 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 11 05:13:54.873586 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 11 05:13:54.873594 kernel: ... version: 0 Sep 11 05:13:54.873602 kernel: ... bit width: 48 Sep 11 05:13:54.873610 kernel: ... generic registers: 6 Sep 11 05:13:54.873619 kernel: ... value mask: 0000ffffffffffff Sep 11 05:13:54.873627 kernel: ... max period: 00007fffffffffff Sep 11 05:13:54.873635 kernel: ... fixed-purpose events: 0 Sep 11 05:13:54.873642 kernel: ... event mask: 000000000000003f Sep 11 05:13:54.873650 kernel: signal: max sigframe size: 1776 Sep 11 05:13:54.873658 kernel: rcu: Hierarchical SRCU implementation. Sep 11 05:13:54.873666 kernel: rcu: Max phase no-delay instances is 400. Sep 11 05:13:54.873677 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 11 05:13:54.873685 kernel: smp: Bringing up secondary CPUs ... Sep 11 05:13:54.873694 kernel: smpboot: x86: Booting SMP configuration: Sep 11 05:13:54.873702 kernel: .... node #0, CPUs: #1 #2 #3 Sep 11 05:13:54.873710 kernel: smp: Brought up 1 node, 4 CPUs Sep 11 05:13:54.873718 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Sep 11 05:13:54.873736 kernel: Memory: 2422672K/2565800K available (14336K kernel code, 2429K rwdata, 9988K rodata, 54052K init, 2916K bss, 137196K reserved, 0K cma-reserved) Sep 11 05:13:54.873744 kernel: devtmpfs: initialized Sep 11 05:13:54.873752 kernel: x86/mm: Memory block size: 128MB Sep 11 05:13:54.873760 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Sep 11 05:13:54.873768 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Sep 11 05:13:54.873778 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Sep 11 05:13:54.873786 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Sep 11 05:13:54.873793 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Sep 11 05:13:54.873802 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Sep 11 05:13:54.873810 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 11 05:13:54.873818 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 11 05:13:54.873826 kernel: pinctrl core: initialized pinctrl subsystem Sep 11 05:13:54.873834 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 11 05:13:54.873842 kernel: audit: initializing netlink subsys (disabled) Sep 11 05:13:54.873852 kernel: audit: type=2000 audit(1757567631.661:1): state=initialized audit_enabled=0 res=1 Sep 11 05:13:54.873860 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 11 05:13:54.873868 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 11 05:13:54.873876 kernel: cpuidle: using governor menu Sep 11 05:13:54.873884 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 11 05:13:54.873891 kernel: dca service started, version 1.12.1 Sep 11 05:13:54.873899 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Sep 11 05:13:54.873907 kernel: PCI: Using configuration type 1 for base access Sep 11 05:13:54.873915 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 11 05:13:54.873925 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 11 05:13:54.873933 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 11 05:13:54.873959 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 11 05:13:54.873967 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 11 05:13:54.873975 kernel: ACPI: Added _OSI(Module Device) Sep 11 05:13:54.873983 kernel: ACPI: Added _OSI(Processor Device) Sep 11 05:13:54.873991 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 11 05:13:54.873998 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 11 05:13:54.874006 kernel: ACPI: Interpreter enabled Sep 11 05:13:54.874017 kernel: ACPI: PM: (supports S0 S3 S5) Sep 11 05:13:54.874024 kernel: ACPI: Using IOAPIC for interrupt routing Sep 11 05:13:54.874032 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 11 05:13:54.874041 kernel: PCI: Using E820 reservations for host bridge windows Sep 11 05:13:54.874048 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 11 05:13:54.874056 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 11 05:13:54.874284 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 11 05:13:54.874414 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 11 05:13:54.874542 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 11 05:13:54.874552 kernel: PCI host bridge to bus 0000:00 Sep 11 05:13:54.874697 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 11 05:13:54.874823 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 11 05:13:54.874936 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 11 05:13:54.875092 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Sep 11 05:13:54.875248 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Sep 11 05:13:54.875411 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Sep 11 05:13:54.875525 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 11 05:13:54.875684 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Sep 11 05:13:54.875841 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Sep 11 05:13:54.876035 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Sep 11 05:13:54.876186 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Sep 11 05:13:54.876319 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Sep 11 05:13:54.876441 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 11 05:13:54.876615 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 11 05:13:54.876755 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Sep 11 05:13:54.878069 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Sep 11 05:13:54.878320 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Sep 11 05:13:54.878494 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Sep 11 05:13:54.878652 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Sep 11 05:13:54.878818 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Sep 11 05:13:54.878996 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Sep 11 05:13:54.879686 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Sep 11 05:13:54.879867 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Sep 11 05:13:54.880026 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Sep 11 05:13:54.880155 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Sep 11 05:13:54.880287 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Sep 11 05:13:54.880434 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Sep 11 05:13:54.880559 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 11 05:13:54.880699 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Sep 11 05:13:54.880873 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Sep 11 05:13:54.881064 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Sep 11 05:13:54.881215 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Sep 11 05:13:54.881354 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Sep 11 05:13:54.881368 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 11 05:13:54.881379 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 11 05:13:54.881390 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 11 05:13:54.881401 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 11 05:13:54.881409 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 11 05:13:54.881418 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 11 05:13:54.881430 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 11 05:13:54.881437 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 11 05:13:54.881446 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 11 05:13:54.881454 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 11 05:13:54.881462 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 11 05:13:54.881470 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 11 05:13:54.881478 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 11 05:13:54.881486 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 11 05:13:54.881494 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 11 05:13:54.881504 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 11 05:13:54.881512 kernel: iommu: Default domain type: Translated Sep 11 05:13:54.881520 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 11 05:13:54.881528 kernel: efivars: Registered efivars operations Sep 11 05:13:54.881536 kernel: PCI: Using ACPI for IRQ routing Sep 11 05:13:54.881544 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 11 05:13:54.881554 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Sep 11 05:13:54.881564 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Sep 11 05:13:54.881575 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Sep 11 05:13:54.881585 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Sep 11 05:13:54.881597 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Sep 11 05:13:54.881607 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Sep 11 05:13:54.881617 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Sep 11 05:13:54.881627 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Sep 11 05:13:54.881777 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 11 05:13:54.881903 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 11 05:13:54.882063 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 11 05:13:54.882080 kernel: vgaarb: loaded Sep 11 05:13:54.882089 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 11 05:13:54.882097 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 11 05:13:54.882105 kernel: clocksource: Switched to clocksource kvm-clock Sep 11 05:13:54.882114 kernel: VFS: Disk quotas dquot_6.6.0 Sep 11 05:13:54.882123 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 11 05:13:54.882131 kernel: pnp: PnP ACPI init Sep 11 05:13:54.882305 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Sep 11 05:13:54.882324 kernel: pnp: PnP ACPI: found 6 devices Sep 11 05:13:54.882332 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 11 05:13:54.882341 kernel: NET: Registered PF_INET protocol family Sep 11 05:13:54.882349 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 11 05:13:54.882358 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 11 05:13:54.882366 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 11 05:13:54.882375 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 11 05:13:54.882383 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 11 05:13:54.882391 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 11 05:13:54.882402 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 11 05:13:54.882411 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 11 05:13:54.882419 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 11 05:13:54.882428 kernel: NET: Registered PF_XDP protocol family Sep 11 05:13:54.882554 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Sep 11 05:13:54.882680 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Sep 11 05:13:54.882808 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 11 05:13:54.882921 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 11 05:13:54.883093 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 11 05:13:54.883210 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Sep 11 05:13:54.883332 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Sep 11 05:13:54.883447 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Sep 11 05:13:54.883458 kernel: PCI: CLS 0 bytes, default 64 Sep 11 05:13:54.883468 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Sep 11 05:13:54.883476 kernel: Initialise system trusted keyrings Sep 11 05:13:54.883489 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 11 05:13:54.883497 kernel: Key type asymmetric registered Sep 11 05:13:54.883506 kernel: Asymmetric key parser 'x509' registered Sep 11 05:13:54.883514 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 11 05:13:54.883523 kernel: io scheduler mq-deadline registered Sep 11 05:13:54.883531 kernel: io scheduler kyber registered Sep 11 05:13:54.883539 kernel: io scheduler bfq registered Sep 11 05:13:54.883550 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 11 05:13:54.883561 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 11 05:13:54.883570 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 11 05:13:54.883578 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 11 05:13:54.883587 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 11 05:13:54.883595 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 11 05:13:54.883604 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 11 05:13:54.883613 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 11 05:13:54.883621 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 11 05:13:54.883629 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 11 05:13:54.883781 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 11 05:13:54.883915 kernel: rtc_cmos 00:04: registered as rtc0 Sep 11 05:13:54.884067 kernel: rtc_cmos 00:04: setting system clock to 2025-09-11T05:13:54 UTC (1757567634) Sep 11 05:13:54.884193 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Sep 11 05:13:54.884204 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 11 05:13:54.884213 kernel: efifb: probing for efifb Sep 11 05:13:54.884221 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Sep 11 05:13:54.884234 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Sep 11 05:13:54.884243 kernel: efifb: scrolling: redraw Sep 11 05:13:54.884251 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 11 05:13:54.884259 kernel: Console: switching to colour frame buffer device 160x50 Sep 11 05:13:54.884268 kernel: fb0: EFI VGA frame buffer device Sep 11 05:13:54.884276 kernel: pstore: Using crash dump compression: deflate Sep 11 05:13:54.884285 kernel: pstore: Registered efi_pstore as persistent store backend Sep 11 05:13:54.884293 kernel: NET: Registered PF_INET6 protocol family Sep 11 05:13:54.884302 kernel: Segment Routing with IPv6 Sep 11 05:13:54.884312 kernel: In-situ OAM (IOAM) with IPv6 Sep 11 05:13:54.884321 kernel: NET: Registered PF_PACKET protocol family Sep 11 05:13:54.884329 kernel: Key type dns_resolver registered Sep 11 05:13:54.884337 kernel: IPI shorthand broadcast: enabled Sep 11 05:13:54.884346 kernel: sched_clock: Marking stable (4241002513, 288136185)->(4579490846, -50352148) Sep 11 05:13:54.884354 kernel: registered taskstats version 1 Sep 11 05:13:54.884362 kernel: Loading compiled-in X.509 certificates Sep 11 05:13:54.884371 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.46-flatcar: b3c5745a53a01317291c111a24ea357b5efeca7e' Sep 11 05:13:54.884379 kernel: Demotion targets for Node 0: null Sep 11 05:13:54.884387 kernel: Key type .fscrypt registered Sep 11 05:13:54.884398 kernel: Key type fscrypt-provisioning registered Sep 11 05:13:54.884406 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 11 05:13:54.884414 kernel: ima: Allocated hash algorithm: sha1 Sep 11 05:13:54.884422 kernel: ima: No architecture policies found Sep 11 05:13:54.884431 kernel: clk: Disabling unused clocks Sep 11 05:13:54.884439 kernel: Warning: unable to open an initial console. Sep 11 05:13:54.884448 kernel: Freeing unused kernel image (initmem) memory: 54052K Sep 11 05:13:54.884456 kernel: Write protecting the kernel read-only data: 24576k Sep 11 05:13:54.884466 kernel: Freeing unused kernel image (rodata/data gap) memory: 252K Sep 11 05:13:54.884475 kernel: Run /init as init process Sep 11 05:13:54.884483 kernel: with arguments: Sep 11 05:13:54.884491 kernel: /init Sep 11 05:13:54.884499 kernel: with environment: Sep 11 05:13:54.884507 kernel: HOME=/ Sep 11 05:13:54.884515 kernel: TERM=linux Sep 11 05:13:54.884523 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 11 05:13:54.884537 systemd[1]: Successfully made /usr/ read-only. Sep 11 05:13:54.884552 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 11 05:13:54.884562 systemd[1]: Detected virtualization kvm. Sep 11 05:13:54.884570 systemd[1]: Detected architecture x86-64. Sep 11 05:13:54.884579 systemd[1]: Running in initrd. Sep 11 05:13:54.884588 systemd[1]: No hostname configured, using default hostname. Sep 11 05:13:54.884596 systemd[1]: Hostname set to . Sep 11 05:13:54.884605 systemd[1]: Initializing machine ID from VM UUID. Sep 11 05:13:54.884616 systemd[1]: Queued start job for default target initrd.target. Sep 11 05:13:54.884624 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 05:13:54.884633 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 05:13:54.884642 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 11 05:13:54.884651 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 11 05:13:54.884660 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 11 05:13:54.884670 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 11 05:13:54.884682 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 11 05:13:54.884691 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 11 05:13:54.884699 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 05:13:54.884708 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 11 05:13:54.884717 systemd[1]: Reached target paths.target - Path Units. Sep 11 05:13:54.884732 systemd[1]: Reached target slices.target - Slice Units. Sep 11 05:13:54.884741 systemd[1]: Reached target swap.target - Swaps. Sep 11 05:13:54.884750 systemd[1]: Reached target timers.target - Timer Units. Sep 11 05:13:54.884758 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 11 05:13:54.884770 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 11 05:13:54.884779 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 11 05:13:54.884787 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 11 05:13:54.884796 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 11 05:13:54.884805 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 11 05:13:54.884814 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 05:13:54.884822 systemd[1]: Reached target sockets.target - Socket Units. Sep 11 05:13:54.884831 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 11 05:13:54.884842 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 11 05:13:54.884851 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 11 05:13:54.884860 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 11 05:13:54.884869 systemd[1]: Starting systemd-fsck-usr.service... Sep 11 05:13:54.884878 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 11 05:13:54.884887 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 11 05:13:54.884896 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:13:54.884905 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 11 05:13:54.884916 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 05:13:54.884926 systemd[1]: Finished systemd-fsck-usr.service. Sep 11 05:13:54.884980 systemd-journald[220]: Collecting audit messages is disabled. Sep 11 05:13:54.885006 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 11 05:13:54.885016 systemd-journald[220]: Journal started Sep 11 05:13:54.885038 systemd-journald[220]: Runtime Journal (/run/log/journal/7f93d0939d374d2d95c3537a12394fab) is 6M, max 48.4M, 42.4M free. Sep 11 05:13:54.885081 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:13:54.878238 systemd-modules-load[223]: Inserted module 'overlay' Sep 11 05:13:54.890108 systemd[1]: Started systemd-journald.service - Journal Service. Sep 11 05:13:54.894183 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 11 05:13:54.896634 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 11 05:13:54.900403 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 11 05:13:54.906245 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 11 05:13:54.917986 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 11 05:13:54.920600 systemd-modules-load[223]: Inserted module 'br_netfilter' Sep 11 05:13:54.922147 kernel: Bridge firewalling registered Sep 11 05:13:54.922405 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 11 05:13:54.924829 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 11 05:13:54.932075 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 05:13:54.933302 systemd-tmpfiles[239]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 11 05:13:54.938797 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 05:13:54.939541 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 11 05:13:54.942872 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 11 05:13:54.953740 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 11 05:13:54.955063 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 11 05:13:54.978855 dracut-cmdline[264]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=51f9391273345312ef1dd27c8e30359e1ffc3b3227e5ed433bbb9ca7d53ecc1c Sep 11 05:13:54.994192 systemd-resolved[259]: Positive Trust Anchors: Sep 11 05:13:54.994226 systemd-resolved[259]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 11 05:13:54.994268 systemd-resolved[259]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 11 05:13:54.997185 systemd-resolved[259]: Defaulting to hostname 'linux'. Sep 11 05:13:54.999014 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 11 05:13:55.004524 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 11 05:13:55.134030 kernel: SCSI subsystem initialized Sep 11 05:13:55.163068 kernel: Loading iSCSI transport class v2.0-870. Sep 11 05:13:55.175976 kernel: iscsi: registered transport (tcp) Sep 11 05:13:55.232285 kernel: iscsi: registered transport (qla4xxx) Sep 11 05:13:55.232365 kernel: QLogic iSCSI HBA Driver Sep 11 05:13:55.260128 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 11 05:13:55.326450 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 05:13:55.327803 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 11 05:13:55.409788 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 11 05:13:55.411602 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 11 05:13:55.505021 kernel: raid6: avx2x4 gen() 27782 MB/s Sep 11 05:13:55.546002 kernel: raid6: avx2x2 gen() 26041 MB/s Sep 11 05:13:55.581979 kernel: raid6: avx2x1 gen() 21645 MB/s Sep 11 05:13:55.582030 kernel: raid6: using algorithm avx2x4 gen() 27782 MB/s Sep 11 05:13:55.606339 kernel: raid6: .... xor() 7181 MB/s, rmw enabled Sep 11 05:13:55.606452 kernel: raid6: using avx2x2 recovery algorithm Sep 11 05:13:55.629017 kernel: xor: automatically using best checksumming function avx Sep 11 05:13:55.802992 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 11 05:13:55.812148 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 11 05:13:55.815290 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 05:13:55.844656 systemd-udevd[473]: Using default interface naming scheme 'v255'. Sep 11 05:13:55.850651 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 05:13:55.853842 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 11 05:13:55.882312 dracut-pre-trigger[480]: rd.md=0: removing MD RAID activation Sep 11 05:13:55.919654 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 11 05:13:55.922482 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 11 05:13:56.009376 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 05:13:56.010902 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 11 05:13:56.064320 kernel: cryptd: max_cpu_qlen set to 1000 Sep 11 05:13:56.064399 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 11 05:13:56.068964 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 11 05:13:56.078015 kernel: AES CTR mode by8 optimization enabled Sep 11 05:13:56.081738 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 11 05:13:56.081782 kernel: GPT:9289727 != 19775487 Sep 11 05:13:56.081796 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 11 05:13:56.081810 kernel: GPT:9289727 != 19775487 Sep 11 05:13:56.081831 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 11 05:13:56.081844 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 05:13:56.088973 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Sep 11 05:13:56.089975 kernel: libata version 3.00 loaded. Sep 11 05:13:56.108350 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 05:13:56.109802 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:13:56.114751 kernel: ahci 0000:00:1f.2: version 3.0 Sep 11 05:13:56.114985 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 11 05:13:56.117092 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:13:56.123994 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Sep 11 05:13:56.124201 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Sep 11 05:13:56.124346 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 11 05:13:56.124737 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:13:56.127204 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 11 05:13:56.139989 kernel: scsi host0: ahci Sep 11 05:13:56.143961 kernel: scsi host1: ahci Sep 11 05:13:56.149059 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 11 05:13:56.155977 kernel: scsi host2: ahci Sep 11 05:13:56.156179 kernel: scsi host3: ahci Sep 11 05:13:56.157006 kernel: scsi host4: ahci Sep 11 05:13:56.161984 kernel: scsi host5: ahci Sep 11 05:13:56.162234 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 1 Sep 11 05:13:56.162247 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 1 Sep 11 05:13:56.162259 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 1 Sep 11 05:13:56.163018 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 1 Sep 11 05:13:56.164064 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 11 05:13:56.166986 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 1 Sep 11 05:13:56.167003 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 1 Sep 11 05:13:56.197208 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 11 05:13:56.198819 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 11 05:13:56.209974 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 11 05:13:56.212811 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 11 05:13:56.214173 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 05:13:56.214244 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:13:56.217228 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:13:56.222268 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:13:56.223863 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 11 05:13:56.250345 disk-uuid[636]: Primary Header is updated. Sep 11 05:13:56.250345 disk-uuid[636]: Secondary Entries is updated. Sep 11 05:13:56.250345 disk-uuid[636]: Secondary Header is updated. Sep 11 05:13:56.255996 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 05:13:56.259793 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:13:56.262646 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 05:13:56.469994 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 11 05:13:56.478416 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 11 05:13:56.478506 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 11 05:13:56.478517 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 11 05:13:56.479991 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 11 05:13:56.480985 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 11 05:13:56.481994 kernel: ata3.00: LPM support broken, forcing max_power Sep 11 05:13:56.482014 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 11 05:13:56.482518 kernel: ata3.00: applying bridge limits Sep 11 05:13:56.483987 kernel: ata3.00: LPM support broken, forcing max_power Sep 11 05:13:56.484011 kernel: ata3.00: configured for UDMA/100 Sep 11 05:13:56.486035 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 11 05:13:56.545040 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 11 05:13:56.545410 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 11 05:13:56.565995 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 11 05:13:56.977764 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 11 05:13:56.979845 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 11 05:13:56.981422 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 05:13:56.984075 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 11 05:13:56.987728 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 11 05:13:57.017335 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 11 05:13:57.316985 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 05:13:57.317818 disk-uuid[639]: The operation has completed successfully. Sep 11 05:13:57.343826 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 11 05:13:57.343975 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 11 05:13:57.389443 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 11 05:13:57.414566 sh[670]: Success Sep 11 05:13:57.433027 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 11 05:13:57.433109 kernel: device-mapper: uevent: version 1.0.3 Sep 11 05:13:57.434115 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 11 05:13:57.444989 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Sep 11 05:13:57.476728 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 11 05:13:57.479034 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 11 05:13:57.496122 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 11 05:13:57.547992 kernel: BTRFS: device fsid f9edc670-4788-4a25-9ed8-a5f7aa85e9c1 devid 1 transid 36 /dev/mapper/usr (253:0) scanned by mount (682) Sep 11 05:13:57.549972 kernel: BTRFS info (device dm-0): first mount of filesystem f9edc670-4788-4a25-9ed8-a5f7aa85e9c1 Sep 11 05:13:57.549999 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 11 05:13:57.555963 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 11 05:13:57.556005 kernel: BTRFS info (device dm-0): enabling free space tree Sep 11 05:13:57.557594 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 11 05:13:57.559202 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 11 05:13:57.560861 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 11 05:13:57.561782 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 11 05:13:57.563760 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 11 05:13:57.590989 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (715) Sep 11 05:13:57.593021 kernel: BTRFS info (device vda6): first mount of filesystem 9a74a431-8bf9-47d2-a780-b26dc4b9d116 Sep 11 05:13:57.593051 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 11 05:13:57.596317 kernel: BTRFS info (device vda6): turning on async discard Sep 11 05:13:57.596355 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 05:13:57.601985 kernel: BTRFS info (device vda6): last unmount of filesystem 9a74a431-8bf9-47d2-a780-b26dc4b9d116 Sep 11 05:13:57.603061 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 11 05:13:57.605043 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 11 05:13:57.740105 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 11 05:13:57.744507 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 11 05:13:57.755678 ignition[756]: Ignition 2.22.0 Sep 11 05:13:57.755692 ignition[756]: Stage: fetch-offline Sep 11 05:13:57.755730 ignition[756]: no configs at "/usr/lib/ignition/base.d" Sep 11 05:13:57.755740 ignition[756]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 05:13:57.755839 ignition[756]: parsed url from cmdline: "" Sep 11 05:13:57.755843 ignition[756]: no config URL provided Sep 11 05:13:57.755848 ignition[756]: reading system config file "/usr/lib/ignition/user.ign" Sep 11 05:13:57.755857 ignition[756]: no config at "/usr/lib/ignition/user.ign" Sep 11 05:13:57.755884 ignition[756]: op(1): [started] loading QEMU firmware config module Sep 11 05:13:57.755889 ignition[756]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 11 05:13:57.765924 ignition[756]: op(1): [finished] loading QEMU firmware config module Sep 11 05:13:57.806192 ignition[756]: parsing config with SHA512: 174a60f277f63dfd125840e1dcafdcd3512662e189928a230cffd5ce84d681349deb8fa4ec36a95f9ca9943727f23b0cd4d14bb92fc432cd493c01b7c75b6f86 Sep 11 05:13:57.807209 systemd-networkd[858]: lo: Link UP Sep 11 05:13:57.807222 systemd-networkd[858]: lo: Gained carrier Sep 11 05:13:57.809211 systemd-networkd[858]: Enumeration completed Sep 11 05:13:57.809618 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 11 05:13:57.810018 systemd-networkd[858]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 05:13:57.810023 systemd-networkd[858]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 11 05:13:57.810578 systemd-networkd[858]: eth0: Link UP Sep 11 05:13:57.812992 systemd-networkd[858]: eth0: Gained carrier Sep 11 05:13:57.813004 systemd-networkd[858]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 05:13:57.815372 systemd[1]: Reached target network.target - Network. Sep 11 05:13:57.820912 ignition[756]: fetch-offline: fetch-offline passed Sep 11 05:13:57.820569 unknown[756]: fetched base config from "system" Sep 11 05:13:57.821005 ignition[756]: Ignition finished successfully Sep 11 05:13:57.820576 unknown[756]: fetched user config from "qemu" Sep 11 05:13:57.830511 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 11 05:13:57.830825 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 11 05:13:57.831018 systemd-networkd[858]: eth0: DHCPv4 address 10.0.0.55/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 11 05:13:57.832058 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 11 05:13:57.898373 ignition[865]: Ignition 2.22.0 Sep 11 05:13:57.898386 ignition[865]: Stage: kargs Sep 11 05:13:57.898586 ignition[865]: no configs at "/usr/lib/ignition/base.d" Sep 11 05:13:57.898598 ignition[865]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 05:13:57.899341 ignition[865]: kargs: kargs passed Sep 11 05:13:57.899390 ignition[865]: Ignition finished successfully Sep 11 05:13:57.907694 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 11 05:13:57.909904 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 11 05:13:58.016381 ignition[873]: Ignition 2.22.0 Sep 11 05:13:58.016396 ignition[873]: Stage: disks Sep 11 05:13:58.016587 ignition[873]: no configs at "/usr/lib/ignition/base.d" Sep 11 05:13:58.016599 ignition[873]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 05:13:58.020935 ignition[873]: disks: disks passed Sep 11 05:13:58.021750 ignition[873]: Ignition finished successfully Sep 11 05:13:58.026036 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 11 05:13:58.027414 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 11 05:13:58.029596 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 11 05:13:58.030918 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 11 05:13:58.032990 systemd[1]: Reached target sysinit.target - System Initialization. Sep 11 05:13:58.035279 systemd[1]: Reached target basic.target - Basic System. Sep 11 05:13:58.038563 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 11 05:13:58.082783 systemd-resolved[259]: Detected conflict on linux IN A 10.0.0.55 Sep 11 05:13:58.082798 systemd-resolved[259]: Hostname conflict, changing published hostname from 'linux' to 'linux9'. Sep 11 05:13:58.086729 systemd-fsck[884]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 11 05:13:58.096857 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 11 05:13:58.099387 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 11 05:13:58.255989 kernel: EXT4-fs (vda9): mounted filesystem 2d35e175-f9a8-40f4-b761-76a7828f45d4 r/w with ordered data mode. Quota mode: none. Sep 11 05:13:58.256689 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 11 05:13:58.258363 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 11 05:13:58.261234 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 11 05:13:58.263227 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 11 05:13:58.264393 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 11 05:13:58.264445 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 11 05:13:58.264470 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 11 05:13:58.280841 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 11 05:13:58.282728 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 11 05:13:58.289077 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (892) Sep 11 05:13:58.289180 kernel: BTRFS info (device vda6): first mount of filesystem 9a74a431-8bf9-47d2-a780-b26dc4b9d116 Sep 11 05:13:58.289203 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 11 05:13:58.293009 kernel: BTRFS info (device vda6): turning on async discard Sep 11 05:13:58.293051 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 05:13:58.296162 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 11 05:13:58.339516 initrd-setup-root[916]: cut: /sysroot/etc/passwd: No such file or directory Sep 11 05:13:58.345435 initrd-setup-root[923]: cut: /sysroot/etc/group: No such file or directory Sep 11 05:13:58.352129 initrd-setup-root[930]: cut: /sysroot/etc/shadow: No such file or directory Sep 11 05:13:58.358530 initrd-setup-root[937]: cut: /sysroot/etc/gshadow: No such file or directory Sep 11 05:13:58.479282 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 11 05:13:58.483411 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 11 05:13:58.486460 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 11 05:13:58.523722 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 11 05:13:58.525327 kernel: BTRFS info (device vda6): last unmount of filesystem 9a74a431-8bf9-47d2-a780-b26dc4b9d116 Sep 11 05:13:58.539166 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 11 05:13:58.569332 ignition[1006]: INFO : Ignition 2.22.0 Sep 11 05:13:58.569332 ignition[1006]: INFO : Stage: mount Sep 11 05:13:58.571244 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 05:13:58.571244 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 05:13:58.571244 ignition[1006]: INFO : mount: mount passed Sep 11 05:13:58.571244 ignition[1006]: INFO : Ignition finished successfully Sep 11 05:13:58.577897 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 11 05:13:58.579124 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 11 05:13:58.601669 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 11 05:13:58.662389 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1018) Sep 11 05:13:58.662445 kernel: BTRFS info (device vda6): first mount of filesystem 9a74a431-8bf9-47d2-a780-b26dc4b9d116 Sep 11 05:13:58.662461 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 11 05:13:58.666997 kernel: BTRFS info (device vda6): turning on async discard Sep 11 05:13:58.667068 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 05:13:58.668729 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 11 05:13:58.709058 ignition[1035]: INFO : Ignition 2.22.0 Sep 11 05:13:58.709058 ignition[1035]: INFO : Stage: files Sep 11 05:13:58.711418 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 05:13:58.711418 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 05:13:58.711418 ignition[1035]: DEBUG : files: compiled without relabeling support, skipping Sep 11 05:13:58.716455 ignition[1035]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 11 05:13:58.716455 ignition[1035]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 11 05:13:58.722597 ignition[1035]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 11 05:13:58.724459 ignition[1035]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 11 05:13:58.726843 unknown[1035]: wrote ssh authorized keys file for user: core Sep 11 05:13:58.728478 ignition[1035]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 11 05:13:58.730239 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 11 05:13:58.732868 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Sep 11 05:13:58.781412 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 11 05:13:59.050015 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 11 05:13:59.050015 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 11 05:13:59.054777 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 11 05:13:59.063223 systemd-networkd[858]: eth0: Gained IPv6LL Sep 11 05:13:59.226295 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 11 05:13:59.228517 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 11 05:13:59.228517 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 05:13:59.297635 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 05:13:59.297635 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 05:13:59.302423 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Sep 11 05:13:59.603825 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 11 05:14:00.511106 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 05:14:00.511106 ignition[1035]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 11 05:14:00.517168 ignition[1035]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 11 05:14:00.520841 ignition[1035]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 11 05:14:00.520841 ignition[1035]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 11 05:14:00.520841 ignition[1035]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 11 05:14:00.526029 ignition[1035]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 11 05:14:00.526029 ignition[1035]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 11 05:14:00.526029 ignition[1035]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 11 05:14:00.526029 ignition[1035]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 11 05:14:00.549871 ignition[1035]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 11 05:14:00.558096 ignition[1035]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 11 05:14:00.559842 ignition[1035]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 11 05:14:00.559842 ignition[1035]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 11 05:14:00.559842 ignition[1035]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 11 05:14:00.559842 ignition[1035]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 11 05:14:00.559842 ignition[1035]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 11 05:14:00.559842 ignition[1035]: INFO : files: files passed Sep 11 05:14:00.559842 ignition[1035]: INFO : Ignition finished successfully Sep 11 05:14:00.566280 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 11 05:14:00.568880 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 11 05:14:00.573915 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 11 05:14:00.596643 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 11 05:14:00.596962 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 11 05:14:00.600177 initrd-setup-root-after-ignition[1064]: grep: /sysroot/oem/oem-release: No such file or directory Sep 11 05:14:00.604513 initrd-setup-root-after-ignition[1066]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 11 05:14:00.604513 initrd-setup-root-after-ignition[1066]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 11 05:14:00.607896 initrd-setup-root-after-ignition[1070]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 11 05:14:00.611172 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 11 05:14:00.612678 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 11 05:14:00.615789 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 11 05:14:00.680982 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 11 05:14:00.681149 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 11 05:14:00.685275 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 11 05:14:00.685397 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 11 05:14:00.689522 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 11 05:14:00.690973 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 11 05:14:00.735770 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 11 05:14:00.737546 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 11 05:14:00.761347 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 11 05:14:00.761527 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 05:14:00.765074 systemd[1]: Stopped target timers.target - Timer Units. Sep 11 05:14:00.767201 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 11 05:14:00.767335 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 11 05:14:00.768496 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 11 05:14:00.768830 systemd[1]: Stopped target basic.target - Basic System. Sep 11 05:14:00.774125 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 11 05:14:00.789255 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 11 05:14:00.789594 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 11 05:14:00.789957 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 11 05:14:00.796658 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 11 05:14:00.797813 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 11 05:14:00.798350 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 11 05:14:00.798721 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 11 05:14:00.799243 systemd[1]: Stopped target swap.target - Swaps. Sep 11 05:14:00.799539 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 11 05:14:00.799670 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 11 05:14:00.810552 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 11 05:14:00.811864 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 05:14:00.812992 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 11 05:14:00.815284 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 05:14:00.817579 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 11 05:14:00.817719 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 11 05:14:00.820974 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 11 05:14:00.821131 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 11 05:14:00.822248 systemd[1]: Stopped target paths.target - Path Units. Sep 11 05:14:00.825217 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 11 05:14:00.830146 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 05:14:00.833241 systemd[1]: Stopped target slices.target - Slice Units. Sep 11 05:14:00.833457 systemd[1]: Stopped target sockets.target - Socket Units. Sep 11 05:14:00.835451 systemd[1]: iscsid.socket: Deactivated successfully. Sep 11 05:14:00.835620 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 11 05:14:00.839630 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 11 05:14:00.839778 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 11 05:14:00.840847 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 11 05:14:00.841053 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 11 05:14:00.843153 systemd[1]: ignition-files.service: Deactivated successfully. Sep 11 05:14:00.843330 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 11 05:14:00.852360 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 11 05:14:00.854532 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 11 05:14:00.854686 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 05:14:00.857370 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 11 05:14:00.861156 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 11 05:14:00.862548 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 05:14:00.865379 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 11 05:14:00.866548 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 11 05:14:00.874935 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 11 05:14:00.919264 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 11 05:14:00.945453 ignition[1090]: INFO : Ignition 2.22.0 Sep 11 05:14:00.945453 ignition[1090]: INFO : Stage: umount Sep 11 05:14:00.955050 ignition[1090]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 05:14:00.955050 ignition[1090]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 05:14:00.955050 ignition[1090]: INFO : umount: umount passed Sep 11 05:14:00.955050 ignition[1090]: INFO : Ignition finished successfully Sep 11 05:14:00.950041 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 11 05:14:00.950994 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 11 05:14:00.951128 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 11 05:14:00.955156 systemd[1]: Stopped target network.target - Network. Sep 11 05:14:00.958383 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 11 05:14:00.958492 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 11 05:14:00.959365 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 11 05:14:00.959444 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 11 05:14:00.962261 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 11 05:14:00.962318 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 11 05:14:00.964626 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 11 05:14:00.964692 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 11 05:14:00.968667 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 11 05:14:00.969781 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 11 05:14:00.982018 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 11 05:14:00.982186 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 11 05:14:00.987046 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 11 05:14:00.987340 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 11 05:14:00.987477 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 11 05:14:00.991663 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 11 05:14:00.992466 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 11 05:14:00.992957 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 11 05:14:00.993014 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 11 05:14:01.000811 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 11 05:14:01.002848 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 11 05:14:01.002932 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 11 05:14:01.003267 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 11 05:14:01.003317 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 11 05:14:01.010371 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 11 05:14:01.010500 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 11 05:14:01.011700 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 11 05:14:01.011754 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 05:14:01.016668 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 05:14:01.021915 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 11 05:14:01.022228 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 11 05:14:01.051335 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 11 05:14:01.051541 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 11 05:14:01.057918 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 11 05:14:01.058145 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 05:14:01.060593 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 11 05:14:01.060654 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 11 05:14:01.061788 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 11 05:14:01.061827 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 05:14:01.062296 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 11 05:14:01.062345 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 11 05:14:01.063180 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 11 05:14:01.063230 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 11 05:14:01.070630 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 11 05:14:01.070695 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 11 05:14:01.073157 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 11 05:14:01.075240 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 11 05:14:01.075294 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 05:14:01.078739 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 11 05:14:01.078787 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 05:14:01.081510 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 05:14:01.081574 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:14:01.087828 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Sep 11 05:14:01.087896 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 11 05:14:01.087973 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 11 05:14:01.102973 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 11 05:14:01.103112 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 11 05:14:01.284259 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 11 05:14:01.284433 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 11 05:14:01.287028 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 11 05:14:01.288293 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 11 05:14:01.288385 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 11 05:14:01.291254 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 11 05:14:01.322993 systemd[1]: Switching root. Sep 11 05:14:01.361271 systemd-journald[220]: Journal stopped Sep 11 05:14:02.812250 systemd-journald[220]: Received SIGTERM from PID 1 (systemd). Sep 11 05:14:02.812324 kernel: SELinux: policy capability network_peer_controls=1 Sep 11 05:14:02.812338 kernel: SELinux: policy capability open_perms=1 Sep 11 05:14:02.812350 kernel: SELinux: policy capability extended_socket_class=1 Sep 11 05:14:02.812365 kernel: SELinux: policy capability always_check_network=0 Sep 11 05:14:02.812377 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 11 05:14:02.812389 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 11 05:14:02.812401 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 11 05:14:02.812415 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 11 05:14:02.812426 kernel: SELinux: policy capability userspace_initial_context=0 Sep 11 05:14:02.812438 kernel: audit: type=1403 audit(1757567641.814:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 11 05:14:02.812458 systemd[1]: Successfully loaded SELinux policy in 70.999ms. Sep 11 05:14:02.812479 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 9.524ms. Sep 11 05:14:02.812493 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 11 05:14:02.812516 systemd[1]: Detected virtualization kvm. Sep 11 05:14:02.812528 systemd[1]: Detected architecture x86-64. Sep 11 05:14:02.812868 systemd[1]: Detected first boot. Sep 11 05:14:02.812884 systemd[1]: Initializing machine ID from VM UUID. Sep 11 05:14:02.812896 zram_generator::config[1139]: No configuration found. Sep 11 05:14:02.812910 kernel: Guest personality initialized and is inactive Sep 11 05:14:02.812922 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 11 05:14:02.812936 kernel: Initialized host personality Sep 11 05:14:02.812974 kernel: NET: Registered PF_VSOCK protocol family Sep 11 05:14:02.812989 systemd[1]: Populated /etc with preset unit settings. Sep 11 05:14:02.813013 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 11 05:14:02.813032 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 11 05:14:02.813047 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 11 05:14:02.813063 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 11 05:14:02.813079 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 11 05:14:02.813095 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 11 05:14:02.813110 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 11 05:14:02.813123 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 11 05:14:02.813135 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 11 05:14:02.813148 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 11 05:14:02.813163 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 11 05:14:02.813176 systemd[1]: Created slice user.slice - User and Session Slice. Sep 11 05:14:02.813188 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 05:14:02.813210 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 05:14:02.813223 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 11 05:14:02.813235 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 11 05:14:02.813248 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 11 05:14:02.813263 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 11 05:14:02.813275 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 11 05:14:02.813288 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 05:14:02.813300 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 11 05:14:02.813313 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 11 05:14:02.813326 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 11 05:14:02.813338 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 11 05:14:02.813356 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 11 05:14:02.813369 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 05:14:02.813383 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 11 05:14:02.813396 systemd[1]: Reached target slices.target - Slice Units. Sep 11 05:14:02.813408 systemd[1]: Reached target swap.target - Swaps. Sep 11 05:14:02.813426 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 11 05:14:02.813439 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 11 05:14:02.813452 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 11 05:14:02.813465 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 11 05:14:02.813477 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 11 05:14:02.813497 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 05:14:02.813519 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 11 05:14:02.813534 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 11 05:14:02.813547 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 11 05:14:02.813559 systemd[1]: Mounting media.mount - External Media Directory... Sep 11 05:14:02.813572 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 05:14:02.813584 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 11 05:14:02.813597 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 11 05:14:02.813610 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 11 05:14:02.813623 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 11 05:14:02.813638 systemd[1]: Reached target machines.target - Containers. Sep 11 05:14:02.813650 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 11 05:14:02.813663 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 05:14:02.813676 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 11 05:14:02.813688 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 11 05:14:02.813700 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 05:14:02.813712 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 11 05:14:02.813725 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 05:14:02.813739 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 11 05:14:02.813752 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 05:14:02.813765 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 11 05:14:02.813785 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 11 05:14:02.813798 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 11 05:14:02.813810 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 11 05:14:02.813823 systemd[1]: Stopped systemd-fsck-usr.service. Sep 11 05:14:02.813836 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 05:14:02.813848 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 11 05:14:02.813864 kernel: fuse: init (API version 7.41) Sep 11 05:14:02.813875 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 11 05:14:02.813889 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 11 05:14:02.813906 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 11 05:14:02.813928 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 11 05:14:02.813969 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 11 05:14:02.813992 systemd[1]: verity-setup.service: Deactivated successfully. Sep 11 05:14:02.814007 systemd[1]: Stopped verity-setup.service. Sep 11 05:14:02.814023 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 05:14:02.814037 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 11 05:14:02.814052 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 11 05:14:02.814064 systemd[1]: Mounted media.mount - External Media Directory. Sep 11 05:14:02.814077 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 11 05:14:02.814089 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 11 05:14:02.814102 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 11 05:14:02.814114 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 05:14:02.814126 kernel: ACPI: bus type drm_connector registered Sep 11 05:14:02.814138 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 11 05:14:02.814151 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 11 05:14:02.814166 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 05:14:02.814178 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 05:14:02.814191 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 11 05:14:02.814203 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 11 05:14:02.814215 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 05:14:02.814227 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 05:14:02.814246 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 11 05:14:02.814258 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 11 05:14:02.814271 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 11 05:14:02.814286 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 11 05:14:02.814299 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 05:14:02.814311 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 11 05:14:02.814324 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 11 05:14:02.814360 systemd-journald[1212]: Collecting audit messages is disabled. Sep 11 05:14:02.814383 kernel: loop: module loaded Sep 11 05:14:02.814398 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 11 05:14:02.814411 systemd-journald[1212]: Journal started Sep 11 05:14:02.814436 systemd-journald[1212]: Runtime Journal (/run/log/journal/7f93d0939d374d2d95c3537a12394fab) is 6M, max 48.4M, 42.4M free. Sep 11 05:14:02.462835 systemd[1]: Queued start job for default target multi-user.target. Sep 11 05:14:02.490753 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 11 05:14:02.491354 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 11 05:14:02.817026 systemd[1]: Started systemd-journald.service - Journal Service. Sep 11 05:14:02.818862 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 05:14:02.819170 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 05:14:02.821049 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 11 05:14:02.822705 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 11 05:14:02.830633 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 11 05:14:02.846580 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 11 05:14:02.848204 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 11 05:14:02.848237 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 11 05:14:02.850713 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 11 05:14:02.854057 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 11 05:14:02.855803 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 05:14:02.858009 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 11 05:14:02.862080 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 11 05:14:02.863610 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 11 05:14:02.868653 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 11 05:14:02.873744 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 11 05:14:02.875359 systemd-journald[1212]: Time spent on flushing to /var/log/journal/7f93d0939d374d2d95c3537a12394fab is 21.724ms for 1074 entries. Sep 11 05:14:02.875359 systemd-journald[1212]: System Journal (/var/log/journal/7f93d0939d374d2d95c3537a12394fab) is 8M, max 195.6M, 187.6M free. Sep 11 05:14:02.922651 systemd-journald[1212]: Received client request to flush runtime journal. Sep 11 05:14:02.877109 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 11 05:14:02.881224 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 11 05:14:02.885747 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 11 05:14:02.897231 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 11 05:14:02.899214 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 11 05:14:02.905349 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 11 05:14:02.932742 kernel: loop0: detected capacity change from 0 to 110984 Sep 11 05:14:02.919731 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 05:14:02.931402 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 11 05:14:02.954053 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 11 05:14:02.957489 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 11 05:14:02.965102 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 11 05:14:02.977594 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 11 05:14:02.981107 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 11 05:14:02.986993 kernel: loop1: detected capacity change from 0 to 128016 Sep 11 05:14:03.020990 kernel: loop2: detected capacity change from 0 to 229808 Sep 11 05:14:03.023823 systemd-tmpfiles[1273]: ACLs are not supported, ignoring. Sep 11 05:14:03.023848 systemd-tmpfiles[1273]: ACLs are not supported, ignoring. Sep 11 05:14:03.031364 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 05:14:03.061121 kernel: loop3: detected capacity change from 0 to 110984 Sep 11 05:14:03.073996 kernel: loop4: detected capacity change from 0 to 128016 Sep 11 05:14:03.090993 kernel: loop5: detected capacity change from 0 to 229808 Sep 11 05:14:03.105347 (sd-merge)[1278]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 11 05:14:03.106240 (sd-merge)[1278]: Merged extensions into '/usr'. Sep 11 05:14:03.114276 systemd[1]: Reload requested from client PID 1258 ('systemd-sysext') (unit systemd-sysext.service)... Sep 11 05:14:03.114306 systemd[1]: Reloading... Sep 11 05:14:03.189985 zram_generator::config[1303]: No configuration found. Sep 11 05:14:03.499163 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 11 05:14:03.499915 systemd[1]: Reloading finished in 384 ms. Sep 11 05:14:03.527318 ldconfig[1253]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 11 05:14:03.533681 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 11 05:14:03.554930 systemd[1]: Starting ensure-sysext.service... Sep 11 05:14:03.557010 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 11 05:14:03.580532 systemd[1]: Reload requested from client PID 1340 ('systemctl') (unit ensure-sysext.service)... Sep 11 05:14:03.580552 systemd[1]: Reloading... Sep 11 05:14:03.600633 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 11 05:14:03.600680 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 11 05:14:03.601066 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 11 05:14:03.601446 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 11 05:14:03.602941 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 11 05:14:03.603411 systemd-tmpfiles[1341]: ACLs are not supported, ignoring. Sep 11 05:14:03.603636 systemd-tmpfiles[1341]: ACLs are not supported, ignoring. Sep 11 05:14:03.614166 systemd-tmpfiles[1341]: Detected autofs mount point /boot during canonicalization of boot. Sep 11 05:14:03.614182 systemd-tmpfiles[1341]: Skipping /boot Sep 11 05:14:03.639758 systemd-tmpfiles[1341]: Detected autofs mount point /boot during canonicalization of boot. Sep 11 05:14:03.640166 systemd-tmpfiles[1341]: Skipping /boot Sep 11 05:14:03.646703 zram_generator::config[1375]: No configuration found. Sep 11 05:14:03.840621 systemd[1]: Reloading finished in 259 ms. Sep 11 05:14:03.862799 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 11 05:14:03.891861 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 05:14:03.905566 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 11 05:14:03.910229 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 11 05:14:03.917380 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 11 05:14:03.924028 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 11 05:14:03.929234 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 11 05:14:03.936236 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 05:14:03.936424 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 05:14:03.941578 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 05:14:03.945589 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 05:14:03.948319 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 05:14:03.951138 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 05:14:03.951254 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 05:14:03.951349 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 05:14:03.964783 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 11 05:14:03.967009 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 11 05:14:03.969091 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 05:14:03.969378 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 05:14:03.971650 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 05:14:03.971873 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 05:14:03.984963 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 11 05:14:03.987557 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 05:14:03.988295 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 05:14:03.997845 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 05:14:03.998216 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 05:14:04.000104 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 05:14:04.002366 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 11 05:14:04.004458 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 05:14:04.026212 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 05:14:04.027416 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 05:14:04.027538 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 05:14:04.027676 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 05:14:04.035041 systemd[1]: Finished ensure-sysext.service. Sep 11 05:14:04.039769 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 11 05:14:04.041481 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 05:14:04.041710 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 05:14:04.043356 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 05:14:04.043583 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 05:14:04.047183 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 05:14:04.047690 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 05:14:04.049449 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 11 05:14:04.050015 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 11 05:14:04.055597 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 11 05:14:04.056083 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 11 05:14:04.056932 augenrules[1452]: No rules Sep 11 05:14:04.060567 systemd[1]: audit-rules.service: Deactivated successfully. Sep 11 05:14:04.061749 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 11 05:14:04.063869 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 11 05:14:04.068524 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 05:14:04.071001 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 11 05:14:04.072478 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 11 05:14:04.073429 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 11 05:14:04.074039 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 11 05:14:04.093123 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 11 05:14:04.112409 systemd-udevd[1459]: Using default interface naming scheme 'v255'. Sep 11 05:14:04.133299 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 05:14:04.138045 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 11 05:14:04.165554 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 11 05:14:04.167229 systemd[1]: Reached target time-set.target - System Time Set. Sep 11 05:14:04.172382 systemd-resolved[1411]: Positive Trust Anchors: Sep 11 05:14:04.172756 systemd-resolved[1411]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 11 05:14:04.172801 systemd-resolved[1411]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 11 05:14:04.181568 systemd-resolved[1411]: Defaulting to hostname 'linux'. Sep 11 05:14:04.185286 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 11 05:14:04.188617 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 11 05:14:04.190089 systemd[1]: Reached target sysinit.target - System Initialization. Sep 11 05:14:04.191748 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 11 05:14:04.193452 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 11 05:14:04.195235 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Sep 11 05:14:04.196903 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 11 05:14:04.198270 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 11 05:14:04.199920 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 11 05:14:04.201386 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 11 05:14:04.201428 systemd[1]: Reached target paths.target - Path Units. Sep 11 05:14:04.203021 systemd[1]: Reached target timers.target - Timer Units. Sep 11 05:14:04.204928 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 11 05:14:04.208862 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 11 05:14:04.214272 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 11 05:14:04.216301 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 11 05:14:04.217922 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 11 05:14:04.222702 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 11 05:14:04.224720 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 11 05:14:04.229023 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 11 05:14:04.236093 systemd[1]: Reached target sockets.target - Socket Units. Sep 11 05:14:04.237908 systemd[1]: Reached target basic.target - Basic System. Sep 11 05:14:04.239351 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 11 05:14:04.239395 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 11 05:14:04.272993 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 11 05:14:04.277339 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 11 05:14:04.312405 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 11 05:14:04.318244 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 11 05:14:04.319552 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 11 05:14:04.323987 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Sep 11 05:14:04.329030 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 11 05:14:04.331740 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 11 05:14:04.336320 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 11 05:14:04.339222 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 11 05:14:04.348450 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 11 05:14:04.350878 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 11 05:14:04.351468 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 11 05:14:04.352218 systemd[1]: Starting update-engine.service - Update Engine... Sep 11 05:14:04.360109 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 11 05:14:04.360338 google_oslogin_nss_cache[1509]: oslogin_cache_refresh[1509]: Refreshing passwd entry cache Sep 11 05:14:04.360753 oslogin_cache_refresh[1509]: Refreshing passwd entry cache Sep 11 05:14:04.367360 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 11 05:14:04.368137 google_oslogin_nss_cache[1509]: oslogin_cache_refresh[1509]: Failure getting users, quitting Sep 11 05:14:04.368691 oslogin_cache_refresh[1509]: Failure getting users, quitting Sep 11 05:14:04.369036 google_oslogin_nss_cache[1509]: oslogin_cache_refresh[1509]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 11 05:14:04.369077 oslogin_cache_refresh[1509]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 11 05:14:04.369168 google_oslogin_nss_cache[1509]: oslogin_cache_refresh[1509]: Refreshing group entry cache Sep 11 05:14:04.369327 oslogin_cache_refresh[1509]: Refreshing group entry cache Sep 11 05:14:04.369884 google_oslogin_nss_cache[1509]: oslogin_cache_refresh[1509]: Failure getting groups, quitting Sep 11 05:14:04.370985 oslogin_cache_refresh[1509]: Failure getting groups, quitting Sep 11 05:14:04.371050 google_oslogin_nss_cache[1509]: oslogin_cache_refresh[1509]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 11 05:14:04.371085 oslogin_cache_refresh[1509]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 11 05:14:04.372559 jq[1519]: true Sep 11 05:14:04.373515 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 11 05:14:04.373824 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 11 05:14:04.378368 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Sep 11 05:14:04.378808 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Sep 11 05:14:04.389213 jq[1506]: false Sep 11 05:14:04.391354 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 11 05:14:04.395938 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 11 05:14:04.400593 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 11 05:14:04.404648 jq[1523]: true Sep 11 05:14:04.405507 update_engine[1518]: I20250911 05:14:04.405426 1518 main.cc:92] Flatcar Update Engine starting Sep 11 05:14:04.427364 tar[1521]: linux-amd64/LICENSE Sep 11 05:14:04.427641 tar[1521]: linux-amd64/helm Sep 11 05:14:04.430709 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 11 05:14:04.436229 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 11 05:14:04.448974 extend-filesystems[1508]: Found /dev/vda6 Sep 11 05:14:04.452167 extend-filesystems[1508]: Found /dev/vda9 Sep 11 05:14:04.457495 extend-filesystems[1508]: Checking size of /dev/vda9 Sep 11 05:14:04.477635 extend-filesystems[1508]: Resized partition /dev/vda9 Sep 11 05:14:04.482923 extend-filesystems[1565]: resize2fs 1.47.3 (8-Jul-2025) Sep 11 05:14:04.486983 kernel: mousedev: PS/2 mouse device common for all mice Sep 11 05:14:04.495086 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 11 05:14:04.508967 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 11 05:14:04.547289 kernel: ACPI: button: Power Button [PWRF] Sep 11 05:14:04.547326 update_engine[1518]: I20250911 05:14:04.517755 1518 update_check_scheduler.cc:74] Next update check in 11m49s Sep 11 05:14:04.513314 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 11 05:14:04.513059 dbus-daemon[1502]: [system] SELinux support is enabled Sep 11 05:14:04.518774 systemd-networkd[1477]: lo: Link UP Sep 11 05:14:04.518780 systemd-networkd[1477]: lo: Gained carrier Sep 11 05:14:04.518906 systemd[1]: motdgen.service: Deactivated successfully. Sep 11 05:14:04.519345 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 11 05:14:04.529083 systemd-networkd[1477]: Enumeration completed Sep 11 05:14:04.530400 systemd-networkd[1477]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 05:14:04.530404 systemd-networkd[1477]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 11 05:14:04.531033 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 11 05:14:04.531197 systemd-networkd[1477]: eth0: Link UP Sep 11 05:14:04.531510 systemd-networkd[1477]: eth0: Gained carrier Sep 11 05:14:04.531525 systemd-networkd[1477]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 05:14:04.533770 systemd[1]: Reached target network.target - Network. Sep 11 05:14:04.537192 systemd[1]: Starting containerd.service - containerd container runtime... Sep 11 05:14:04.537371 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 11 05:14:04.537396 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 11 05:14:04.540669 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 11 05:14:04.546123 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 11 05:14:04.547121 systemd-networkd[1477]: eth0: DHCPv4 address 10.0.0.55/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 11 05:14:04.550494 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 11 05:14:04.550526 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 11 05:14:04.553567 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 11 05:14:04.561645 systemd-timesyncd[1446]: Network configuration changed, trying to establish connection. Sep 11 05:14:05.874958 systemd-resolved[1411]: Clock change detected. Flushing caches. Sep 11 05:14:05.875109 systemd-timesyncd[1446]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 11 05:14:05.875191 systemd-timesyncd[1446]: Initial clock synchronization to Thu 2025-09-11 05:14:05.874796 UTC. Sep 11 05:14:05.878246 systemd[1]: Started update-engine.service - Update Engine. Sep 11 05:14:05.887182 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 11 05:14:05.936588 (ntainerd)[1579]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 11 05:14:06.037785 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 11 05:14:06.054642 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 11 05:14:06.054001 systemd-logind[1516]: New seat seat0. Sep 11 05:14:06.055666 systemd[1]: Started systemd-logind.service - User Login Management. Sep 11 05:14:06.070592 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:14:06.086000 extend-filesystems[1565]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 11 05:14:06.086000 extend-filesystems[1565]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 11 05:14:06.086000 extend-filesystems[1565]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 11 05:14:06.090488 extend-filesystems[1508]: Resized filesystem in /dev/vda9 Sep 11 05:14:06.089592 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 11 05:14:06.091511 bash[1568]: Updated "/home/core/.ssh/authorized_keys" Sep 11 05:14:06.092322 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 11 05:14:06.103122 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 11 05:14:06.134663 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 11 05:14:06.223741 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Sep 11 05:14:06.224281 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 11 05:14:06.224461 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 11 05:14:06.225375 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 05:14:06.231297 sshd_keygen[1527]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 11 05:14:06.225803 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:14:06.229446 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 05:14:06.248253 locksmithd[1575]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 11 05:14:06.252026 systemd-logind[1516]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 11 05:14:06.303428 systemd-logind[1516]: Watching system buttons on /dev/input/event2 (Power Button) Sep 11 05:14:06.319322 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 11 05:14:06.325035 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 11 05:14:06.352556 systemd[1]: issuegen.service: Deactivated successfully. Sep 11 05:14:06.352841 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 11 05:14:06.356212 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 11 05:14:06.396026 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 11 05:14:06.430133 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 05:14:06.436896 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 11 05:14:06.443197 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 11 05:14:06.444795 systemd[1]: Reached target getty.target - Login Prompts. Sep 11 05:14:06.450769 kernel: kvm_amd: TSC scaling supported Sep 11 05:14:06.450813 kernel: kvm_amd: Nested Virtualization enabled Sep 11 05:14:06.450827 kernel: kvm_amd: Nested Paging enabled Sep 11 05:14:06.450839 kernel: kvm_amd: LBR virtualization supported Sep 11 05:14:06.452177 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 11 05:14:06.452205 kernel: kvm_amd: Virtual GIF supported Sep 11 05:14:06.511950 kernel: EDAC MC: Ver: 3.0.0 Sep 11 05:14:06.590384 containerd[1579]: time="2025-09-11T05:14:06Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 11 05:14:06.591939 containerd[1579]: time="2025-09-11T05:14:06.591856826Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 11 05:14:06.609229 containerd[1579]: time="2025-09-11T05:14:06.609147909Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="15.479µs" Sep 11 05:14:06.609229 containerd[1579]: time="2025-09-11T05:14:06.609207661Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 11 05:14:06.609229 containerd[1579]: time="2025-09-11T05:14:06.609244210Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 11 05:14:06.609565 containerd[1579]: time="2025-09-11T05:14:06.609528132Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 11 05:14:06.609565 containerd[1579]: time="2025-09-11T05:14:06.609557888Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 11 05:14:06.609614 containerd[1579]: time="2025-09-11T05:14:06.609592072Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 11 05:14:06.609710 containerd[1579]: time="2025-09-11T05:14:06.609677623Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 11 05:14:06.609710 containerd[1579]: time="2025-09-11T05:14:06.609698722Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 11 05:14:06.610958 containerd[1579]: time="2025-09-11T05:14:06.610299068Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 11 05:14:06.610958 containerd[1579]: time="2025-09-11T05:14:06.610390871Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 11 05:14:06.610958 containerd[1579]: time="2025-09-11T05:14:06.610445082Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 11 05:14:06.610958 containerd[1579]: time="2025-09-11T05:14:06.610535873Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 11 05:14:06.610958 containerd[1579]: time="2025-09-11T05:14:06.610864840Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 11 05:14:06.611563 containerd[1579]: time="2025-09-11T05:14:06.611525880Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 11 05:14:06.611643 containerd[1579]: time="2025-09-11T05:14:06.611606200Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 11 05:14:06.611683 containerd[1579]: time="2025-09-11T05:14:06.611647077Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 11 05:14:06.611731 containerd[1579]: time="2025-09-11T05:14:06.611703643Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 11 05:14:06.669420 tar[1521]: linux-amd64/README.md Sep 11 05:14:06.682113 containerd[1579]: time="2025-09-11T05:14:06.682028442Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 11 05:14:06.682272 containerd[1579]: time="2025-09-11T05:14:06.682251210Z" level=info msg="metadata content store policy set" policy=shared Sep 11 05:14:06.689111 containerd[1579]: time="2025-09-11T05:14:06.689042659Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 11 05:14:06.689224 containerd[1579]: time="2025-09-11T05:14:06.689145572Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 11 05:14:06.689224 containerd[1579]: time="2025-09-11T05:14:06.689171791Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 11 05:14:06.689224 containerd[1579]: time="2025-09-11T05:14:06.689193692Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 11 05:14:06.689224 containerd[1579]: time="2025-09-11T05:14:06.689212087Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689227926Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689250048Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689281797Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689300262Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689311964Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689330298Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 11 05:14:06.689461 containerd[1579]: time="2025-09-11T05:14:06.689358652Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689590096Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689631844Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689653825Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689672200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689705141Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689723736Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689748342Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 11 05:14:06.689792 containerd[1579]: time="2025-09-11T05:14:06.689774051Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 11 05:14:06.690046 containerd[1579]: time="2025-09-11T05:14:06.689801252Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 11 05:14:06.690046 containerd[1579]: time="2025-09-11T05:14:06.689816781Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 11 05:14:06.690046 containerd[1579]: time="2025-09-11T05:14:06.689840295Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 11 05:14:06.690046 containerd[1579]: time="2025-09-11T05:14:06.689978604Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 11 05:14:06.690046 containerd[1579]: time="2025-09-11T05:14:06.690011145Z" level=info msg="Start snapshots syncer" Sep 11 05:14:06.690186 containerd[1579]: time="2025-09-11T05:14:06.690054867Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 11 05:14:06.690574 containerd[1579]: time="2025-09-11T05:14:06.690507987Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 11 05:14:06.690708 containerd[1579]: time="2025-09-11T05:14:06.690602194Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 11 05:14:06.690742 containerd[1579]: time="2025-09-11T05:14:06.690713122Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 11 05:14:06.690884 containerd[1579]: time="2025-09-11T05:14:06.690845531Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 11 05:14:06.690884 containerd[1579]: time="2025-09-11T05:14:06.690881528Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 11 05:14:06.690986 containerd[1579]: time="2025-09-11T05:14:06.690907457Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 11 05:14:06.690986 containerd[1579]: time="2025-09-11T05:14:06.690941971Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 11 05:14:06.690986 containerd[1579]: time="2025-09-11T05:14:06.690963442Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 11 05:14:06.690986 containerd[1579]: time="2025-09-11T05:14:06.690976396Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 11 05:14:06.691074 containerd[1579]: time="2025-09-11T05:14:06.690991875Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 11 05:14:06.691074 containerd[1579]: time="2025-09-11T05:14:06.691028814Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 11 05:14:06.691074 containerd[1579]: time="2025-09-11T05:14:06.691043792Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 11 05:14:06.691074 containerd[1579]: time="2025-09-11T05:14:06.691056045Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 11 05:14:06.691172 containerd[1579]: time="2025-09-11T05:14:06.691107312Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 11 05:14:06.691172 containerd[1579]: time="2025-09-11T05:14:06.691126558Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 11 05:14:06.691172 containerd[1579]: time="2025-09-11T05:14:06.691135735Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 11 05:14:06.691172 containerd[1579]: time="2025-09-11T05:14:06.691144531Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 11 05:14:06.691172 containerd[1579]: time="2025-09-11T05:14:06.691152065Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 11 05:14:06.691172 containerd[1579]: time="2025-09-11T05:14:06.691163377Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 11 05:14:06.691363 containerd[1579]: time="2025-09-11T05:14:06.691182082Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 11 05:14:06.691363 containerd[1579]: time="2025-09-11T05:14:06.691252514Z" level=info msg="runtime interface created" Sep 11 05:14:06.691363 containerd[1579]: time="2025-09-11T05:14:06.691261621Z" level=info msg="created NRI interface" Sep 11 05:14:06.691363 containerd[1579]: time="2025-09-11T05:14:06.691297869Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 11 05:14:06.691363 containerd[1579]: time="2025-09-11T05:14:06.691317065Z" level=info msg="Connect containerd service" Sep 11 05:14:06.691363 containerd[1579]: time="2025-09-11T05:14:06.691362731Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 11 05:14:06.692580 containerd[1579]: time="2025-09-11T05:14:06.692539438Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 11 05:14:06.695289 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 11 05:14:06.851230 containerd[1579]: time="2025-09-11T05:14:06.851089202Z" level=info msg="Start subscribing containerd event" Sep 11 05:14:06.851230 containerd[1579]: time="2025-09-11T05:14:06.851194279Z" level=info msg="Start recovering state" Sep 11 05:14:06.851387 containerd[1579]: time="2025-09-11T05:14:06.851368576Z" level=info msg="Start event monitor" Sep 11 05:14:06.851436 containerd[1579]: time="2025-09-11T05:14:06.851422317Z" level=info msg="Start cni network conf syncer for default" Sep 11 05:14:06.851460 containerd[1579]: time="2025-09-11T05:14:06.851440250Z" level=info msg="Start streaming server" Sep 11 05:14:06.851564 containerd[1579]: time="2025-09-11T05:14:06.851537372Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 11 05:14:06.851564 containerd[1579]: time="2025-09-11T05:14:06.851557610Z" level=info msg="runtime interface starting up..." Sep 11 05:14:06.851605 containerd[1579]: time="2025-09-11T05:14:06.851584481Z" level=info msg="starting plugins..." Sep 11 05:14:06.851626 containerd[1579]: time="2025-09-11T05:14:06.851612423Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 11 05:14:06.851732 containerd[1579]: time="2025-09-11T05:14:06.851668258Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 11 05:14:06.852054 containerd[1579]: time="2025-09-11T05:14:06.851807169Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 11 05:14:06.852224 containerd[1579]: time="2025-09-11T05:14:06.852182543Z" level=info msg="containerd successfully booted in 0.262593s" Sep 11 05:14:06.852336 systemd[1]: Started containerd.service - containerd container runtime. Sep 11 05:14:07.223296 systemd-networkd[1477]: eth0: Gained IPv6LL Sep 11 05:14:07.227078 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 11 05:14:07.229305 systemd[1]: Reached target network-online.target - Network is Online. Sep 11 05:14:07.232625 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 11 05:14:07.235486 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:07.238527 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 11 05:14:07.269373 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 11 05:14:07.291441 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 11 05:14:07.291792 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 11 05:14:07.293570 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 11 05:14:08.861495 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:08.863243 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 11 05:14:08.864576 systemd[1]: Startup finished in 4.303s (kernel) + 7.160s (initrd) + 5.806s (userspace) = 17.269s. Sep 11 05:14:08.882477 (kubelet)[1678]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 05:14:09.496649 kubelet[1678]: E0911 05:14:09.496560 1678 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 05:14:09.501435 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 05:14:09.501698 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 05:14:09.502283 systemd[1]: kubelet.service: Consumed 1.993s CPU time, 268.6M memory peak. Sep 11 05:14:10.066605 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 11 05:14:10.068422 systemd[1]: Started sshd@0-10.0.0.55:22-10.0.0.1:49896.service - OpenSSH per-connection server daemon (10.0.0.1:49896). Sep 11 05:14:10.153815 sshd[1691]: Accepted publickey for core from 10.0.0.1 port 49896 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:10.156228 sshd-session[1691]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:10.163734 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 11 05:14:10.165018 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 11 05:14:10.173239 systemd-logind[1516]: New session 1 of user core. Sep 11 05:14:10.212466 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 11 05:14:10.216878 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 11 05:14:10.246258 (systemd)[1696]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 11 05:14:10.250100 systemd-logind[1516]: New session c1 of user core. Sep 11 05:14:10.453113 systemd[1696]: Queued start job for default target default.target. Sep 11 05:14:10.465687 systemd[1696]: Created slice app.slice - User Application Slice. Sep 11 05:14:10.465727 systemd[1696]: Reached target paths.target - Paths. Sep 11 05:14:10.465788 systemd[1696]: Reached target timers.target - Timers. Sep 11 05:14:10.467989 systemd[1696]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 11 05:14:10.483704 systemd[1696]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 11 05:14:10.483898 systemd[1696]: Reached target sockets.target - Sockets. Sep 11 05:14:10.484006 systemd[1696]: Reached target basic.target - Basic System. Sep 11 05:14:10.484070 systemd[1696]: Reached target default.target - Main User Target. Sep 11 05:14:10.484123 systemd[1696]: Startup finished in 224ms. Sep 11 05:14:10.484489 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 11 05:14:10.487098 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 11 05:14:10.554289 systemd[1]: Started sshd@1-10.0.0.55:22-10.0.0.1:49902.service - OpenSSH per-connection server daemon (10.0.0.1:49902). Sep 11 05:14:10.614727 sshd[1707]: Accepted publickey for core from 10.0.0.1 port 49902 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:10.616130 sshd-session[1707]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:10.620896 systemd-logind[1516]: New session 2 of user core. Sep 11 05:14:10.631173 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 11 05:14:10.686948 sshd[1710]: Connection closed by 10.0.0.1 port 49902 Sep 11 05:14:10.687361 sshd-session[1707]: pam_unix(sshd:session): session closed for user core Sep 11 05:14:10.706339 systemd[1]: sshd@1-10.0.0.55:22-10.0.0.1:49902.service: Deactivated successfully. Sep 11 05:14:10.708673 systemd[1]: session-2.scope: Deactivated successfully. Sep 11 05:14:10.709490 systemd-logind[1516]: Session 2 logged out. Waiting for processes to exit. Sep 11 05:14:10.712647 systemd[1]: Started sshd@2-10.0.0.55:22-10.0.0.1:49904.service - OpenSSH per-connection server daemon (10.0.0.1:49904). Sep 11 05:14:10.713312 systemd-logind[1516]: Removed session 2. Sep 11 05:14:10.775747 sshd[1716]: Accepted publickey for core from 10.0.0.1 port 49904 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:10.777594 sshd-session[1716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:10.783162 systemd-logind[1516]: New session 3 of user core. Sep 11 05:14:10.793112 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 11 05:14:10.843767 sshd[1719]: Connection closed by 10.0.0.1 port 49904 Sep 11 05:14:10.844316 sshd-session[1716]: pam_unix(sshd:session): session closed for user core Sep 11 05:14:10.858764 systemd[1]: sshd@2-10.0.0.55:22-10.0.0.1:49904.service: Deactivated successfully. Sep 11 05:14:10.861333 systemd[1]: session-3.scope: Deactivated successfully. Sep 11 05:14:10.863228 systemd-logind[1516]: Session 3 logged out. Waiting for processes to exit. Sep 11 05:14:10.868413 systemd[1]: Started sshd@3-10.0.0.55:22-10.0.0.1:49912.service - OpenSSH per-connection server daemon (10.0.0.1:49912). Sep 11 05:14:10.869367 systemd-logind[1516]: Removed session 3. Sep 11 05:14:10.929804 sshd[1725]: Accepted publickey for core from 10.0.0.1 port 49912 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:10.931266 sshd-session[1725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:10.936561 systemd-logind[1516]: New session 4 of user core. Sep 11 05:14:10.952199 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 11 05:14:11.008577 sshd[1728]: Connection closed by 10.0.0.1 port 49912 Sep 11 05:14:11.009112 sshd-session[1725]: pam_unix(sshd:session): session closed for user core Sep 11 05:14:11.023834 systemd[1]: sshd@3-10.0.0.55:22-10.0.0.1:49912.service: Deactivated successfully. Sep 11 05:14:11.026023 systemd[1]: session-4.scope: Deactivated successfully. Sep 11 05:14:11.026748 systemd-logind[1516]: Session 4 logged out. Waiting for processes to exit. Sep 11 05:14:11.029962 systemd[1]: Started sshd@4-10.0.0.55:22-10.0.0.1:49928.service - OpenSSH per-connection server daemon (10.0.0.1:49928). Sep 11 05:14:11.030704 systemd-logind[1516]: Removed session 4. Sep 11 05:14:11.088705 sshd[1734]: Accepted publickey for core from 10.0.0.1 port 49928 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:11.090195 sshd-session[1734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:11.094817 systemd-logind[1516]: New session 5 of user core. Sep 11 05:14:11.105061 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 11 05:14:11.165193 sudo[1738]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 11 05:14:11.165614 sudo[1738]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 05:14:11.186401 sudo[1738]: pam_unix(sudo:session): session closed for user root Sep 11 05:14:11.187980 sshd[1737]: Connection closed by 10.0.0.1 port 49928 Sep 11 05:14:11.188388 sshd-session[1734]: pam_unix(sshd:session): session closed for user core Sep 11 05:14:11.197694 systemd[1]: sshd@4-10.0.0.55:22-10.0.0.1:49928.service: Deactivated successfully. Sep 11 05:14:11.199961 systemd[1]: session-5.scope: Deactivated successfully. Sep 11 05:14:11.200809 systemd-logind[1516]: Session 5 logged out. Waiting for processes to exit. Sep 11 05:14:11.203956 systemd[1]: Started sshd@5-10.0.0.55:22-10.0.0.1:49934.service - OpenSSH per-connection server daemon (10.0.0.1:49934). Sep 11 05:14:11.204590 systemd-logind[1516]: Removed session 5. Sep 11 05:14:11.268407 sshd[1744]: Accepted publickey for core from 10.0.0.1 port 49934 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:11.270098 sshd-session[1744]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:11.274638 systemd-logind[1516]: New session 6 of user core. Sep 11 05:14:11.285069 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 11 05:14:11.338778 sudo[1749]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 11 05:14:11.339137 sudo[1749]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 05:14:11.991316 sudo[1749]: pam_unix(sudo:session): session closed for user root Sep 11 05:14:11.999388 sudo[1748]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 11 05:14:11.999740 sudo[1748]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 05:14:12.013006 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 11 05:14:12.064481 augenrules[1771]: No rules Sep 11 05:14:12.066544 systemd[1]: audit-rules.service: Deactivated successfully. Sep 11 05:14:12.066875 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 11 05:14:12.068247 sudo[1748]: pam_unix(sudo:session): session closed for user root Sep 11 05:14:12.070067 sshd[1747]: Connection closed by 10.0.0.1 port 49934 Sep 11 05:14:12.070549 sshd-session[1744]: pam_unix(sshd:session): session closed for user core Sep 11 05:14:12.088800 systemd[1]: sshd@5-10.0.0.55:22-10.0.0.1:49934.service: Deactivated successfully. Sep 11 05:14:12.091076 systemd[1]: session-6.scope: Deactivated successfully. Sep 11 05:14:12.092136 systemd-logind[1516]: Session 6 logged out. Waiting for processes to exit. Sep 11 05:14:12.095388 systemd[1]: Started sshd@6-10.0.0.55:22-10.0.0.1:49964.service - OpenSSH per-connection server daemon (10.0.0.1:49964). Sep 11 05:14:12.096142 systemd-logind[1516]: Removed session 6. Sep 11 05:14:12.162711 sshd[1780]: Accepted publickey for core from 10.0.0.1 port 49964 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:14:12.164515 sshd-session[1780]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:14:12.169598 systemd-logind[1516]: New session 7 of user core. Sep 11 05:14:12.183085 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 11 05:14:12.238671 sudo[1784]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 11 05:14:12.239011 sudo[1784]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 05:14:12.879085 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 11 05:14:12.906533 (dockerd)[1805]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 11 05:14:13.449659 dockerd[1805]: time="2025-09-11T05:14:13.449585054Z" level=info msg="Starting up" Sep 11 05:14:13.450755 dockerd[1805]: time="2025-09-11T05:14:13.450678846Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 11 05:14:13.495337 dockerd[1805]: time="2025-09-11T05:14:13.495263534Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 11 05:14:14.260377 dockerd[1805]: time="2025-09-11T05:14:14.260320071Z" level=info msg="Loading containers: start." Sep 11 05:14:14.272983 kernel: Initializing XFRM netlink socket Sep 11 05:14:14.700047 systemd-networkd[1477]: docker0: Link UP Sep 11 05:14:15.014272 dockerd[1805]: time="2025-09-11T05:14:15.014067626Z" level=info msg="Loading containers: done." Sep 11 05:14:15.032572 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2347510931-merged.mount: Deactivated successfully. Sep 11 05:14:15.036358 dockerd[1805]: time="2025-09-11T05:14:15.036265162Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 11 05:14:15.036571 dockerd[1805]: time="2025-09-11T05:14:15.036538525Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 11 05:14:15.036759 dockerd[1805]: time="2025-09-11T05:14:15.036726969Z" level=info msg="Initializing buildkit" Sep 11 05:14:15.082071 dockerd[1805]: time="2025-09-11T05:14:15.081999006Z" level=info msg="Completed buildkit initialization" Sep 11 05:14:15.090405 dockerd[1805]: time="2025-09-11T05:14:15.090334842Z" level=info msg="Daemon has completed initialization" Sep 11 05:14:15.090609 dockerd[1805]: time="2025-09-11T05:14:15.090453654Z" level=info msg="API listen on /run/docker.sock" Sep 11 05:14:15.090736 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 11 05:14:16.151347 containerd[1579]: time="2025-09-11T05:14:16.151268279Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\"" Sep 11 05:14:16.863507 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount422176890.mount: Deactivated successfully. Sep 11 05:14:18.965749 containerd[1579]: time="2025-09-11T05:14:18.965659603Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:18.967133 containerd[1579]: time="2025-09-11T05:14:18.967081230Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.5: active requests=0, bytes read=30114893" Sep 11 05:14:18.968466 containerd[1579]: time="2025-09-11T05:14:18.968410523Z" level=info msg="ImageCreate event name:\"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:18.971293 containerd[1579]: time="2025-09-11T05:14:18.971248276Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:18.972278 containerd[1579]: time="2025-09-11T05:14:18.972245577Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.5\" with image id \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\", size \"30111492\" in 2.82089896s" Sep 11 05:14:18.972334 containerd[1579]: time="2025-09-11T05:14:18.972303766Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\" returns image reference \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\"" Sep 11 05:14:18.975556 containerd[1579]: time="2025-09-11T05:14:18.975508747Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\"" Sep 11 05:14:19.752816 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 11 05:14:19.755273 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:20.102794 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:20.107814 (kubelet)[2090]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 05:14:20.752637 kubelet[2090]: E0911 05:14:20.752541 2090 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 05:14:20.760731 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 05:14:20.761387 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 05:14:20.762177 systemd[1]: kubelet.service: Consumed 349ms CPU time, 110.7M memory peak. Sep 11 05:14:21.312104 containerd[1579]: time="2025-09-11T05:14:21.312023331Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:21.313151 containerd[1579]: time="2025-09-11T05:14:21.313027524Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.5: active requests=0, bytes read=26020844" Sep 11 05:14:21.314580 containerd[1579]: time="2025-09-11T05:14:21.314527308Z" level=info msg="ImageCreate event name:\"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:21.320238 containerd[1579]: time="2025-09-11T05:14:21.320183146Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:21.321505 containerd[1579]: time="2025-09-11T05:14:21.321472875Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.5\" with image id \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\", size \"27681301\" in 2.345931096s" Sep 11 05:14:21.321505 containerd[1579]: time="2025-09-11T05:14:21.321505577Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\" returns image reference \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\"" Sep 11 05:14:21.322159 containerd[1579]: time="2025-09-11T05:14:21.322022126Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\"" Sep 11 05:14:24.618942 containerd[1579]: time="2025-09-11T05:14:24.618838170Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:24.620205 containerd[1579]: time="2025-09-11T05:14:24.620158166Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.5: active requests=0, bytes read=20155568" Sep 11 05:14:24.621880 containerd[1579]: time="2025-09-11T05:14:24.621759049Z" level=info msg="ImageCreate event name:\"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:24.625135 containerd[1579]: time="2025-09-11T05:14:24.625010167Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:24.626864 containerd[1579]: time="2025-09-11T05:14:24.626260122Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.5\" with image id \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\", size \"21816043\" in 3.304198773s" Sep 11 05:14:24.626864 containerd[1579]: time="2025-09-11T05:14:24.626289707Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\" returns image reference \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\"" Sep 11 05:14:24.626864 containerd[1579]: time="2025-09-11T05:14:24.626836363Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\"" Sep 11 05:14:26.047481 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3564715422.mount: Deactivated successfully. Sep 11 05:14:27.122263 containerd[1579]: time="2025-09-11T05:14:27.122176928Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:27.123252 containerd[1579]: time="2025-09-11T05:14:27.123135246Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.5: active requests=0, bytes read=31929469" Sep 11 05:14:27.125046 containerd[1579]: time="2025-09-11T05:14:27.125004943Z" level=info msg="ImageCreate event name:\"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:27.127298 containerd[1579]: time="2025-09-11T05:14:27.127244363Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:27.127852 containerd[1579]: time="2025-09-11T05:14:27.127805305Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.5\" with image id \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\", repo tag \"registry.k8s.io/kube-proxy:v1.33.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\", size \"31928488\" in 2.500932775s" Sep 11 05:14:27.127900 containerd[1579]: time="2025-09-11T05:14:27.127857363Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\" returns image reference \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\"" Sep 11 05:14:27.128680 containerd[1579]: time="2025-09-11T05:14:27.128645772Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 11 05:14:27.650514 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4012944154.mount: Deactivated successfully. Sep 11 05:14:29.302951 containerd[1579]: time="2025-09-11T05:14:29.302846641Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:29.305150 containerd[1579]: time="2025-09-11T05:14:29.305061426Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Sep 11 05:14:29.327761 containerd[1579]: time="2025-09-11T05:14:29.327701372Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:29.368911 containerd[1579]: time="2025-09-11T05:14:29.368847360Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:29.370442 containerd[1579]: time="2025-09-11T05:14:29.370370136Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 2.241686373s" Sep 11 05:14:29.370442 containerd[1579]: time="2025-09-11T05:14:29.370427955Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Sep 11 05:14:29.371089 containerd[1579]: time="2025-09-11T05:14:29.371059750Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 11 05:14:30.022264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount829388334.mount: Deactivated successfully. Sep 11 05:14:30.028420 containerd[1579]: time="2025-09-11T05:14:30.028334599Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 05:14:30.029643 containerd[1579]: time="2025-09-11T05:14:30.029590435Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 11 05:14:30.030697 containerd[1579]: time="2025-09-11T05:14:30.030650834Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 05:14:30.033241 containerd[1579]: time="2025-09-11T05:14:30.033160762Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 05:14:30.033849 containerd[1579]: time="2025-09-11T05:14:30.033804129Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 662.70764ms" Sep 11 05:14:30.033849 containerd[1579]: time="2025-09-11T05:14:30.033845416Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 11 05:14:30.034433 containerd[1579]: time="2025-09-11T05:14:30.034407701Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 11 05:14:30.785111 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 11 05:14:30.786896 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:30.993364 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:30.999401 (kubelet)[2178]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 05:14:31.257270 kubelet[2178]: E0911 05:14:31.257045 2178 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 05:14:31.262657 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 05:14:31.262896 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 05:14:31.263374 systemd[1]: kubelet.service: Consumed 233ms CPU time, 110.1M memory peak. Sep 11 05:14:31.621294 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1421570369.mount: Deactivated successfully. Sep 11 05:14:34.170897 containerd[1579]: time="2025-09-11T05:14:34.170815479Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:34.171799 containerd[1579]: time="2025-09-11T05:14:34.171773877Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58378433" Sep 11 05:14:34.173357 containerd[1579]: time="2025-09-11T05:14:34.173316270Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:34.176410 containerd[1579]: time="2025-09-11T05:14:34.176382061Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:14:34.178048 containerd[1579]: time="2025-09-11T05:14:34.177974217Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 4.143532303s" Sep 11 05:14:34.178115 containerd[1579]: time="2025-09-11T05:14:34.178051262Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Sep 11 05:14:38.136271 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:38.136451 systemd[1]: kubelet.service: Consumed 233ms CPU time, 110.1M memory peak. Sep 11 05:14:38.138763 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:38.166612 systemd[1]: Reload requested from client PID 2273 ('systemctl') (unit session-7.scope)... Sep 11 05:14:38.166627 systemd[1]: Reloading... Sep 11 05:14:38.249000 zram_generator::config[2315]: No configuration found. Sep 11 05:14:38.871788 systemd[1]: Reloading finished in 704 ms. Sep 11 05:14:38.938952 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 11 05:14:38.939081 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 11 05:14:38.939466 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:38.939536 systemd[1]: kubelet.service: Consumed 166ms CPU time, 98.3M memory peak. Sep 11 05:14:38.941885 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:39.200185 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:39.221424 (kubelet)[2363]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 11 05:14:39.270073 kubelet[2363]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 05:14:39.270073 kubelet[2363]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 11 05:14:39.270073 kubelet[2363]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 05:14:39.270520 kubelet[2363]: I0911 05:14:39.270136 2363 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 11 05:14:39.741375 kubelet[2363]: I0911 05:14:39.741309 2363 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 11 05:14:39.741375 kubelet[2363]: I0911 05:14:39.741353 2363 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 11 05:14:39.741753 kubelet[2363]: I0911 05:14:39.741723 2363 server.go:956] "Client rotation is on, will bootstrap in background" Sep 11 05:14:39.790529 kubelet[2363]: E0911 05:14:39.790462 2363 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.55:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 11 05:14:39.791111 kubelet[2363]: I0911 05:14:39.791066 2363 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 05:14:39.800535 kubelet[2363]: I0911 05:14:39.800462 2363 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 11 05:14:39.807598 kubelet[2363]: I0911 05:14:39.807530 2363 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 11 05:14:39.807992 kubelet[2363]: I0911 05:14:39.807953 2363 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 11 05:14:39.808236 kubelet[2363]: I0911 05:14:39.807989 2363 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 11 05:14:39.808436 kubelet[2363]: I0911 05:14:39.808245 2363 topology_manager.go:138] "Creating topology manager with none policy" Sep 11 05:14:39.808436 kubelet[2363]: I0911 05:14:39.808256 2363 container_manager_linux.go:303] "Creating device plugin manager" Sep 11 05:14:39.808527 kubelet[2363]: I0911 05:14:39.808500 2363 state_mem.go:36] "Initialized new in-memory state store" Sep 11 05:14:39.810416 kubelet[2363]: I0911 05:14:39.810382 2363 kubelet.go:480] "Attempting to sync node with API server" Sep 11 05:14:39.810416 kubelet[2363]: I0911 05:14:39.810405 2363 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 11 05:14:39.812115 kubelet[2363]: I0911 05:14:39.812020 2363 kubelet.go:386] "Adding apiserver pod source" Sep 11 05:14:39.814220 kubelet[2363]: I0911 05:14:39.813938 2363 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 11 05:14:39.820334 kubelet[2363]: E0911 05:14:39.820024 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.55:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 11 05:14:39.820632 kubelet[2363]: I0911 05:14:39.820608 2363 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 11 05:14:39.821277 kubelet[2363]: I0911 05:14:39.821253 2363 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 11 05:14:39.822013 kubelet[2363]: E0911 05:14:39.821967 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.55:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 11 05:14:39.823605 kubelet[2363]: W0911 05:14:39.823576 2363 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 11 05:14:39.828259 kubelet[2363]: I0911 05:14:39.828227 2363 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 11 05:14:39.828343 kubelet[2363]: I0911 05:14:39.828303 2363 server.go:1289] "Started kubelet" Sep 11 05:14:39.828705 kubelet[2363]: I0911 05:14:39.828636 2363 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 11 05:14:39.829199 kubelet[2363]: I0911 05:14:39.829181 2363 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 11 05:14:39.829306 kubelet[2363]: I0911 05:14:39.829231 2363 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 11 05:14:39.830660 kubelet[2363]: I0911 05:14:39.830597 2363 server.go:317] "Adding debug handlers to kubelet server" Sep 11 05:14:39.830963 kubelet[2363]: I0911 05:14:39.830885 2363 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 11 05:14:39.832059 kubelet[2363]: I0911 05:14:39.831479 2363 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 11 05:14:39.836248 kubelet[2363]: E0911 05:14:39.836202 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:39.836352 kubelet[2363]: I0911 05:14:39.836328 2363 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 11 05:14:39.837069 kubelet[2363]: I0911 05:14:39.837042 2363 reconciler.go:26] "Reconciler: start to sync state" Sep 11 05:14:39.837268 kubelet[2363]: I0911 05:14:39.837248 2363 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 11 05:14:39.837741 kubelet[2363]: E0911 05:14:39.837700 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.55:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 11 05:14:39.838820 kubelet[2363]: I0911 05:14:39.838790 2363 factory.go:223] Registration of the systemd container factory successfully Sep 11 05:14:39.838946 kubelet[2363]: I0911 05:14:39.838903 2363 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 11 05:14:39.839339 kubelet[2363]: E0911 05:14:39.838249 2363 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.55:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.55:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1864227e28deeaf9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-11 05:14:39.828257529 +0000 UTC m=+0.601662031,LastTimestamp:2025-09-11 05:14:39.828257529 +0000 UTC m=+0.601662031,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 11 05:14:39.840954 kubelet[2363]: E0911 05:14:39.839683 2363 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.55:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.55:6443: connect: connection refused" interval="200ms" Sep 11 05:14:39.840954 kubelet[2363]: E0911 05:14:39.839858 2363 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 11 05:14:39.842693 kubelet[2363]: I0911 05:14:39.842646 2363 factory.go:223] Registration of the containerd container factory successfully Sep 11 05:14:39.861572 kubelet[2363]: I0911 05:14:39.861530 2363 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 11 05:14:39.861572 kubelet[2363]: I0911 05:14:39.861558 2363 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 11 05:14:39.861572 kubelet[2363]: I0911 05:14:39.861581 2363 state_mem.go:36] "Initialized new in-memory state store" Sep 11 05:14:39.862867 kubelet[2363]: I0911 05:14:39.862826 2363 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 11 05:14:39.864959 kubelet[2363]: I0911 05:14:39.864442 2363 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 11 05:14:39.864959 kubelet[2363]: I0911 05:14:39.864478 2363 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 11 05:14:39.864959 kubelet[2363]: I0911 05:14:39.864512 2363 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 11 05:14:39.864959 kubelet[2363]: I0911 05:14:39.864532 2363 kubelet.go:2436] "Starting kubelet main sync loop" Sep 11 05:14:39.864959 kubelet[2363]: E0911 05:14:39.864578 2363 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 11 05:14:39.865770 kubelet[2363]: E0911 05:14:39.865746 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.55:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 11 05:14:39.921721 kubelet[2363]: I0911 05:14:39.921659 2363 policy_none.go:49] "None policy: Start" Sep 11 05:14:39.921721 kubelet[2363]: I0911 05:14:39.921708 2363 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 11 05:14:39.921721 kubelet[2363]: I0911 05:14:39.921728 2363 state_mem.go:35] "Initializing new in-memory state store" Sep 11 05:14:39.936370 kubelet[2363]: E0911 05:14:39.936325 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:39.950325 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 11 05:14:39.965020 kubelet[2363]: E0911 05:14:39.964985 2363 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 11 05:14:39.969085 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 11 05:14:39.974006 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 11 05:14:39.996101 kubelet[2363]: E0911 05:14:39.995714 2363 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 11 05:14:39.996101 kubelet[2363]: I0911 05:14:39.996076 2363 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 11 05:14:39.996278 kubelet[2363]: I0911 05:14:39.996109 2363 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 11 05:14:39.996823 kubelet[2363]: I0911 05:14:39.996787 2363 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 11 05:14:39.998609 kubelet[2363]: E0911 05:14:39.998515 2363 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 11 05:14:39.998609 kubelet[2363]: E0911 05:14:39.998575 2363 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 11 05:14:40.040843 kubelet[2363]: E0911 05:14:40.040774 2363 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.55:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.55:6443: connect: connection refused" interval="400ms" Sep 11 05:14:40.098376 kubelet[2363]: I0911 05:14:40.098281 2363 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:40.098817 kubelet[2363]: E0911 05:14:40.098751 2363 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.55:6443/api/v1/nodes\": dial tcp 10.0.0.55:6443: connect: connection refused" node="localhost" Sep 11 05:14:40.181190 systemd[1]: Created slice kubepods-burstable-pod258786ef9318cb16f0ed51ed242229c4.slice - libcontainer container kubepods-burstable-pod258786ef9318cb16f0ed51ed242229c4.slice. Sep 11 05:14:40.197198 kubelet[2363]: E0911 05:14:40.197164 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:40.200771 systemd[1]: Created slice kubepods-burstable-podb678d5c6713e936e66aa5bb73166297e.slice - libcontainer container kubepods-burstable-podb678d5c6713e936e66aa5bb73166297e.slice. Sep 11 05:14:40.203035 kubelet[2363]: E0911 05:14:40.202992 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:40.204699 systemd[1]: Created slice kubepods-burstable-pod7b968cf906b2d9d713a362c43868bef2.slice - libcontainer container kubepods-burstable-pod7b968cf906b2d9d713a362c43868bef2.slice. Sep 11 05:14:40.206720 kubelet[2363]: E0911 05:14:40.206690 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:40.240028 kubelet[2363]: I0911 05:14:40.239988 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/258786ef9318cb16f0ed51ed242229c4-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"258786ef9318cb16f0ed51ed242229c4\") " pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:40.240126 kubelet[2363]: I0911 05:14:40.240037 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/258786ef9318cb16f0ed51ed242229c4-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"258786ef9318cb16f0ed51ed242229c4\") " pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:40.240126 kubelet[2363]: I0911 05:14:40.240065 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:40.240126 kubelet[2363]: I0911 05:14:40.240089 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:40.240126 kubelet[2363]: I0911 05:14:40.240113 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:40.240283 kubelet[2363]: I0911 05:14:40.240161 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:40.240283 kubelet[2363]: I0911 05:14:40.240185 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/258786ef9318cb16f0ed51ed242229c4-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"258786ef9318cb16f0ed51ed242229c4\") " pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:40.240283 kubelet[2363]: I0911 05:14:40.240207 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:40.240283 kubelet[2363]: I0911 05:14:40.240231 2363 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7b968cf906b2d9d713a362c43868bef2-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7b968cf906b2d9d713a362c43868bef2\") " pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:40.301464 kubelet[2363]: I0911 05:14:40.301424 2363 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:40.302024 kubelet[2363]: E0911 05:14:40.301943 2363 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.55:6443/api/v1/nodes\": dial tcp 10.0.0.55:6443: connect: connection refused" node="localhost" Sep 11 05:14:40.441873 kubelet[2363]: E0911 05:14:40.441803 2363 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.55:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.55:6443: connect: connection refused" interval="800ms" Sep 11 05:14:40.498196 kubelet[2363]: E0911 05:14:40.498124 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:40.499053 containerd[1579]: time="2025-09-11T05:14:40.498982225Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:258786ef9318cb16f0ed51ed242229c4,Namespace:kube-system,Attempt:0,}" Sep 11 05:14:40.504099 kubelet[2363]: E0911 05:14:40.504070 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:40.504547 containerd[1579]: time="2025-09-11T05:14:40.504492854Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b678d5c6713e936e66aa5bb73166297e,Namespace:kube-system,Attempt:0,}" Sep 11 05:14:40.507979 kubelet[2363]: E0911 05:14:40.507911 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:40.508672 containerd[1579]: time="2025-09-11T05:14:40.508512434Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7b968cf906b2d9d713a362c43868bef2,Namespace:kube-system,Attempt:0,}" Sep 11 05:14:40.555046 containerd[1579]: time="2025-09-11T05:14:40.554843840Z" level=info msg="connecting to shim e860334969afdc56430ba3ff563375c19cd47e932776ec2ffae10f631e687d68" address="unix:///run/containerd/s/9278f4655e4924c928586afb12d57684148784206dff1d9f2b78b99805a095e6" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:14:40.555528 containerd[1579]: time="2025-09-11T05:14:40.555175027Z" level=info msg="connecting to shim 455c0594c77995fe7d1a629ebcc4787a50115873837c0c96f717211c68d4ada7" address="unix:///run/containerd/s/ac3c703d772e9178629dc758f45d09fafe6512ddd6ae2d0f3590e3af5ca980eb" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:14:40.566196 containerd[1579]: time="2025-09-11T05:14:40.564900502Z" level=info msg="connecting to shim c954e7cd1fe5f091f89751fd38ed9e3eababa6e7aa059a02b48d8e092458ccf4" address="unix:///run/containerd/s/4d0fdae1701d03960d57fb555241095ebf08af939d9ded4fd64c915aaf5e8868" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:14:40.704942 kubelet[2363]: I0911 05:14:40.704850 2363 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:40.705564 kubelet[2363]: E0911 05:14:40.705462 2363 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.55:6443/api/v1/nodes\": dial tcp 10.0.0.55:6443: connect: connection refused" node="localhost" Sep 11 05:14:40.876791 kubelet[2363]: E0911 05:14:40.876621 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.55:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 11 05:14:40.876791 kubelet[2363]: E0911 05:14:40.876633 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.55:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 11 05:14:40.913250 systemd[1]: Started cri-containerd-e860334969afdc56430ba3ff563375c19cd47e932776ec2ffae10f631e687d68.scope - libcontainer container e860334969afdc56430ba3ff563375c19cd47e932776ec2ffae10f631e687d68. Sep 11 05:14:40.918819 systemd[1]: Started cri-containerd-c954e7cd1fe5f091f89751fd38ed9e3eababa6e7aa059a02b48d8e092458ccf4.scope - libcontainer container c954e7cd1fe5f091f89751fd38ed9e3eababa6e7aa059a02b48d8e092458ccf4. Sep 11 05:14:40.951122 systemd[1]: Started cri-containerd-455c0594c77995fe7d1a629ebcc4787a50115873837c0c96f717211c68d4ada7.scope - libcontainer container 455c0594c77995fe7d1a629ebcc4787a50115873837c0c96f717211c68d4ada7. Sep 11 05:14:41.017484 kubelet[2363]: E0911 05:14:41.017442 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.55:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 11 05:14:41.164838 containerd[1579]: time="2025-09-11T05:14:41.164674799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b678d5c6713e936e66aa5bb73166297e,Namespace:kube-system,Attempt:0,} returns sandbox id \"e860334969afdc56430ba3ff563375c19cd47e932776ec2ffae10f631e687d68\"" Sep 11 05:14:41.166052 kubelet[2363]: E0911 05:14:41.165995 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:41.237999 kubelet[2363]: E0911 05:14:41.237916 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.55:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 11 05:14:41.242771 kubelet[2363]: E0911 05:14:41.242714 2363 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.55:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.55:6443: connect: connection refused" interval="1.6s" Sep 11 05:14:41.490649 containerd[1579]: time="2025-09-11T05:14:41.490487068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7b968cf906b2d9d713a362c43868bef2,Namespace:kube-system,Attempt:0,} returns sandbox id \"c954e7cd1fe5f091f89751fd38ed9e3eababa6e7aa059a02b48d8e092458ccf4\"" Sep 11 05:14:41.491532 kubelet[2363]: E0911 05:14:41.491472 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:41.507725 kubelet[2363]: I0911 05:14:41.507673 2363 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:41.508172 kubelet[2363]: E0911 05:14:41.508122 2363 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.55:6443/api/v1/nodes\": dial tcp 10.0.0.55:6443: connect: connection refused" node="localhost" Sep 11 05:14:41.527896 containerd[1579]: time="2025-09-11T05:14:41.527845516Z" level=info msg="CreateContainer within sandbox \"e860334969afdc56430ba3ff563375c19cd47e932776ec2ffae10f631e687d68\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 11 05:14:41.796757 kubelet[2363]: E0911 05:14:41.796711 2363 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.55:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 11 05:14:41.889476 containerd[1579]: time="2025-09-11T05:14:41.889420060Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:258786ef9318cb16f0ed51ed242229c4,Namespace:kube-system,Attempt:0,} returns sandbox id \"455c0594c77995fe7d1a629ebcc4787a50115873837c0c96f717211c68d4ada7\"" Sep 11 05:14:41.890370 kubelet[2363]: E0911 05:14:41.890338 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:41.903363 containerd[1579]: time="2025-09-11T05:14:41.903295223Z" level=info msg="CreateContainer within sandbox \"c954e7cd1fe5f091f89751fd38ed9e3eababa6e7aa059a02b48d8e092458ccf4\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 11 05:14:42.429898 containerd[1579]: time="2025-09-11T05:14:42.429838532Z" level=info msg="CreateContainer within sandbox \"455c0594c77995fe7d1a629ebcc4787a50115873837c0c96f717211c68d4ada7\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 11 05:14:42.843908 kubelet[2363]: E0911 05:14:42.843838 2363 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.55:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.55:6443: connect: connection refused" interval="3.2s" Sep 11 05:14:43.110326 kubelet[2363]: I0911 05:14:43.110204 2363 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:43.110693 kubelet[2363]: E0911 05:14:43.110659 2363 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.55:6443/api/v1/nodes\": dial tcp 10.0.0.55:6443: connect: connection refused" node="localhost" Sep 11 05:14:43.110833 kubelet[2363]: E0911 05:14:43.110806 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.55:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 11 05:14:43.569253 kubelet[2363]: E0911 05:14:43.569196 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.55:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 11 05:14:43.762122 kubelet[2363]: E0911 05:14:43.762059 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.55:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 11 05:14:44.001721 kubelet[2363]: E0911 05:14:44.001548 2363 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.55:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.55:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 11 05:14:44.214070 containerd[1579]: time="2025-09-11T05:14:44.214013977Z" level=info msg="Container b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:14:44.223289 containerd[1579]: time="2025-09-11T05:14:44.223222798Z" level=info msg="Container c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:14:44.228131 containerd[1579]: time="2025-09-11T05:14:44.228078655Z" level=info msg="Container 76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:14:44.238879 containerd[1579]: time="2025-09-11T05:14:44.238809107Z" level=info msg="CreateContainer within sandbox \"c954e7cd1fe5f091f89751fd38ed9e3eababa6e7aa059a02b48d8e092458ccf4\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d\"" Sep 11 05:14:44.240015 containerd[1579]: time="2025-09-11T05:14:44.239966550Z" level=info msg="StartContainer for \"c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d\"" Sep 11 05:14:44.241452 containerd[1579]: time="2025-09-11T05:14:44.241298357Z" level=info msg="connecting to shim c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d" address="unix:///run/containerd/s/4d0fdae1701d03960d57fb555241095ebf08af939d9ded4fd64c915aaf5e8868" protocol=ttrpc version=3 Sep 11 05:14:44.244287 containerd[1579]: time="2025-09-11T05:14:44.244062533Z" level=info msg="CreateContainer within sandbox \"e860334969afdc56430ba3ff563375c19cd47e932776ec2ffae10f631e687d68\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00\"" Sep 11 05:14:44.245261 containerd[1579]: time="2025-09-11T05:14:44.245198817Z" level=info msg="StartContainer for \"b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00\"" Sep 11 05:14:44.248803 containerd[1579]: time="2025-09-11T05:14:44.248409767Z" level=info msg="CreateContainer within sandbox \"455c0594c77995fe7d1a629ebcc4787a50115873837c0c96f717211c68d4ada7\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861\"" Sep 11 05:14:44.248911 containerd[1579]: time="2025-09-11T05:14:44.248888692Z" level=info msg="StartContainer for \"76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861\"" Sep 11 05:14:44.251905 containerd[1579]: time="2025-09-11T05:14:44.251732822Z" level=info msg="connecting to shim b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00" address="unix:///run/containerd/s/9278f4655e4924c928586afb12d57684148784206dff1d9f2b78b99805a095e6" protocol=ttrpc version=3 Sep 11 05:14:44.252663 containerd[1579]: time="2025-09-11T05:14:44.252611863Z" level=info msg="connecting to shim 76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861" address="unix:///run/containerd/s/ac3c703d772e9178629dc758f45d09fafe6512ddd6ae2d0f3590e3af5ca980eb" protocol=ttrpc version=3 Sep 11 05:14:44.273132 systemd[1]: Started cri-containerd-c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d.scope - libcontainer container c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d. Sep 11 05:14:44.284188 systemd[1]: Started cri-containerd-b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00.scope - libcontainer container b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00. Sep 11 05:14:44.288419 systemd[1]: Started cri-containerd-76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861.scope - libcontainer container 76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861. Sep 11 05:14:44.504297 containerd[1579]: time="2025-09-11T05:14:44.503471946Z" level=info msg="StartContainer for \"c791a5ccc055de036d461557f42ded7696c9a123ee22862a31b81519f3bad27d\" returns successfully" Sep 11 05:14:44.508176 containerd[1579]: time="2025-09-11T05:14:44.508139542Z" level=info msg="StartContainer for \"76eda11b3d274d25d54773aef68a1bbf5200ed85c7789fa27082f72500c45861\" returns successfully" Sep 11 05:14:44.509232 containerd[1579]: time="2025-09-11T05:14:44.509171155Z" level=info msg="StartContainer for \"b2dd49af52fb6edfc553b4e6df91ba73aa72cb7a61e084aa5a3d90bbaff53c00\" returns successfully" Sep 11 05:14:44.882709 kubelet[2363]: E0911 05:14:44.882260 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:44.882709 kubelet[2363]: E0911 05:14:44.882385 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:44.885890 kubelet[2363]: E0911 05:14:44.885714 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:44.885890 kubelet[2363]: E0911 05:14:44.885816 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:44.888762 kubelet[2363]: E0911 05:14:44.888552 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:44.888762 kubelet[2363]: E0911 05:14:44.888665 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:45.891951 kubelet[2363]: E0911 05:14:45.891452 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:45.891951 kubelet[2363]: E0911 05:14:45.891514 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:45.891951 kubelet[2363]: E0911 05:14:45.891665 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:45.891951 kubelet[2363]: E0911 05:14:45.891665 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:45.891951 kubelet[2363]: E0911 05:14:45.891889 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:45.893016 kubelet[2363]: E0911 05:14:45.892939 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:46.312696 kubelet[2363]: I0911 05:14:46.312633 2363 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:46.760509 kubelet[2363]: E0911 05:14:46.760355 2363 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 11 05:14:46.892530 kubelet[2363]: E0911 05:14:46.892493 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:46.892530 kubelet[2363]: E0911 05:14:46.892494 2363 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 05:14:46.893053 kubelet[2363]: E0911 05:14:46.892628 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:46.893053 kubelet[2363]: E0911 05:14:46.892656 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:47.066959 kubelet[2363]: I0911 05:14:47.066746 2363 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 11 05:14:47.066959 kubelet[2363]: E0911 05:14:47.066803 2363 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 11 05:14:47.067409 kubelet[2363]: E0911 05:14:47.067285 2363 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1864227e28deeaf9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-11 05:14:39.828257529 +0000 UTC m=+0.601662031,LastTimestamp:2025-09-11 05:14:39.828257529 +0000 UTC m=+0.601662031,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 11 05:14:47.085141 kubelet[2363]: E0911 05:14:47.085085 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.186175 kubelet[2363]: E0911 05:14:47.186093 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.286689 kubelet[2363]: E0911 05:14:47.286617 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.387805 kubelet[2363]: E0911 05:14:47.387637 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.488020 kubelet[2363]: E0911 05:14:47.487777 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.588957 kubelet[2363]: E0911 05:14:47.588869 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.690166 kubelet[2363]: E0911 05:14:47.689969 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.790710 kubelet[2363]: E0911 05:14:47.790647 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.891403 kubelet[2363]: E0911 05:14:47.891311 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:47.991726 kubelet[2363]: E0911 05:14:47.991532 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:48.092502 kubelet[2363]: E0911 05:14:48.092431 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:48.193611 kubelet[2363]: E0911 05:14:48.193536 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:48.294662 kubelet[2363]: E0911 05:14:48.294580 2363 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 05:14:48.438138 kubelet[2363]: I0911 05:14:48.438068 2363 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:48.452040 kubelet[2363]: I0911 05:14:48.451988 2363 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:48.458462 kubelet[2363]: I0911 05:14:48.458381 2363 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:48.819784 kubelet[2363]: I0911 05:14:48.819741 2363 apiserver.go:52] "Watching apiserver" Sep 11 05:14:48.822316 kubelet[2363]: E0911 05:14:48.822284 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:48.823009 kubelet[2363]: E0911 05:14:48.822555 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:48.823009 kubelet[2363]: E0911 05:14:48.822564 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:48.838343 kubelet[2363]: I0911 05:14:48.838286 2363 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 11 05:14:49.635672 kubelet[2363]: E0911 05:14:49.635622 2363 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:49.772367 systemd[1]: Reload requested from client PID 2649 ('systemctl') (unit session-7.scope)... Sep 11 05:14:49.772395 systemd[1]: Reloading... Sep 11 05:14:49.872972 zram_generator::config[2692]: No configuration found. Sep 11 05:14:49.927342 kubelet[2363]: I0911 05:14:49.927134 2363 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.927108094 podStartE2EDuration="1.927108094s" podCreationTimestamp="2025-09-11 05:14:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 05:14:49.916582282 +0000 UTC m=+10.689986794" watchObservedRunningTime="2025-09-11 05:14:49.927108094 +0000 UTC m=+10.700512596" Sep 11 05:14:49.939192 kubelet[2363]: I0911 05:14:49.939103 2363 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.939079488 podStartE2EDuration="1.939079488s" podCreationTimestamp="2025-09-11 05:14:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 05:14:49.927548232 +0000 UTC m=+10.700952744" watchObservedRunningTime="2025-09-11 05:14:49.939079488 +0000 UTC m=+10.712483990" Sep 11 05:14:49.939431 kubelet[2363]: I0911 05:14:49.939203 2363 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.939197582 podStartE2EDuration="1.939197582s" podCreationTimestamp="2025-09-11 05:14:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 05:14:49.938173575 +0000 UTC m=+10.711578087" watchObservedRunningTime="2025-09-11 05:14:49.939197582 +0000 UTC m=+10.712602104" Sep 11 05:14:50.145709 systemd[1]: Reloading finished in 372 ms. Sep 11 05:14:50.173010 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:50.191191 systemd[1]: kubelet.service: Deactivated successfully. Sep 11 05:14:50.191672 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:50.191750 systemd[1]: kubelet.service: Consumed 1.209s CPU time, 132.5M memory peak. Sep 11 05:14:50.194139 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 05:14:50.426573 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 05:14:50.443389 (kubelet)[2739]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 11 05:14:50.490627 kubelet[2739]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 05:14:50.491128 kubelet[2739]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 11 05:14:50.491128 kubelet[2739]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 05:14:50.491128 kubelet[2739]: I0911 05:14:50.490768 2739 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 11 05:14:50.501036 kubelet[2739]: I0911 05:14:50.500972 2739 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 11 05:14:50.501036 kubelet[2739]: I0911 05:14:50.501016 2739 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 11 05:14:50.501356 kubelet[2739]: I0911 05:14:50.501328 2739 server.go:956] "Client rotation is on, will bootstrap in background" Sep 11 05:14:50.503060 kubelet[2739]: I0911 05:14:50.503024 2739 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 11 05:14:50.512242 kubelet[2739]: I0911 05:14:50.512164 2739 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 05:14:50.514199 kubelet[2739]: I0911 05:14:50.514176 2739 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 11 05:14:50.520796 kubelet[2739]: I0911 05:14:50.520745 2739 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 11 05:14:50.521123 kubelet[2739]: I0911 05:14:50.521079 2739 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 11 05:14:50.521322 kubelet[2739]: I0911 05:14:50.521116 2739 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 11 05:14:50.521437 kubelet[2739]: I0911 05:14:50.521328 2739 topology_manager.go:138] "Creating topology manager with none policy" Sep 11 05:14:50.521437 kubelet[2739]: I0911 05:14:50.521340 2739 container_manager_linux.go:303] "Creating device plugin manager" Sep 11 05:14:50.521437 kubelet[2739]: I0911 05:14:50.521402 2739 state_mem.go:36] "Initialized new in-memory state store" Sep 11 05:14:50.521660 kubelet[2739]: I0911 05:14:50.521634 2739 kubelet.go:480] "Attempting to sync node with API server" Sep 11 05:14:50.521660 kubelet[2739]: I0911 05:14:50.521654 2739 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 11 05:14:50.521721 kubelet[2739]: I0911 05:14:50.521697 2739 kubelet.go:386] "Adding apiserver pod source" Sep 11 05:14:50.521721 kubelet[2739]: I0911 05:14:50.521716 2739 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 11 05:14:50.525754 kubelet[2739]: I0911 05:14:50.523369 2739 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 11 05:14:50.525754 kubelet[2739]: I0911 05:14:50.523978 2739 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 11 05:14:50.528775 kubelet[2739]: I0911 05:14:50.528754 2739 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 11 05:14:50.528857 kubelet[2739]: I0911 05:14:50.528845 2739 server.go:1289] "Started kubelet" Sep 11 05:14:50.530953 kubelet[2739]: I0911 05:14:50.529454 2739 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 11 05:14:50.530953 kubelet[2739]: I0911 05:14:50.530841 2739 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 11 05:14:50.530953 kubelet[2739]: I0911 05:14:50.530951 2739 server.go:317] "Adding debug handlers to kubelet server" Sep 11 05:14:50.532272 kubelet[2739]: I0911 05:14:50.529484 2739 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 11 05:14:50.536988 kubelet[2739]: I0911 05:14:50.532721 2739 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 11 05:14:50.536988 kubelet[2739]: I0911 05:14:50.534280 2739 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 11 05:14:50.544417 kubelet[2739]: I0911 05:14:50.544380 2739 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 11 05:14:50.544945 kubelet[2739]: I0911 05:14:50.544840 2739 reconciler.go:26] "Reconciler: start to sync state" Sep 11 05:14:50.546124 kubelet[2739]: I0911 05:14:50.544848 2739 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 11 05:14:50.552002 kubelet[2739]: I0911 05:14:50.551243 2739 factory.go:223] Registration of the systemd container factory successfully Sep 11 05:14:50.552002 kubelet[2739]: I0911 05:14:50.551449 2739 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 11 05:14:50.553792 kubelet[2739]: E0911 05:14:50.553750 2739 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 11 05:14:50.558371 kubelet[2739]: I0911 05:14:50.558331 2739 factory.go:223] Registration of the containerd container factory successfully Sep 11 05:14:50.564581 kubelet[2739]: I0911 05:14:50.564512 2739 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 11 05:14:50.574671 kubelet[2739]: I0911 05:14:50.574519 2739 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 11 05:14:50.574671 kubelet[2739]: I0911 05:14:50.574659 2739 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 11 05:14:50.574838 kubelet[2739]: I0911 05:14:50.574683 2739 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 11 05:14:50.574838 kubelet[2739]: I0911 05:14:50.574793 2739 kubelet.go:2436] "Starting kubelet main sync loop" Sep 11 05:14:50.574907 kubelet[2739]: E0911 05:14:50.574837 2739 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 11 05:14:50.595113 update_engine[1518]: I20250911 05:14:50.595036 1518 update_attempter.cc:509] Updating boot flags... Sep 11 05:14:50.606562 kubelet[2739]: I0911 05:14:50.606522 2739 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 11 05:14:50.606562 kubelet[2739]: I0911 05:14:50.606542 2739 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 11 05:14:50.606562 kubelet[2739]: I0911 05:14:50.606564 2739 state_mem.go:36] "Initialized new in-memory state store" Sep 11 05:14:50.606770 kubelet[2739]: I0911 05:14:50.606747 2739 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 11 05:14:50.606849 kubelet[2739]: I0911 05:14:50.606767 2739 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 11 05:14:50.606849 kubelet[2739]: I0911 05:14:50.606794 2739 policy_none.go:49] "None policy: Start" Sep 11 05:14:50.606849 kubelet[2739]: I0911 05:14:50.606807 2739 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 11 05:14:50.606849 kubelet[2739]: I0911 05:14:50.606823 2739 state_mem.go:35] "Initializing new in-memory state store" Sep 11 05:14:50.607125 kubelet[2739]: I0911 05:14:50.607068 2739 state_mem.go:75] "Updated machine memory state" Sep 11 05:14:50.611877 kubelet[2739]: E0911 05:14:50.611360 2739 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 11 05:14:50.611877 kubelet[2739]: I0911 05:14:50.611572 2739 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 11 05:14:50.611877 kubelet[2739]: I0911 05:14:50.611587 2739 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 11 05:14:50.611877 kubelet[2739]: I0911 05:14:50.611803 2739 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 11 05:14:50.612856 kubelet[2739]: E0911 05:14:50.612824 2739 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 11 05:14:50.676655 kubelet[2739]: I0911 05:14:50.676577 2739 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:50.676655 kubelet[2739]: I0911 05:14:50.676652 2739 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.676851 kubelet[2739]: I0911 05:14:50.676690 2739 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:50.727199 kubelet[2739]: I0911 05:14:50.727033 2739 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 05:14:50.748115 kubelet[2739]: I0911 05:14:50.748041 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/258786ef9318cb16f0ed51ed242229c4-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"258786ef9318cb16f0ed51ed242229c4\") " pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:50.748115 kubelet[2739]: I0911 05:14:50.748097 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/258786ef9318cb16f0ed51ed242229c4-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"258786ef9318cb16f0ed51ed242229c4\") " pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:50.748376 kubelet[2739]: I0911 05:14:50.748136 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.748376 kubelet[2739]: I0911 05:14:50.748162 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.748376 kubelet[2739]: I0911 05:14:50.748206 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.748376 kubelet[2739]: I0911 05:14:50.748258 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/258786ef9318cb16f0ed51ed242229c4-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"258786ef9318cb16f0ed51ed242229c4\") " pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:50.748376 kubelet[2739]: I0911 05:14:50.748318 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.748653 kubelet[2739]: I0911 05:14:50.748360 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.748653 kubelet[2739]: I0911 05:14:50.748405 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7b968cf906b2d9d713a362c43868bef2-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7b968cf906b2d9d713a362c43868bef2\") " pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:50.895798 kubelet[2739]: E0911 05:14:50.895729 2739 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:50.896051 kubelet[2739]: E0911 05:14:50.896023 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:50.897986 kubelet[2739]: E0911 05:14:50.897917 2739 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:50.898060 kubelet[2739]: E0911 05:14:50.898022 2739 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 11 05:14:50.898200 kubelet[2739]: E0911 05:14:50.898172 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:50.898428 kubelet[2739]: E0911 05:14:50.898361 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:50.901327 kubelet[2739]: I0911 05:14:50.901277 2739 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 11 05:14:50.901672 kubelet[2739]: I0911 05:14:50.901611 2739 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 11 05:14:51.522493 kubelet[2739]: I0911 05:14:51.522378 2739 apiserver.go:52] "Watching apiserver" Sep 11 05:14:51.546705 kubelet[2739]: I0911 05:14:51.546211 2739 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 11 05:14:51.589389 kubelet[2739]: I0911 05:14:51.589337 2739 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:51.589389 kubelet[2739]: I0911 05:14:51.589397 2739 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:51.589854 kubelet[2739]: E0911 05:14:51.589640 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:51.858255 kubelet[2739]: E0911 05:14:51.857954 2739 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 11 05:14:51.858255 kubelet[2739]: E0911 05:14:51.858089 2739 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 11 05:14:51.858255 kubelet[2739]: E0911 05:14:51.858262 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:51.858582 kubelet[2739]: E0911 05:14:51.858432 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:52.591040 kubelet[2739]: E0911 05:14:52.591008 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:52.591445 kubelet[2739]: E0911 05:14:52.591159 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:54.104022 kubelet[2739]: I0911 05:14:54.103963 2739 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 11 05:14:54.104630 kubelet[2739]: I0911 05:14:54.104574 2739 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 11 05:14:54.104664 containerd[1579]: time="2025-09-11T05:14:54.104333360Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 11 05:14:55.385576 systemd[1]: Created slice kubepods-besteffort-pod76e4c769_6394_49d5_93c3_5c1ae385d41a.slice - libcontainer container kubepods-besteffort-pod76e4c769_6394_49d5_93c3_5c1ae385d41a.slice. Sep 11 05:14:55.417166 systemd[1]: Created slice kubepods-besteffort-pod464351fa_bc0b_4434_b3bc_2aa8ea50462f.slice - libcontainer container kubepods-besteffort-pod464351fa_bc0b_4434_b3bc_2aa8ea50462f.slice. Sep 11 05:14:55.478295 kubelet[2739]: I0911 05:14:55.478214 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/76e4c769-6394-49d5-93c3-5c1ae385d41a-lib-modules\") pod \"kube-proxy-tj542\" (UID: \"76e4c769-6394-49d5-93c3-5c1ae385d41a\") " pod="kube-system/kube-proxy-tj542" Sep 11 05:14:55.478295 kubelet[2739]: I0911 05:14:55.478280 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xxk7\" (UniqueName: \"kubernetes.io/projected/464351fa-bc0b-4434-b3bc-2aa8ea50462f-kube-api-access-6xxk7\") pod \"tigera-operator-755d956888-886qb\" (UID: \"464351fa-bc0b-4434-b3bc-2aa8ea50462f\") " pod="tigera-operator/tigera-operator-755d956888-886qb" Sep 11 05:14:55.478838 kubelet[2739]: I0911 05:14:55.478322 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/76e4c769-6394-49d5-93c3-5c1ae385d41a-kube-proxy\") pod \"kube-proxy-tj542\" (UID: \"76e4c769-6394-49d5-93c3-5c1ae385d41a\") " pod="kube-system/kube-proxy-tj542" Sep 11 05:14:55.478838 kubelet[2739]: I0911 05:14:55.478352 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqxnq\" (UniqueName: \"kubernetes.io/projected/76e4c769-6394-49d5-93c3-5c1ae385d41a-kube-api-access-rqxnq\") pod \"kube-proxy-tj542\" (UID: \"76e4c769-6394-49d5-93c3-5c1ae385d41a\") " pod="kube-system/kube-proxy-tj542" Sep 11 05:14:55.478838 kubelet[2739]: I0911 05:14:55.478376 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/464351fa-bc0b-4434-b3bc-2aa8ea50462f-var-lib-calico\") pod \"tigera-operator-755d956888-886qb\" (UID: \"464351fa-bc0b-4434-b3bc-2aa8ea50462f\") " pod="tigera-operator/tigera-operator-755d956888-886qb" Sep 11 05:14:55.478838 kubelet[2739]: I0911 05:14:55.478395 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/76e4c769-6394-49d5-93c3-5c1ae385d41a-xtables-lock\") pod \"kube-proxy-tj542\" (UID: \"76e4c769-6394-49d5-93c3-5c1ae385d41a\") " pod="kube-system/kube-proxy-tj542" Sep 11 05:14:55.697710 kubelet[2739]: E0911 05:14:55.697531 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:55.698445 containerd[1579]: time="2025-09-11T05:14:55.698405370Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-tj542,Uid:76e4c769-6394-49d5-93c3-5c1ae385d41a,Namespace:kube-system,Attempt:0,}" Sep 11 05:14:55.721150 containerd[1579]: time="2025-09-11T05:14:55.721080822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-886qb,Uid:464351fa-bc0b-4434-b3bc-2aa8ea50462f,Namespace:tigera-operator,Attempt:0,}" Sep 11 05:14:55.959220 containerd[1579]: time="2025-09-11T05:14:55.959048838Z" level=info msg="connecting to shim c0530c72c36871ec1aa8945670d71cd2006812f45f3d3d938c9ae037fa8defdb" address="unix:///run/containerd/s/54f18984507c488fe81467939ca8f599ea4e9928ee14ca3ed9398436ef3d9d84" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:14:55.971272 containerd[1579]: time="2025-09-11T05:14:55.971146759Z" level=info msg="connecting to shim 3e5ce8ac1a0114a133f5ffc1b5b144b26fa8ffdfee0581bfbac9b0481e4f74a2" address="unix:///run/containerd/s/122d7c7d684610993316613d7498d07e818124487f05cca25b81a8a7ec817a32" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:14:55.992140 systemd[1]: Started cri-containerd-c0530c72c36871ec1aa8945670d71cd2006812f45f3d3d938c9ae037fa8defdb.scope - libcontainer container c0530c72c36871ec1aa8945670d71cd2006812f45f3d3d938c9ae037fa8defdb. Sep 11 05:14:55.996432 systemd[1]: Started cri-containerd-3e5ce8ac1a0114a133f5ffc1b5b144b26fa8ffdfee0581bfbac9b0481e4f74a2.scope - libcontainer container 3e5ce8ac1a0114a133f5ffc1b5b144b26fa8ffdfee0581bfbac9b0481e4f74a2. Sep 11 05:14:56.043499 containerd[1579]: time="2025-09-11T05:14:56.043449656Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-tj542,Uid:76e4c769-6394-49d5-93c3-5c1ae385d41a,Namespace:kube-system,Attempt:0,} returns sandbox id \"c0530c72c36871ec1aa8945670d71cd2006812f45f3d3d938c9ae037fa8defdb\"" Sep 11 05:14:56.044782 kubelet[2739]: E0911 05:14:56.044538 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:56.057478 containerd[1579]: time="2025-09-11T05:14:56.057417402Z" level=info msg="CreateContainer within sandbox \"c0530c72c36871ec1aa8945670d71cd2006812f45f3d3d938c9ae037fa8defdb\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 11 05:14:56.069175 containerd[1579]: time="2025-09-11T05:14:56.069117416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-886qb,Uid:464351fa-bc0b-4434-b3bc-2aa8ea50462f,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"3e5ce8ac1a0114a133f5ffc1b5b144b26fa8ffdfee0581bfbac9b0481e4f74a2\"" Sep 11 05:14:56.071095 containerd[1579]: time="2025-09-11T05:14:56.070914257Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 11 05:14:56.073619 containerd[1579]: time="2025-09-11T05:14:56.073590873Z" level=info msg="Container 29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:14:56.085337 containerd[1579]: time="2025-09-11T05:14:56.085186098Z" level=info msg="CreateContainer within sandbox \"c0530c72c36871ec1aa8945670d71cd2006812f45f3d3d938c9ae037fa8defdb\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5\"" Sep 11 05:14:56.086043 containerd[1579]: time="2025-09-11T05:14:56.086003865Z" level=info msg="StartContainer for \"29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5\"" Sep 11 05:14:56.087787 containerd[1579]: time="2025-09-11T05:14:56.087756443Z" level=info msg="connecting to shim 29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5" address="unix:///run/containerd/s/54f18984507c488fe81467939ca8f599ea4e9928ee14ca3ed9398436ef3d9d84" protocol=ttrpc version=3 Sep 11 05:14:56.110211 systemd[1]: Started cri-containerd-29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5.scope - libcontainer container 29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5. Sep 11 05:14:56.165421 containerd[1579]: time="2025-09-11T05:14:56.165376048Z" level=info msg="StartContainer for \"29d74fcbb7e3ad76ebaf431637f886e3f90d5f9ce8acfea7d7bc696a5a45d2a5\" returns successfully" Sep 11 05:14:56.601566 kubelet[2739]: E0911 05:14:56.601502 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:56.614261 kubelet[2739]: I0911 05:14:56.614176 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-tj542" podStartSLOduration=1.614151579 podStartE2EDuration="1.614151579s" podCreationTimestamp="2025-09-11 05:14:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 05:14:56.613763304 +0000 UTC m=+6.164248491" watchObservedRunningTime="2025-09-11 05:14:56.614151579 +0000 UTC m=+6.164636755" Sep 11 05:14:58.478295 kubelet[2739]: E0911 05:14:58.478111 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:58.605833 kubelet[2739]: E0911 05:14:58.605779 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:59.819360 kubelet[2739]: E0911 05:14:59.819319 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:14:59.990505 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4167796070.mount: Deactivated successfully. Sep 11 05:15:00.404689 kubelet[2739]: E0911 05:15:00.404641 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:00.609836 kubelet[2739]: E0911 05:15:00.609779 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:00.610948 kubelet[2739]: E0911 05:15:00.610419 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:00.612084 containerd[1579]: time="2025-09-11T05:15:00.611890718Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:00.621625 containerd[1579]: time="2025-09-11T05:15:00.621564831Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Sep 11 05:15:00.626727 containerd[1579]: time="2025-09-11T05:15:00.626643152Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:00.659964 containerd[1579]: time="2025-09-11T05:15:00.659788557Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:00.660316 containerd[1579]: time="2025-09-11T05:15:00.660268354Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 4.589300276s" Sep 11 05:15:00.660316 containerd[1579]: time="2025-09-11T05:15:00.660312447Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Sep 11 05:15:00.667162 containerd[1579]: time="2025-09-11T05:15:00.667102962Z" level=info msg="CreateContainer within sandbox \"3e5ce8ac1a0114a133f5ffc1b5b144b26fa8ffdfee0581bfbac9b0481e4f74a2\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 11 05:15:00.678073 containerd[1579]: time="2025-09-11T05:15:00.678000506Z" level=info msg="Container 1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:00.686780 containerd[1579]: time="2025-09-11T05:15:00.686708684Z" level=info msg="CreateContainer within sandbox \"3e5ce8ac1a0114a133f5ffc1b5b144b26fa8ffdfee0581bfbac9b0481e4f74a2\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167\"" Sep 11 05:15:00.687548 containerd[1579]: time="2025-09-11T05:15:00.687473178Z" level=info msg="StartContainer for \"1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167\"" Sep 11 05:15:00.688968 containerd[1579]: time="2025-09-11T05:15:00.688912166Z" level=info msg="connecting to shim 1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167" address="unix:///run/containerd/s/122d7c7d684610993316613d7498d07e818124487f05cca25b81a8a7ec817a32" protocol=ttrpc version=3 Sep 11 05:15:00.751135 systemd[1]: Started cri-containerd-1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167.scope - libcontainer container 1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167. Sep 11 05:15:00.788283 containerd[1579]: time="2025-09-11T05:15:00.788224115Z" level=info msg="StartContainer for \"1f5dd9d45e4024f74167cfb0978240c59f756ce4478123aa9b88737503ccb167\" returns successfully" Sep 11 05:15:01.613745 kubelet[2739]: E0911 05:15:01.613690 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:04.470362 sudo[1784]: pam_unix(sudo:session): session closed for user root Sep 11 05:15:04.474950 sshd[1783]: Connection closed by 10.0.0.1 port 49964 Sep 11 05:15:04.478395 sshd-session[1780]: pam_unix(sshd:session): session closed for user core Sep 11 05:15:04.484484 systemd[1]: sshd@6-10.0.0.55:22-10.0.0.1:49964.service: Deactivated successfully. Sep 11 05:15:04.487335 systemd[1]: session-7.scope: Deactivated successfully. Sep 11 05:15:04.487580 systemd[1]: session-7.scope: Consumed 6.886s CPU time, 223.7M memory peak. Sep 11 05:15:04.489865 systemd-logind[1516]: Session 7 logged out. Waiting for processes to exit. Sep 11 05:15:04.492702 systemd-logind[1516]: Removed session 7. Sep 11 05:15:09.397502 kubelet[2739]: I0911 05:15:09.397419 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-886qb" podStartSLOduration=9.806711991 podStartE2EDuration="14.39739002s" podCreationTimestamp="2025-09-11 05:14:55 +0000 UTC" firstStartedPulling="2025-09-11 05:14:56.070603979 +0000 UTC m=+5.621089155" lastFinishedPulling="2025-09-11 05:15:00.661282008 +0000 UTC m=+10.211767184" observedRunningTime="2025-09-11 05:15:01.622972831 +0000 UTC m=+11.173458007" watchObservedRunningTime="2025-09-11 05:15:09.39739002 +0000 UTC m=+18.947875196" Sep 11 05:15:09.968029 systemd[1]: Created slice kubepods-besteffort-pode3ffe175_4f19_4f9c_b627_d660a0f8b22e.slice - libcontainer container kubepods-besteffort-pode3ffe175_4f19_4f9c_b627_d660a0f8b22e.slice. Sep 11 05:15:09.984735 kubelet[2739]: I0911 05:15:09.984664 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/e3ffe175-4f19-4f9c-b627-d660a0f8b22e-typha-certs\") pod \"calico-typha-559b656695-87jhb\" (UID: \"e3ffe175-4f19-4f9c-b627-d660a0f8b22e\") " pod="calico-system/calico-typha-559b656695-87jhb" Sep 11 05:15:09.984735 kubelet[2739]: I0911 05:15:09.984740 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3ffe175-4f19-4f9c-b627-d660a0f8b22e-tigera-ca-bundle\") pod \"calico-typha-559b656695-87jhb\" (UID: \"e3ffe175-4f19-4f9c-b627-d660a0f8b22e\") " pod="calico-system/calico-typha-559b656695-87jhb" Sep 11 05:15:09.984961 kubelet[2739]: I0911 05:15:09.984774 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6q2\" (UniqueName: \"kubernetes.io/projected/e3ffe175-4f19-4f9c-b627-d660a0f8b22e-kube-api-access-9n6q2\") pod \"calico-typha-559b656695-87jhb\" (UID: \"e3ffe175-4f19-4f9c-b627-d660a0f8b22e\") " pod="calico-system/calico-typha-559b656695-87jhb" Sep 11 05:15:10.177238 systemd[1]: Created slice kubepods-besteffort-pod617fd3b1_b716_4dda_8bfd_ba04f5d59b78.slice - libcontainer container kubepods-besteffort-pod617fd3b1_b716_4dda_8bfd_ba04f5d59b78.slice. Sep 11 05:15:10.283229 kubelet[2739]: E0911 05:15:10.282952 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:10.286184 kubelet[2739]: I0911 05:15:10.286122 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-var-lib-calico\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286382 kubelet[2739]: I0911 05:15:10.286194 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-cni-log-dir\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286382 kubelet[2739]: I0911 05:15:10.286220 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-xtables-lock\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286382 kubelet[2739]: I0911 05:15:10.286241 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9j4w\" (UniqueName: \"kubernetes.io/projected/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-kube-api-access-h9j4w\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286382 kubelet[2739]: I0911 05:15:10.286276 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-cni-bin-dir\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286864 kubelet[2739]: I0911 05:15:10.286605 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-cni-net-dir\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286864 kubelet[2739]: I0911 05:15:10.286651 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-flexvol-driver-host\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286864 kubelet[2739]: I0911 05:15:10.286677 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-policysync\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286864 kubelet[2739]: I0911 05:15:10.286734 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-var-run-calico\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.286864 kubelet[2739]: I0911 05:15:10.286759 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-node-certs\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.287112 kubelet[2739]: I0911 05:15:10.286778 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-tigera-ca-bundle\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.287112 kubelet[2739]: I0911 05:15:10.286804 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/617fd3b1-b716-4dda-8bfd-ba04f5d59b78-lib-modules\") pod \"calico-node-rmm9d\" (UID: \"617fd3b1-b716-4dda-8bfd-ba04f5d59b78\") " pod="calico-system/calico-node-rmm9d" Sep 11 05:15:10.290531 kubelet[2739]: E0911 05:15:10.290468 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:10.291897 containerd[1579]: time="2025-09-11T05:15:10.291842697Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-559b656695-87jhb,Uid:e3ffe175-4f19-4f9c-b627-d660a0f8b22e,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:10.352245 containerd[1579]: time="2025-09-11T05:15:10.352184986Z" level=info msg="connecting to shim 312e3d113bb45153b857b550186ea5e313ca3a711a266ecf1e7b580a9ba593f9" address="unix:///run/containerd/s/26248250fbad33668e7b791a16829c4d62062f12ca7a90d8e8e477a55b5b52f2" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:10.387600 kubelet[2739]: I0911 05:15:10.387556 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/01cab480-4fed-4f83-9e86-d02437c18ae7-registration-dir\") pod \"csi-node-driver-kqvtp\" (UID: \"01cab480-4fed-4f83-9e86-d02437c18ae7\") " pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:10.390340 kubelet[2739]: I0911 05:15:10.389233 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbftq\" (UniqueName: \"kubernetes.io/projected/01cab480-4fed-4f83-9e86-d02437c18ae7-kube-api-access-pbftq\") pod \"csi-node-driver-kqvtp\" (UID: \"01cab480-4fed-4f83-9e86-d02437c18ae7\") " pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:10.390340 kubelet[2739]: I0911 05:15:10.389289 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/01cab480-4fed-4f83-9e86-d02437c18ae7-varrun\") pod \"csi-node-driver-kqvtp\" (UID: \"01cab480-4fed-4f83-9e86-d02437c18ae7\") " pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:10.390340 kubelet[2739]: I0911 05:15:10.389383 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01cab480-4fed-4f83-9e86-d02437c18ae7-kubelet-dir\") pod \"csi-node-driver-kqvtp\" (UID: \"01cab480-4fed-4f83-9e86-d02437c18ae7\") " pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:10.390340 kubelet[2739]: I0911 05:15:10.389408 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/01cab480-4fed-4f83-9e86-d02437c18ae7-socket-dir\") pod \"csi-node-driver-kqvtp\" (UID: \"01cab480-4fed-4f83-9e86-d02437c18ae7\") " pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:10.397038 kubelet[2739]: E0911 05:15:10.397011 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.397999 kubelet[2739]: W0911 05:15:10.397978 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.401159 kubelet[2739]: E0911 05:15:10.401109 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.401767 kubelet[2739]: E0911 05:15:10.401727 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.401767 kubelet[2739]: W0911 05:15:10.401753 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.401961 kubelet[2739]: E0911 05:15:10.401776 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.407125 systemd[1]: Started cri-containerd-312e3d113bb45153b857b550186ea5e313ca3a711a266ecf1e7b580a9ba593f9.scope - libcontainer container 312e3d113bb45153b857b550186ea5e313ca3a711a266ecf1e7b580a9ba593f9. Sep 11 05:15:10.469298 containerd[1579]: time="2025-09-11T05:15:10.469237736Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-559b656695-87jhb,Uid:e3ffe175-4f19-4f9c-b627-d660a0f8b22e,Namespace:calico-system,Attempt:0,} returns sandbox id \"312e3d113bb45153b857b550186ea5e313ca3a711a266ecf1e7b580a9ba593f9\"" Sep 11 05:15:10.470180 kubelet[2739]: E0911 05:15:10.470131 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:10.470948 containerd[1579]: time="2025-09-11T05:15:10.470900115Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 11 05:15:10.481479 containerd[1579]: time="2025-09-11T05:15:10.481427845Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-rmm9d,Uid:617fd3b1-b716-4dda-8bfd-ba04f5d59b78,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:10.490343 kubelet[2739]: E0911 05:15:10.490275 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.490343 kubelet[2739]: W0911 05:15:10.490299 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.490343 kubelet[2739]: E0911 05:15:10.490331 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.490642 kubelet[2739]: E0911 05:15:10.490583 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.490642 kubelet[2739]: W0911 05:15:10.490604 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.490642 kubelet[2739]: E0911 05:15:10.490617 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.490986 kubelet[2739]: E0911 05:15:10.490901 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.490986 kubelet[2739]: W0911 05:15:10.490947 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.490986 kubelet[2739]: E0911 05:15:10.490958 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.491241 kubelet[2739]: E0911 05:15:10.491221 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.491241 kubelet[2739]: W0911 05:15:10.491233 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.491241 kubelet[2739]: E0911 05:15:10.491243 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.491593 kubelet[2739]: E0911 05:15:10.491548 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.491593 kubelet[2739]: W0911 05:15:10.491578 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.491684 kubelet[2739]: E0911 05:15:10.491615 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.491983 kubelet[2739]: E0911 05:15:10.491956 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.491983 kubelet[2739]: W0911 05:15:10.491979 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.492064 kubelet[2739]: E0911 05:15:10.491994 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.492232 kubelet[2739]: E0911 05:15:10.492207 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.492232 kubelet[2739]: W0911 05:15:10.492220 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.492296 kubelet[2739]: E0911 05:15:10.492230 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.492458 kubelet[2739]: E0911 05:15:10.492443 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.492458 kubelet[2739]: W0911 05:15:10.492455 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.492520 kubelet[2739]: E0911 05:15:10.492472 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.492702 kubelet[2739]: E0911 05:15:10.492686 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.492702 kubelet[2739]: W0911 05:15:10.492698 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.492767 kubelet[2739]: E0911 05:15:10.492708 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.492970 kubelet[2739]: E0911 05:15:10.492942 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.492970 kubelet[2739]: W0911 05:15:10.492960 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.493080 kubelet[2739]: E0911 05:15:10.492973 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.493248 kubelet[2739]: E0911 05:15:10.493215 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.493248 kubelet[2739]: W0911 05:15:10.493232 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.493248 kubelet[2739]: E0911 05:15:10.493245 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.493491 kubelet[2739]: E0911 05:15:10.493470 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.493491 kubelet[2739]: W0911 05:15:10.493482 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.493491 kubelet[2739]: E0911 05:15:10.493492 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.493938 kubelet[2739]: E0911 05:15:10.493854 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.493938 kubelet[2739]: W0911 05:15:10.493874 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.493938 kubelet[2739]: E0911 05:15:10.493889 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.494526 kubelet[2739]: E0911 05:15:10.494135 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.494526 kubelet[2739]: W0911 05:15:10.494150 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.494526 kubelet[2739]: E0911 05:15:10.494164 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.494526 kubelet[2739]: E0911 05:15:10.494421 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.494526 kubelet[2739]: W0911 05:15:10.494433 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.494526 kubelet[2739]: E0911 05:15:10.494443 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.494707 kubelet[2739]: E0911 05:15:10.494661 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.494707 kubelet[2739]: W0911 05:15:10.494672 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.494707 kubelet[2739]: E0911 05:15:10.494681 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.494898 kubelet[2739]: E0911 05:15:10.494875 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.494898 kubelet[2739]: W0911 05:15:10.494887 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.494898 kubelet[2739]: E0911 05:15:10.494896 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.497397 kubelet[2739]: E0911 05:15:10.497355 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.497397 kubelet[2739]: W0911 05:15:10.497376 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.497397 kubelet[2739]: E0911 05:15:10.497393 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.498381 kubelet[2739]: E0911 05:15:10.498362 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.498381 kubelet[2739]: W0911 05:15:10.498376 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.498495 kubelet[2739]: E0911 05:15:10.498389 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.498610 kubelet[2739]: E0911 05:15:10.498592 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.498610 kubelet[2739]: W0911 05:15:10.498605 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.498711 kubelet[2739]: E0911 05:15:10.498617 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.498938 kubelet[2739]: E0911 05:15:10.498894 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.498938 kubelet[2739]: W0911 05:15:10.498915 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.499029 kubelet[2739]: E0911 05:15:10.498943 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.499356 kubelet[2739]: E0911 05:15:10.499301 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.499356 kubelet[2739]: W0911 05:15:10.499354 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.499450 kubelet[2739]: E0911 05:15:10.499386 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.499744 kubelet[2739]: E0911 05:15:10.499723 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.499744 kubelet[2739]: W0911 05:15:10.499737 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.499830 kubelet[2739]: E0911 05:15:10.499749 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.500141 kubelet[2739]: E0911 05:15:10.500118 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.500141 kubelet[2739]: W0911 05:15:10.500134 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.500237 kubelet[2739]: E0911 05:15:10.500146 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.500772 kubelet[2739]: E0911 05:15:10.500748 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.500772 kubelet[2739]: W0911 05:15:10.500764 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.500894 kubelet[2739]: E0911 05:15:10.500777 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.510732 containerd[1579]: time="2025-09-11T05:15:10.510674935Z" level=info msg="connecting to shim 27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b" address="unix:///run/containerd/s/106216578970caa810b4db0dfd6cfb2bdd6acb7cd1832b3a055a698908e9c5d6" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:10.511949 kubelet[2739]: E0911 05:15:10.511029 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:10.511949 kubelet[2739]: W0911 05:15:10.511056 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:10.511949 kubelet[2739]: E0911 05:15:10.511084 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:10.541099 systemd[1]: Started cri-containerd-27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b.scope - libcontainer container 27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b. Sep 11 05:15:10.580671 containerd[1579]: time="2025-09-11T05:15:10.580552494Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-rmm9d,Uid:617fd3b1-b716-4dda-8bfd-ba04f5d59b78,Namespace:calico-system,Attempt:0,} returns sandbox id \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\"" Sep 11 05:15:12.578941 kubelet[2739]: E0911 05:15:12.578827 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:13.619426 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount327740840.mount: Deactivated successfully. Sep 11 05:15:14.484289 containerd[1579]: time="2025-09-11T05:15:14.484223333Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:14.485388 containerd[1579]: time="2025-09-11T05:15:14.485342518Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Sep 11 05:15:14.486444 containerd[1579]: time="2025-09-11T05:15:14.486397562Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:14.489330 containerd[1579]: time="2025-09-11T05:15:14.489259666Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:14.489771 containerd[1579]: time="2025-09-11T05:15:14.489726624Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 4.018780913s" Sep 11 05:15:14.489771 containerd[1579]: time="2025-09-11T05:15:14.489760588Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Sep 11 05:15:14.490874 containerd[1579]: time="2025-09-11T05:15:14.490846982Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 11 05:15:14.511700 containerd[1579]: time="2025-09-11T05:15:14.511643289Z" level=info msg="CreateContainer within sandbox \"312e3d113bb45153b857b550186ea5e313ca3a711a266ecf1e7b580a9ba593f9\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 11 05:15:14.520699 containerd[1579]: time="2025-09-11T05:15:14.520648779Z" level=info msg="Container 8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:14.530365 containerd[1579]: time="2025-09-11T05:15:14.530299542Z" level=info msg="CreateContainer within sandbox \"312e3d113bb45153b857b550186ea5e313ca3a711a266ecf1e7b580a9ba593f9\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9\"" Sep 11 05:15:14.530943 containerd[1579]: time="2025-09-11T05:15:14.530895734Z" level=info msg="StartContainer for \"8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9\"" Sep 11 05:15:14.532027 containerd[1579]: time="2025-09-11T05:15:14.532001183Z" level=info msg="connecting to shim 8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9" address="unix:///run/containerd/s/26248250fbad33668e7b791a16829c4d62062f12ca7a90d8e8e477a55b5b52f2" protocol=ttrpc version=3 Sep 11 05:15:14.565358 systemd[1]: Started cri-containerd-8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9.scope - libcontainer container 8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9. Sep 11 05:15:14.575446 kubelet[2739]: E0911 05:15:14.575391 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:14.639410 containerd[1579]: time="2025-09-11T05:15:14.639268044Z" level=info msg="StartContainer for \"8a4b13d710f40507ac1353c665621f411d8d6052b4fc2ce5219005f61fcc4dc9\" returns successfully" Sep 11 05:15:15.646595 kubelet[2739]: E0911 05:15:15.646530 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:15.700351 kubelet[2739]: E0911 05:15:15.700290 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.700351 kubelet[2739]: W0911 05:15:15.700326 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.700351 kubelet[2739]: E0911 05:15:15.700353 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.700613 kubelet[2739]: E0911 05:15:15.700582 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.700613 kubelet[2739]: W0911 05:15:15.700596 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.700613 kubelet[2739]: E0911 05:15:15.700607 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.700837 kubelet[2739]: E0911 05:15:15.700805 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.700837 kubelet[2739]: W0911 05:15:15.700821 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.700837 kubelet[2739]: E0911 05:15:15.700831 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.701105 kubelet[2739]: E0911 05:15:15.701077 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.701105 kubelet[2739]: W0911 05:15:15.701093 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.701105 kubelet[2739]: E0911 05:15:15.701104 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.701344 kubelet[2739]: E0911 05:15:15.701322 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.701344 kubelet[2739]: W0911 05:15:15.701337 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.701406 kubelet[2739]: E0911 05:15:15.701349 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.701568 kubelet[2739]: E0911 05:15:15.701549 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.701568 kubelet[2739]: W0911 05:15:15.701563 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.701636 kubelet[2739]: E0911 05:15:15.701574 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.701912 kubelet[2739]: E0911 05:15:15.701864 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.701912 kubelet[2739]: W0911 05:15:15.701894 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.701912 kubelet[2739]: E0911 05:15:15.701942 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.702278 kubelet[2739]: E0911 05:15:15.702255 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.702278 kubelet[2739]: W0911 05:15:15.702272 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.702353 kubelet[2739]: E0911 05:15:15.702284 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.702571 kubelet[2739]: E0911 05:15:15.702545 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.702571 kubelet[2739]: W0911 05:15:15.702564 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.702648 kubelet[2739]: E0911 05:15:15.702581 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.702835 kubelet[2739]: E0911 05:15:15.702816 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.702835 kubelet[2739]: W0911 05:15:15.702830 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.702835 kubelet[2739]: E0911 05:15:15.702841 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.703057 kubelet[2739]: E0911 05:15:15.703038 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.703057 kubelet[2739]: W0911 05:15:15.703053 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.703110 kubelet[2739]: E0911 05:15:15.703063 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.703307 kubelet[2739]: E0911 05:15:15.703288 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.703307 kubelet[2739]: W0911 05:15:15.703304 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.703360 kubelet[2739]: E0911 05:15:15.703314 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.703581 kubelet[2739]: E0911 05:15:15.703562 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.703613 kubelet[2739]: W0911 05:15:15.703579 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.703613 kubelet[2739]: E0911 05:15:15.703594 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.703859 kubelet[2739]: E0911 05:15:15.703842 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.703891 kubelet[2739]: W0911 05:15:15.703857 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.703891 kubelet[2739]: E0911 05:15:15.703870 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.704112 kubelet[2739]: E0911 05:15:15.704095 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.704112 kubelet[2739]: W0911 05:15:15.704110 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.704172 kubelet[2739]: E0911 05:15:15.704121 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.727958 kubelet[2739]: E0911 05:15:15.727887 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.727958 kubelet[2739]: W0911 05:15:15.727909 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.727958 kubelet[2739]: E0911 05:15:15.727948 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.728219 kubelet[2739]: E0911 05:15:15.728190 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.728219 kubelet[2739]: W0911 05:15:15.728212 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.728275 kubelet[2739]: E0911 05:15:15.728221 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.728439 kubelet[2739]: E0911 05:15:15.728413 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.728439 kubelet[2739]: W0911 05:15:15.728424 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.728439 kubelet[2739]: E0911 05:15:15.728433 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.728667 kubelet[2739]: E0911 05:15:15.728651 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.728667 kubelet[2739]: W0911 05:15:15.728662 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.728742 kubelet[2739]: E0911 05:15:15.728672 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.728869 kubelet[2739]: E0911 05:15:15.728855 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.728869 kubelet[2739]: W0911 05:15:15.728865 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.728869 kubelet[2739]: E0911 05:15:15.728874 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.729132 kubelet[2739]: E0911 05:15:15.729106 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.729132 kubelet[2739]: W0911 05:15:15.729127 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.729241 kubelet[2739]: E0911 05:15:15.729140 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.729492 kubelet[2739]: E0911 05:15:15.729381 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.729492 kubelet[2739]: W0911 05:15:15.729396 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.729492 kubelet[2739]: E0911 05:15:15.729418 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.730149 kubelet[2739]: E0911 05:15:15.730126 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.730149 kubelet[2739]: W0911 05:15:15.730141 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.730327 kubelet[2739]: E0911 05:15:15.730153 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.730392 kubelet[2739]: E0911 05:15:15.730357 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.730392 kubelet[2739]: W0911 05:15:15.730367 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.730392 kubelet[2739]: E0911 05:15:15.730380 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.730914 kubelet[2739]: E0911 05:15:15.730880 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.730914 kubelet[2739]: W0911 05:15:15.730896 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.730914 kubelet[2739]: E0911 05:15:15.730908 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.731288 kubelet[2739]: E0911 05:15:15.731261 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.731288 kubelet[2739]: W0911 05:15:15.731274 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.731288 kubelet[2739]: E0911 05:15:15.731283 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.731672 kubelet[2739]: E0911 05:15:15.731525 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.731672 kubelet[2739]: W0911 05:15:15.731535 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.731672 kubelet[2739]: E0911 05:15:15.731546 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.731801 kubelet[2739]: E0911 05:15:15.731784 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.731801 kubelet[2739]: W0911 05:15:15.731797 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.731893 kubelet[2739]: E0911 05:15:15.731806 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.732013 kubelet[2739]: E0911 05:15:15.732000 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.732051 kubelet[2739]: W0911 05:15:15.732013 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.732051 kubelet[2739]: E0911 05:15:15.732023 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.732214 kubelet[2739]: E0911 05:15:15.732191 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.732248 kubelet[2739]: W0911 05:15:15.732212 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.732248 kubelet[2739]: E0911 05:15:15.732225 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.732442 kubelet[2739]: E0911 05:15:15.732433 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.732442 kubelet[2739]: W0911 05:15:15.732442 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.732754 kubelet[2739]: E0911 05:15:15.732453 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.732754 kubelet[2739]: E0911 05:15:15.732715 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.732754 kubelet[2739]: W0911 05:15:15.732728 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.732754 kubelet[2739]: E0911 05:15:15.732742 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:15.733022 kubelet[2739]: E0911 05:15:15.733007 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:15.733022 kubelet[2739]: W0911 05:15:15.733018 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:15.733089 kubelet[2739]: E0911 05:15:15.733026 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.576226 kubelet[2739]: E0911 05:15:16.576145 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:16.647309 kubelet[2739]: I0911 05:15:16.647261 2739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 05:15:16.647738 kubelet[2739]: E0911 05:15:16.647652 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:16.710222 kubelet[2739]: E0911 05:15:16.710155 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.710222 kubelet[2739]: W0911 05:15:16.710206 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.710415 kubelet[2739]: E0911 05:15:16.710237 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.710614 kubelet[2739]: E0911 05:15:16.710587 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.710660 kubelet[2739]: W0911 05:15:16.710611 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.710660 kubelet[2739]: E0911 05:15:16.710639 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.710965 kubelet[2739]: E0911 05:15:16.710907 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.710965 kubelet[2739]: W0911 05:15:16.710962 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.711044 kubelet[2739]: E0911 05:15:16.710974 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.711249 kubelet[2739]: E0911 05:15:16.711223 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.711249 kubelet[2739]: W0911 05:15:16.711235 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.711249 kubelet[2739]: E0911 05:15:16.711244 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.711467 kubelet[2739]: E0911 05:15:16.711451 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.711467 kubelet[2739]: W0911 05:15:16.711462 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.711529 kubelet[2739]: E0911 05:15:16.711470 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.711736 kubelet[2739]: E0911 05:15:16.711700 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.711736 kubelet[2739]: W0911 05:15:16.711714 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.711736 kubelet[2739]: E0911 05:15:16.711727 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.712053 kubelet[2739]: E0911 05:15:16.712020 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.712053 kubelet[2739]: W0911 05:15:16.712029 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.712053 kubelet[2739]: E0911 05:15:16.712039 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.712377 kubelet[2739]: E0911 05:15:16.712339 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.712434 kubelet[2739]: W0911 05:15:16.712384 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.712434 kubelet[2739]: E0911 05:15:16.712421 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.712767 kubelet[2739]: E0911 05:15:16.712741 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.712767 kubelet[2739]: W0911 05:15:16.712756 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.712825 kubelet[2739]: E0911 05:15:16.712770 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.713045 kubelet[2739]: E0911 05:15:16.713020 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.713045 kubelet[2739]: W0911 05:15:16.713035 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.713045 kubelet[2739]: E0911 05:15:16.713046 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.713261 kubelet[2739]: E0911 05:15:16.713243 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.713261 kubelet[2739]: W0911 05:15:16.713255 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.713261 kubelet[2739]: E0911 05:15:16.713263 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.713456 kubelet[2739]: E0911 05:15:16.713436 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.713456 kubelet[2739]: W0911 05:15:16.713451 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.713539 kubelet[2739]: E0911 05:15:16.713463 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.713677 kubelet[2739]: E0911 05:15:16.713659 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.713677 kubelet[2739]: W0911 05:15:16.713670 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.713742 kubelet[2739]: E0911 05:15:16.713679 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.713864 kubelet[2739]: E0911 05:15:16.713848 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.713864 kubelet[2739]: W0911 05:15:16.713858 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.713958 kubelet[2739]: E0911 05:15:16.713866 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.714070 kubelet[2739]: E0911 05:15:16.714053 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.714070 kubelet[2739]: W0911 05:15:16.714063 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.714070 kubelet[2739]: E0911 05:15:16.714071 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.734756 kubelet[2739]: E0911 05:15:16.734684 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.734756 kubelet[2739]: W0911 05:15:16.734723 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.734756 kubelet[2739]: E0911 05:15:16.734737 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.735059 kubelet[2739]: E0911 05:15:16.735011 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.735059 kubelet[2739]: W0911 05:15:16.735020 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.735059 kubelet[2739]: E0911 05:15:16.735029 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.735325 kubelet[2739]: E0911 05:15:16.735302 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.735325 kubelet[2739]: W0911 05:15:16.735314 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.735325 kubelet[2739]: E0911 05:15:16.735323 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.735699 kubelet[2739]: E0911 05:15:16.735664 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.735699 kubelet[2739]: W0911 05:15:16.735688 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.735792 kubelet[2739]: E0911 05:15:16.735714 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.735988 kubelet[2739]: E0911 05:15:16.735965 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.735988 kubelet[2739]: W0911 05:15:16.735980 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.735988 kubelet[2739]: E0911 05:15:16.735991 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.736240 kubelet[2739]: E0911 05:15:16.736210 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.736240 kubelet[2739]: W0911 05:15:16.736222 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.736303 kubelet[2739]: E0911 05:15:16.736232 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.736634 kubelet[2739]: E0911 05:15:16.736603 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.736634 kubelet[2739]: W0911 05:15:16.736623 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.736731 kubelet[2739]: E0911 05:15:16.736639 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.736939 kubelet[2739]: E0911 05:15:16.736888 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.736939 kubelet[2739]: W0911 05:15:16.736904 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.736939 kubelet[2739]: E0911 05:15:16.736939 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.737169 kubelet[2739]: E0911 05:15:16.737146 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.737169 kubelet[2739]: W0911 05:15:16.737157 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.737259 kubelet[2739]: E0911 05:15:16.737168 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.737530 kubelet[2739]: E0911 05:15:16.737508 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.737530 kubelet[2739]: W0911 05:15:16.737521 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.737618 kubelet[2739]: E0911 05:15:16.737532 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.737861 kubelet[2739]: E0911 05:15:16.737838 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.737979 kubelet[2739]: W0911 05:15:16.737884 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.737979 kubelet[2739]: E0911 05:15:16.737898 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.738166 kubelet[2739]: E0911 05:15:16.738142 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.738166 kubelet[2739]: W0911 05:15:16.738155 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.738166 kubelet[2739]: E0911 05:15:16.738166 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.738417 kubelet[2739]: E0911 05:15:16.738402 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.738417 kubelet[2739]: W0911 05:15:16.738415 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.738487 kubelet[2739]: E0911 05:15:16.738425 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.738694 kubelet[2739]: E0911 05:15:16.738676 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.738694 kubelet[2739]: W0911 05:15:16.738690 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.738764 kubelet[2739]: E0911 05:15:16.738702 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.738983 kubelet[2739]: E0911 05:15:16.738963 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.738983 kubelet[2739]: W0911 05:15:16.738979 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.739059 kubelet[2739]: E0911 05:15:16.738993 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.740037 kubelet[2739]: E0911 05:15:16.740018 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.740037 kubelet[2739]: W0911 05:15:16.740032 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.740126 kubelet[2739]: E0911 05:15:16.740044 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.740699 kubelet[2739]: E0911 05:15:16.740677 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.740699 kubelet[2739]: W0911 05:15:16.740694 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.740781 kubelet[2739]: E0911 05:15:16.740708 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:16.740981 kubelet[2739]: E0911 05:15:16.740965 2739 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 05:15:16.740981 kubelet[2739]: W0911 05:15:16.740978 2739 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 05:15:16.741061 kubelet[2739]: E0911 05:15:16.740988 2739 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 05:15:17.922181 containerd[1579]: time="2025-09-11T05:15:17.922107838Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:18.017612 containerd[1579]: time="2025-09-11T05:15:18.017507613Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Sep 11 05:15:18.097857 containerd[1579]: time="2025-09-11T05:15:18.097771043Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:18.136549 containerd[1579]: time="2025-09-11T05:15:18.136474167Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:18.137230 containerd[1579]: time="2025-09-11T05:15:18.137190283Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 3.646312644s" Sep 11 05:15:18.137230 containerd[1579]: time="2025-09-11T05:15:18.137230269Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Sep 11 05:15:18.397878 containerd[1579]: time="2025-09-11T05:15:18.397798006Z" level=info msg="CreateContainer within sandbox \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 11 05:15:18.576272 kubelet[2739]: E0911 05:15:18.576204 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:18.783541 containerd[1579]: time="2025-09-11T05:15:18.783054042Z" level=info msg="Container 6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:18.970441 containerd[1579]: time="2025-09-11T05:15:18.970370761Z" level=info msg="CreateContainer within sandbox \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\"" Sep 11 05:15:18.971077 containerd[1579]: time="2025-09-11T05:15:18.971050539Z" level=info msg="StartContainer for \"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\"" Sep 11 05:15:18.972554 containerd[1579]: time="2025-09-11T05:15:18.972528688Z" level=info msg="connecting to shim 6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef" address="unix:///run/containerd/s/106216578970caa810b4db0dfd6cfb2bdd6acb7cd1832b3a055a698908e9c5d6" protocol=ttrpc version=3 Sep 11 05:15:18.998159 systemd[1]: Started cri-containerd-6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef.scope - libcontainer container 6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef. Sep 11 05:15:19.061432 systemd[1]: cri-containerd-6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef.scope: Deactivated successfully. Sep 11 05:15:19.063357 containerd[1579]: time="2025-09-11T05:15:19.063317457Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\" id:\"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\" pid:3439 exited_at:{seconds:1757567719 nanos:62689417}" Sep 11 05:15:19.326605 containerd[1579]: time="2025-09-11T05:15:19.326449908Z" level=info msg="received exit event container_id:\"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\" id:\"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\" pid:3439 exited_at:{seconds:1757567719 nanos:62689417}" Sep 11 05:15:19.328720 containerd[1579]: time="2025-09-11T05:15:19.328631658Z" level=info msg="StartContainer for \"6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef\" returns successfully" Sep 11 05:15:19.356260 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6be2e13784e453fcde586c8910f3855f26dafd26f8ffae74888e1647b8a114ef-rootfs.mount: Deactivated successfully. Sep 11 05:15:19.696891 kubelet[2739]: I0911 05:15:19.696646 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-559b656695-87jhb" podStartSLOduration=7.676485211 podStartE2EDuration="11.6966195s" podCreationTimestamp="2025-09-11 05:15:08 +0000 UTC" firstStartedPulling="2025-09-11 05:15:10.470606693 +0000 UTC m=+20.021091859" lastFinishedPulling="2025-09-11 05:15:14.490740972 +0000 UTC m=+24.041226148" observedRunningTime="2025-09-11 05:15:15.658496635 +0000 UTC m=+25.208981811" watchObservedRunningTime="2025-09-11 05:15:19.6966195 +0000 UTC m=+29.247104676" Sep 11 05:15:20.575732 kubelet[2739]: E0911 05:15:20.575645 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:20.661641 containerd[1579]: time="2025-09-11T05:15:20.661588971Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 11 05:15:22.575627 kubelet[2739]: E0911 05:15:22.575558 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:24.575767 kubelet[2739]: E0911 05:15:24.575669 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:26.575490 kubelet[2739]: E0911 05:15:26.575440 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:28.576481 kubelet[2739]: E0911 05:15:28.575965 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:30.575313 kubelet[2739]: E0911 05:15:30.575250 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:30.844043 containerd[1579]: time="2025-09-11T05:15:30.843869995Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:30.864293 containerd[1579]: time="2025-09-11T05:15:30.864204115Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Sep 11 05:15:30.899297 containerd[1579]: time="2025-09-11T05:15:30.899228697Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:30.957593 containerd[1579]: time="2025-09-11T05:15:30.957522760Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:30.958240 containerd[1579]: time="2025-09-11T05:15:30.958210010Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 10.296575083s" Sep 11 05:15:30.958240 containerd[1579]: time="2025-09-11T05:15:30.958238313Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Sep 11 05:15:31.273000 containerd[1579]: time="2025-09-11T05:15:31.272848143Z" level=info msg="CreateContainer within sandbox \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 11 05:15:31.773798 containerd[1579]: time="2025-09-11T05:15:31.773745538Z" level=info msg="Container 328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:31.866279 containerd[1579]: time="2025-09-11T05:15:31.866197245Z" level=info msg="CreateContainer within sandbox \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\"" Sep 11 05:15:31.867536 containerd[1579]: time="2025-09-11T05:15:31.867435290Z" level=info msg="StartContainer for \"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\"" Sep 11 05:15:31.869390 containerd[1579]: time="2025-09-11T05:15:31.869341599Z" level=info msg="connecting to shim 328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248" address="unix:///run/containerd/s/106216578970caa810b4db0dfd6cfb2bdd6acb7cd1832b3a055a698908e9c5d6" protocol=ttrpc version=3 Sep 11 05:15:31.903203 systemd[1]: Started cri-containerd-328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248.scope - libcontainer container 328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248. Sep 11 05:15:31.960220 containerd[1579]: time="2025-09-11T05:15:31.960147936Z" level=info msg="StartContainer for \"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\" returns successfully" Sep 11 05:15:32.576154 kubelet[2739]: E0911 05:15:32.576081 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:34.004397 systemd[1]: cri-containerd-328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248.scope: Deactivated successfully. Sep 11 05:15:34.004954 systemd[1]: cri-containerd-328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248.scope: Consumed 709ms CPU time, 178.5M memory peak, 2.7M read from disk, 171.3M written to disk. Sep 11 05:15:34.006003 containerd[1579]: time="2025-09-11T05:15:34.005940429Z" level=info msg="received exit event container_id:\"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\" id:\"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\" pid:3502 exited_at:{seconds:1757567734 nanos:5452223}" Sep 11 05:15:34.035719 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248-rootfs.mount: Deactivated successfully. Sep 11 05:15:34.109606 containerd[1579]: time="2025-09-11T05:15:34.109505710Z" level=info msg="TaskExit event in podsandbox handler container_id:\"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\" id:\"328894cb9b3dd2d1d38063b9a70490f17e2b165dea0284af20eccea1b9486248\" pid:3502 exited_at:{seconds:1757567734 nanos:5452223}" Sep 11 05:15:34.115947 kubelet[2739]: I0911 05:15:34.115877 2739 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 11 05:15:34.395377 systemd[1]: Created slice kubepods-burstable-pod8901f992_0b72_401d_a222_834b29554fce.slice - libcontainer container kubepods-burstable-pod8901f992_0b72_401d_a222_834b29554fce.slice. Sep 11 05:15:34.405950 systemd[1]: Created slice kubepods-besteffort-pod693f7cea_52aa_44a2_b54f_f6043f50d9a1.slice - libcontainer container kubepods-besteffort-pod693f7cea_52aa_44a2_b54f_f6043f50d9a1.slice. Sep 11 05:15:34.417790 systemd[1]: Created slice kubepods-besteffort-podea4dcec3_7636_4d67_a68d_a7ac8de150e8.slice - libcontainer container kubepods-besteffort-podea4dcec3_7636_4d67_a68d_a7ac8de150e8.slice. Sep 11 05:15:34.426365 systemd[1]: Created slice kubepods-besteffort-pod2387d1c6_af43_4e75_8af9_0dc4355b35c4.slice - libcontainer container kubepods-besteffort-pod2387d1c6_af43_4e75_8af9_0dc4355b35c4.slice. Sep 11 05:15:34.434604 systemd[1]: Created slice kubepods-besteffort-pod76d227b9_ab19_4b27_9577_4e9a40b52e32.slice - libcontainer container kubepods-besteffort-pod76d227b9_ab19_4b27_9577_4e9a40b52e32.slice. Sep 11 05:15:34.441569 systemd[1]: Created slice kubepods-besteffort-pod77c7a7a2_47be_433e_a634_fa7ffa5f9121.slice - libcontainer container kubepods-besteffort-pod77c7a7a2_47be_433e_a634_fa7ffa5f9121.slice. Sep 11 05:15:34.449555 systemd[1]: Created slice kubepods-burstable-pod2f2e53f8_1cc4_4927_8435_c7e54310ce21.slice - libcontainer container kubepods-burstable-pod2f2e53f8_1cc4_4927_8435_c7e54310ce21.slice. Sep 11 05:15:34.452874 kubelet[2739]: I0911 05:15:34.452776 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2387d1c6-af43-4e75-8af9-0dc4355b35c4-config\") pod \"goldmane-54d579b49d-ffrdj\" (UID: \"2387d1c6-af43-4e75-8af9-0dc4355b35c4\") " pod="calico-system/goldmane-54d579b49d-ffrdj" Sep 11 05:15:34.453131 kubelet[2739]: I0911 05:15:34.453084 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/2387d1c6-af43-4e75-8af9-0dc4355b35c4-goldmane-key-pair\") pod \"goldmane-54d579b49d-ffrdj\" (UID: \"2387d1c6-af43-4e75-8af9-0dc4355b35c4\") " pod="calico-system/goldmane-54d579b49d-ffrdj" Sep 11 05:15:34.453251 kubelet[2739]: I0911 05:15:34.453233 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-backend-key-pair\") pod \"whisker-6469d65f9f-ndcjs\" (UID: \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\") " pod="calico-system/whisker-6469d65f9f-ndcjs" Sep 11 05:15:34.453429 kubelet[2739]: I0911 05:15:34.453378 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2e53f8-1cc4-4927-8435-c7e54310ce21-config-volume\") pod \"coredns-674b8bbfcf-2x5tn\" (UID: \"2f2e53f8-1cc4-4927-8435-c7e54310ce21\") " pod="kube-system/coredns-674b8bbfcf-2x5tn" Sep 11 05:15:34.453566 kubelet[2739]: I0911 05:15:34.453408 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/693f7cea-52aa-44a2-b54f-f6043f50d9a1-calico-apiserver-certs\") pod \"calico-apiserver-7c7f7bddcc-jlz6c\" (UID: \"693f7cea-52aa-44a2-b54f-f6043f50d9a1\") " pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jlz6c" Sep 11 05:15:34.453566 kubelet[2739]: I0911 05:15:34.453536 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8901f992-0b72-401d-a222-834b29554fce-config-volume\") pod \"coredns-674b8bbfcf-hv2hn\" (UID: \"8901f992-0b72-401d-a222-834b29554fce\") " pod="kube-system/coredns-674b8bbfcf-hv2hn" Sep 11 05:15:34.453749 kubelet[2739]: I0911 05:15:34.453708 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlxln\" (UniqueName: \"kubernetes.io/projected/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-kube-api-access-vlxln\") pod \"whisker-6469d65f9f-ndcjs\" (UID: \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\") " pod="calico-system/whisker-6469d65f9f-ndcjs" Sep 11 05:15:34.453936 kubelet[2739]: I0911 05:15:34.453866 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwvmk\" (UniqueName: \"kubernetes.io/projected/8901f992-0b72-401d-a222-834b29554fce-kube-api-access-vwvmk\") pod \"coredns-674b8bbfcf-hv2hn\" (UID: \"8901f992-0b72-401d-a222-834b29554fce\") " pod="kube-system/coredns-674b8bbfcf-hv2hn" Sep 11 05:15:34.453936 kubelet[2739]: I0911 05:15:34.453893 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2387d1c6-af43-4e75-8af9-0dc4355b35c4-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-ffrdj\" (UID: \"2387d1c6-af43-4e75-8af9-0dc4355b35c4\") " pod="calico-system/goldmane-54d579b49d-ffrdj" Sep 11 05:15:34.454088 kubelet[2739]: I0911 05:15:34.454069 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtgx\" (UniqueName: \"kubernetes.io/projected/2387d1c6-af43-4e75-8af9-0dc4355b35c4-kube-api-access-vmtgx\") pod \"goldmane-54d579b49d-ffrdj\" (UID: \"2387d1c6-af43-4e75-8af9-0dc4355b35c4\") " pod="calico-system/goldmane-54d579b49d-ffrdj" Sep 11 05:15:34.454269 kubelet[2739]: I0911 05:15:34.454219 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/76d227b9-ab19-4b27-9577-4e9a40b52e32-calico-apiserver-certs\") pod \"calico-apiserver-7c7f7bddcc-jtjtb\" (UID: \"76d227b9-ab19-4b27-9577-4e9a40b52e32\") " pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jtjtb" Sep 11 05:15:34.454416 kubelet[2739]: I0911 05:15:34.454251 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bv9f\" (UniqueName: \"kubernetes.io/projected/2f2e53f8-1cc4-4927-8435-c7e54310ce21-kube-api-access-4bv9f\") pod \"coredns-674b8bbfcf-2x5tn\" (UID: \"2f2e53f8-1cc4-4927-8435-c7e54310ce21\") " pod="kube-system/coredns-674b8bbfcf-2x5tn" Sep 11 05:15:34.454502 kubelet[2739]: I0911 05:15:34.454398 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rvvk\" (UniqueName: \"kubernetes.io/projected/76d227b9-ab19-4b27-9577-4e9a40b52e32-kube-api-access-5rvvk\") pod \"calico-apiserver-7c7f7bddcc-jtjtb\" (UID: \"76d227b9-ab19-4b27-9577-4e9a40b52e32\") " pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jtjtb" Sep 11 05:15:34.454676 kubelet[2739]: I0911 05:15:34.454629 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77c7a7a2-47be-433e-a634-fa7ffa5f9121-tigera-ca-bundle\") pod \"calico-kube-controllers-5f7ffcc67f-f5zqv\" (UID: \"77c7a7a2-47be-433e-a634-fa7ffa5f9121\") " pod="calico-system/calico-kube-controllers-5f7ffcc67f-f5zqv" Sep 11 05:15:34.454793 kubelet[2739]: I0911 05:15:34.454775 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8k67\" (UniqueName: \"kubernetes.io/projected/693f7cea-52aa-44a2-b54f-f6043f50d9a1-kube-api-access-s8k67\") pod \"calico-apiserver-7c7f7bddcc-jlz6c\" (UID: \"693f7cea-52aa-44a2-b54f-f6043f50d9a1\") " pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jlz6c" Sep 11 05:15:34.454982 kubelet[2739]: I0911 05:15:34.454947 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-ca-bundle\") pod \"whisker-6469d65f9f-ndcjs\" (UID: \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\") " pod="calico-system/whisker-6469d65f9f-ndcjs" Sep 11 05:15:34.455076 kubelet[2739]: I0911 05:15:34.455058 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sx5v\" (UniqueName: \"kubernetes.io/projected/77c7a7a2-47be-433e-a634-fa7ffa5f9121-kube-api-access-9sx5v\") pod \"calico-kube-controllers-5f7ffcc67f-f5zqv\" (UID: \"77c7a7a2-47be-433e-a634-fa7ffa5f9121\") " pod="calico-system/calico-kube-controllers-5f7ffcc67f-f5zqv" Sep 11 05:15:34.613790 systemd[1]: Created slice kubepods-besteffort-pod01cab480_4fed_4f83_9e86_d02437c18ae7.slice - libcontainer container kubepods-besteffort-pod01cab480_4fed_4f83_9e86_d02437c18ae7.slice. Sep 11 05:15:34.617677 containerd[1579]: time="2025-09-11T05:15:34.617617682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kqvtp,Uid:01cab480-4fed-4f83-9e86-d02437c18ae7,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:34.695321 containerd[1579]: time="2025-09-11T05:15:34.694064897Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 11 05:15:34.703265 kubelet[2739]: E0911 05:15:34.703209 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:34.703906 containerd[1579]: time="2025-09-11T05:15:34.703861276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hv2hn,Uid:8901f992-0b72-401d-a222-834b29554fce,Namespace:kube-system,Attempt:0,}" Sep 11 05:15:34.711114 containerd[1579]: time="2025-09-11T05:15:34.711055311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jlz6c,Uid:693f7cea-52aa-44a2-b54f-f6043f50d9a1,Namespace:calico-apiserver,Attempt:0,}" Sep 11 05:15:34.723139 containerd[1579]: time="2025-09-11T05:15:34.723085303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6469d65f9f-ndcjs,Uid:ea4dcec3-7636-4d67-a68d-a7ac8de150e8,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:34.731765 containerd[1579]: time="2025-09-11T05:15:34.731705014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-ffrdj,Uid:2387d1c6-af43-4e75-8af9-0dc4355b35c4,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:34.739994 containerd[1579]: time="2025-09-11T05:15:34.739247703Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jtjtb,Uid:76d227b9-ab19-4b27-9577-4e9a40b52e32,Namespace:calico-apiserver,Attempt:0,}" Sep 11 05:15:34.743854 containerd[1579]: time="2025-09-11T05:15:34.743778499Z" level=error msg="Failed to destroy network for sandbox \"d47e0c81714ffac1aee3f100f864338ad37113592cde07dbf324762592270c0b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.748271 containerd[1579]: time="2025-09-11T05:15:34.748188036Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f7ffcc67f-f5zqv,Uid:77c7a7a2-47be-433e-a634-fa7ffa5f9121,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:34.756573 kubelet[2739]: E0911 05:15:34.756047 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:34.758194 containerd[1579]: time="2025-09-11T05:15:34.758148192Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2x5tn,Uid:2f2e53f8-1cc4-4927-8435-c7e54310ce21,Namespace:kube-system,Attempt:0,}" Sep 11 05:15:34.774456 containerd[1579]: time="2025-09-11T05:15:34.774384140Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kqvtp,Uid:01cab480-4fed-4f83-9e86-d02437c18ae7,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d47e0c81714ffac1aee3f100f864338ad37113592cde07dbf324762592270c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.778433 containerd[1579]: time="2025-09-11T05:15:34.777149351Z" level=error msg="Failed to destroy network for sandbox \"012976500b3dbda34057107c5db597b4bc99405c62b40f4c7e925704f6301dbd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.787891 kubelet[2739]: E0911 05:15:34.787311 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d47e0c81714ffac1aee3f100f864338ad37113592cde07dbf324762592270c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.787891 kubelet[2739]: E0911 05:15:34.787479 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d47e0c81714ffac1aee3f100f864338ad37113592cde07dbf324762592270c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:34.787891 kubelet[2739]: E0911 05:15:34.787501 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d47e0c81714ffac1aee3f100f864338ad37113592cde07dbf324762592270c0b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kqvtp" Sep 11 05:15:34.788307 kubelet[2739]: E0911 05:15:34.787560 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kqvtp_calico-system(01cab480-4fed-4f83-9e86-d02437c18ae7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kqvtp_calico-system(01cab480-4fed-4f83-9e86-d02437c18ae7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d47e0c81714ffac1aee3f100f864338ad37113592cde07dbf324762592270c0b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kqvtp" podUID="01cab480-4fed-4f83-9e86-d02437c18ae7" Sep 11 05:15:34.794496 containerd[1579]: time="2025-09-11T05:15:34.794401045Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hv2hn,Uid:8901f992-0b72-401d-a222-834b29554fce,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"012976500b3dbda34057107c5db597b4bc99405c62b40f4c7e925704f6301dbd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.796866 kubelet[2739]: E0911 05:15:34.794689 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"012976500b3dbda34057107c5db597b4bc99405c62b40f4c7e925704f6301dbd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.796866 kubelet[2739]: E0911 05:15:34.794789 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"012976500b3dbda34057107c5db597b4bc99405c62b40f4c7e925704f6301dbd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hv2hn" Sep 11 05:15:34.796866 kubelet[2739]: E0911 05:15:34.794974 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"012976500b3dbda34057107c5db597b4bc99405c62b40f4c7e925704f6301dbd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hv2hn" Sep 11 05:15:34.797061 kubelet[2739]: E0911 05:15:34.795046 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-hv2hn_kube-system(8901f992-0b72-401d-a222-834b29554fce)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-hv2hn_kube-system(8901f992-0b72-401d-a222-834b29554fce)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"012976500b3dbda34057107c5db597b4bc99405c62b40f4c7e925704f6301dbd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-hv2hn" podUID="8901f992-0b72-401d-a222-834b29554fce" Sep 11 05:15:34.830945 containerd[1579]: time="2025-09-11T05:15:34.830873080Z" level=error msg="Failed to destroy network for sandbox \"27fae65bbc617b3c01937c94483bc0c212b9c6dd4d2f1bfaf3a68465917bf44d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.834706 containerd[1579]: time="2025-09-11T05:15:34.834648937Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jlz6c,Uid:693f7cea-52aa-44a2-b54f-f6043f50d9a1,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"27fae65bbc617b3c01937c94483bc0c212b9c6dd4d2f1bfaf3a68465917bf44d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.835599 kubelet[2739]: E0911 05:15:34.835145 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27fae65bbc617b3c01937c94483bc0c212b9c6dd4d2f1bfaf3a68465917bf44d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.835599 kubelet[2739]: E0911 05:15:34.835211 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27fae65bbc617b3c01937c94483bc0c212b9c6dd4d2f1bfaf3a68465917bf44d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jlz6c" Sep 11 05:15:34.835599 kubelet[2739]: E0911 05:15:34.835231 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27fae65bbc617b3c01937c94483bc0c212b9c6dd4d2f1bfaf3a68465917bf44d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jlz6c" Sep 11 05:15:34.835717 kubelet[2739]: E0911 05:15:34.835311 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7c7f7bddcc-jlz6c_calico-apiserver(693f7cea-52aa-44a2-b54f-f6043f50d9a1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7c7f7bddcc-jlz6c_calico-apiserver(693f7cea-52aa-44a2-b54f-f6043f50d9a1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"27fae65bbc617b3c01937c94483bc0c212b9c6dd4d2f1bfaf3a68465917bf44d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jlz6c" podUID="693f7cea-52aa-44a2-b54f-f6043f50d9a1" Sep 11 05:15:34.856840 containerd[1579]: time="2025-09-11T05:15:34.856613288Z" level=error msg="Failed to destroy network for sandbox \"98f9d6643062dbd5742d84cb87f041ec5947c44c4cdb495af1b9cd51e8f11b73\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.861056 containerd[1579]: time="2025-09-11T05:15:34.860980707Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6469d65f9f-ndcjs,Uid:ea4dcec3-7636-4d67-a68d-a7ac8de150e8,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"98f9d6643062dbd5742d84cb87f041ec5947c44c4cdb495af1b9cd51e8f11b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.862419 kubelet[2739]: E0911 05:15:34.861836 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"98f9d6643062dbd5742d84cb87f041ec5947c44c4cdb495af1b9cd51e8f11b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.862419 kubelet[2739]: E0911 05:15:34.862111 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"98f9d6643062dbd5742d84cb87f041ec5947c44c4cdb495af1b9cd51e8f11b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6469d65f9f-ndcjs" Sep 11 05:15:34.862419 kubelet[2739]: E0911 05:15:34.862192 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"98f9d6643062dbd5742d84cb87f041ec5947c44c4cdb495af1b9cd51e8f11b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6469d65f9f-ndcjs" Sep 11 05:15:34.862875 kubelet[2739]: E0911 05:15:34.862345 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6469d65f9f-ndcjs_calico-system(ea4dcec3-7636-4d67-a68d-a7ac8de150e8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6469d65f9f-ndcjs_calico-system(ea4dcec3-7636-4d67-a68d-a7ac8de150e8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"98f9d6643062dbd5742d84cb87f041ec5947c44c4cdb495af1b9cd51e8f11b73\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6469d65f9f-ndcjs" podUID="ea4dcec3-7636-4d67-a68d-a7ac8de150e8" Sep 11 05:15:34.904111 containerd[1579]: time="2025-09-11T05:15:34.904044344Z" level=error msg="Failed to destroy network for sandbox \"ca078fe26cda3bde21b7d0a85c47a56daabf32a38973bd6cf5ec76f837ab1984\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.905198 containerd[1579]: time="2025-09-11T05:15:34.905150381Z" level=error msg="Failed to destroy network for sandbox \"825da08af9ad713accaacfc02eed4e86dde83d6e900deddaef1ad9928fcf74e0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.907081 containerd[1579]: time="2025-09-11T05:15:34.907034768Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-ffrdj,Uid:2387d1c6-af43-4e75-8af9-0dc4355b35c4,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca078fe26cda3bde21b7d0a85c47a56daabf32a38973bd6cf5ec76f837ab1984\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.907367 kubelet[2739]: E0911 05:15:34.907306 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca078fe26cda3bde21b7d0a85c47a56daabf32a38973bd6cf5ec76f837ab1984\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.907438 kubelet[2739]: E0911 05:15:34.907378 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca078fe26cda3bde21b7d0a85c47a56daabf32a38973bd6cf5ec76f837ab1984\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-ffrdj" Sep 11 05:15:34.907438 kubelet[2739]: E0911 05:15:34.907401 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca078fe26cda3bde21b7d0a85c47a56daabf32a38973bd6cf5ec76f837ab1984\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-ffrdj" Sep 11 05:15:34.907510 kubelet[2739]: E0911 05:15:34.907452 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-ffrdj_calico-system(2387d1c6-af43-4e75-8af9-0dc4355b35c4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-ffrdj_calico-system(2387d1c6-af43-4e75-8af9-0dc4355b35c4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ca078fe26cda3bde21b7d0a85c47a56daabf32a38973bd6cf5ec76f837ab1984\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-ffrdj" podUID="2387d1c6-af43-4e75-8af9-0dc4355b35c4" Sep 11 05:15:34.908766 containerd[1579]: time="2025-09-11T05:15:34.908696407Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f7ffcc67f-f5zqv,Uid:77c7a7a2-47be-433e-a634-fa7ffa5f9121,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"825da08af9ad713accaacfc02eed4e86dde83d6e900deddaef1ad9928fcf74e0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.909104 kubelet[2739]: E0911 05:15:34.909060 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"825da08af9ad713accaacfc02eed4e86dde83d6e900deddaef1ad9928fcf74e0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.909255 kubelet[2739]: E0911 05:15:34.909205 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"825da08af9ad713accaacfc02eed4e86dde83d6e900deddaef1ad9928fcf74e0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5f7ffcc67f-f5zqv" Sep 11 05:15:34.909255 kubelet[2739]: E0911 05:15:34.909261 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"825da08af9ad713accaacfc02eed4e86dde83d6e900deddaef1ad9928fcf74e0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5f7ffcc67f-f5zqv" Sep 11 05:15:34.909464 kubelet[2739]: E0911 05:15:34.909325 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5f7ffcc67f-f5zqv_calico-system(77c7a7a2-47be-433e-a634-fa7ffa5f9121)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5f7ffcc67f-f5zqv_calico-system(77c7a7a2-47be-433e-a634-fa7ffa5f9121)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"825da08af9ad713accaacfc02eed4e86dde83d6e900deddaef1ad9928fcf74e0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5f7ffcc67f-f5zqv" podUID="77c7a7a2-47be-433e-a634-fa7ffa5f9121" Sep 11 05:15:34.909680 containerd[1579]: time="2025-09-11T05:15:34.909536814Z" level=error msg="Failed to destroy network for sandbox \"8511b4e816b0b00ec744e91827679351cc6acbae43134111765c5e0d0e25f597\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.911122 containerd[1579]: time="2025-09-11T05:15:34.911047100Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jtjtb,Uid:76d227b9-ab19-4b27-9577-4e9a40b52e32,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8511b4e816b0b00ec744e91827679351cc6acbae43134111765c5e0d0e25f597\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.911321 containerd[1579]: time="2025-09-11T05:15:34.911100120Z" level=error msg="Failed to destroy network for sandbox \"be84bbb8550ec652f4e5d6b0bce69b91f0ba7f0b409e5ef536c1070b37825a12\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.911466 kubelet[2739]: E0911 05:15:34.911431 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8511b4e816b0b00ec744e91827679351cc6acbae43134111765c5e0d0e25f597\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.911516 kubelet[2739]: E0911 05:15:34.911477 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8511b4e816b0b00ec744e91827679351cc6acbae43134111765c5e0d0e25f597\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jtjtb" Sep 11 05:15:34.911516 kubelet[2739]: E0911 05:15:34.911497 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8511b4e816b0b00ec744e91827679351cc6acbae43134111765c5e0d0e25f597\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jtjtb" Sep 11 05:15:34.911575 kubelet[2739]: E0911 05:15:34.911534 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7c7f7bddcc-jtjtb_calico-apiserver(76d227b9-ab19-4b27-9577-4e9a40b52e32)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7c7f7bddcc-jtjtb_calico-apiserver(76d227b9-ab19-4b27-9577-4e9a40b52e32)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8511b4e816b0b00ec744e91827679351cc6acbae43134111765c5e0d0e25f597\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jtjtb" podUID="76d227b9-ab19-4b27-9577-4e9a40b52e32" Sep 11 05:15:34.912593 containerd[1579]: time="2025-09-11T05:15:34.912546965Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2x5tn,Uid:2f2e53f8-1cc4-4927-8435-c7e54310ce21,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"be84bbb8550ec652f4e5d6b0bce69b91f0ba7f0b409e5ef536c1070b37825a12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.912763 kubelet[2739]: E0911 05:15:34.912710 2739 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be84bbb8550ec652f4e5d6b0bce69b91f0ba7f0b409e5ef536c1070b37825a12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 05:15:34.912763 kubelet[2739]: E0911 05:15:34.912758 2739 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be84bbb8550ec652f4e5d6b0bce69b91f0ba7f0b409e5ef536c1070b37825a12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-2x5tn" Sep 11 05:15:34.913047 kubelet[2739]: E0911 05:15:34.912783 2739 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be84bbb8550ec652f4e5d6b0bce69b91f0ba7f0b409e5ef536c1070b37825a12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-2x5tn" Sep 11 05:15:34.913047 kubelet[2739]: E0911 05:15:34.912816 2739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-2x5tn_kube-system(2f2e53f8-1cc4-4927-8435-c7e54310ce21)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-2x5tn_kube-system(2f2e53f8-1cc4-4927-8435-c7e54310ce21)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"be84bbb8550ec652f4e5d6b0bce69b91f0ba7f0b409e5ef536c1070b37825a12\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-2x5tn" podUID="2f2e53f8-1cc4-4927-8435-c7e54310ce21" Sep 11 05:15:35.561693 kubelet[2739]: I0911 05:15:35.561608 2739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 05:15:35.562283 kubelet[2739]: E0911 05:15:35.562154 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:35.695618 kubelet[2739]: E0911 05:15:35.695557 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:45.032280 systemd[1]: Started sshd@7-10.0.0.55:22-10.0.0.1:54836.service - OpenSSH per-connection server daemon (10.0.0.1:54836). Sep 11 05:15:45.132622 sshd[3819]: Accepted publickey for core from 10.0.0.1 port 54836 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:15:45.134800 sshd-session[3819]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:15:45.143154 systemd-logind[1516]: New session 8 of user core. Sep 11 05:15:45.148217 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 11 05:15:45.339938 sshd[3822]: Connection closed by 10.0.0.1 port 54836 Sep 11 05:15:45.340723 sshd-session[3819]: pam_unix(sshd:session): session closed for user core Sep 11 05:15:45.349339 systemd[1]: sshd@7-10.0.0.55:22-10.0.0.1:54836.service: Deactivated successfully. Sep 11 05:15:45.352430 systemd[1]: session-8.scope: Deactivated successfully. Sep 11 05:15:45.357008 systemd-logind[1516]: Session 8 logged out. Waiting for processes to exit. Sep 11 05:15:45.359895 systemd-logind[1516]: Removed session 8. Sep 11 05:15:45.473721 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2484550686.mount: Deactivated successfully. Sep 11 05:15:46.036195 containerd[1579]: time="2025-09-11T05:15:46.036141709Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:46.037647 containerd[1579]: time="2025-09-11T05:15:46.037594142Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Sep 11 05:15:46.038867 containerd[1579]: time="2025-09-11T05:15:46.038823431Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:46.043889 containerd[1579]: time="2025-09-11T05:15:46.043842539Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:46.044381 containerd[1579]: time="2025-09-11T05:15:46.044355502Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 11.35022929s" Sep 11 05:15:46.044427 containerd[1579]: time="2025-09-11T05:15:46.044382745Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Sep 11 05:15:46.078137 containerd[1579]: time="2025-09-11T05:15:46.078076592Z" level=info msg="CreateContainer within sandbox \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 11 05:15:46.095737 containerd[1579]: time="2025-09-11T05:15:46.095660298Z" level=info msg="Container 688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:46.113253 containerd[1579]: time="2025-09-11T05:15:46.113175972Z" level=info msg="CreateContainer within sandbox \"27be69cf98b2aec5f3ad45922a542fb9841e1101088e44ae3d107d6e8b16d50b\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0\"" Sep 11 05:15:46.113900 containerd[1579]: time="2025-09-11T05:15:46.113857231Z" level=info msg="StartContainer for \"688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0\"" Sep 11 05:15:46.115808 containerd[1579]: time="2025-09-11T05:15:46.115769533Z" level=info msg="connecting to shim 688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0" address="unix:///run/containerd/s/106216578970caa810b4db0dfd6cfb2bdd6acb7cd1832b3a055a698908e9c5d6" protocol=ttrpc version=3 Sep 11 05:15:46.141186 systemd[1]: Started cri-containerd-688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0.scope - libcontainer container 688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0. Sep 11 05:15:46.219752 containerd[1579]: time="2025-09-11T05:15:46.219705935Z" level=info msg="StartContainer for \"688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0\" returns successfully" Sep 11 05:15:46.292597 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 11 05:15:46.292746 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 11 05:15:46.438307 kubelet[2739]: I0911 05:15:46.438239 2739 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-backend-key-pair\") pod \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\" (UID: \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\") " Sep 11 05:15:46.438307 kubelet[2739]: I0911 05:15:46.438289 2739 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlxln\" (UniqueName: \"kubernetes.io/projected/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-kube-api-access-vlxln\") pod \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\" (UID: \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\") " Sep 11 05:15:46.438307 kubelet[2739]: I0911 05:15:46.438325 2739 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-ca-bundle\") pod \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\" (UID: \"ea4dcec3-7636-4d67-a68d-a7ac8de150e8\") " Sep 11 05:15:46.439973 kubelet[2739]: I0911 05:15:46.439909 2739 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "ea4dcec3-7636-4d67-a68d-a7ac8de150e8" (UID: "ea4dcec3-7636-4d67-a68d-a7ac8de150e8"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 11 05:15:46.443603 kubelet[2739]: I0911 05:15:46.443547 2739 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-kube-api-access-vlxln" (OuterVolumeSpecName: "kube-api-access-vlxln") pod "ea4dcec3-7636-4d67-a68d-a7ac8de150e8" (UID: "ea4dcec3-7636-4d67-a68d-a7ac8de150e8"). InnerVolumeSpecName "kube-api-access-vlxln". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 11 05:15:46.443603 kubelet[2739]: I0911 05:15:46.443595 2739 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "ea4dcec3-7636-4d67-a68d-a7ac8de150e8" (UID: "ea4dcec3-7636-4d67-a68d-a7ac8de150e8"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 11 05:15:46.473541 systemd[1]: var-lib-kubelet-pods-ea4dcec3\x2d7636\x2d4d67\x2da68d\x2da7ac8de150e8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dvlxln.mount: Deactivated successfully. Sep 11 05:15:46.473673 systemd[1]: var-lib-kubelet-pods-ea4dcec3\x2d7636\x2d4d67\x2da68d\x2da7ac8de150e8-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 11 05:15:46.538947 kubelet[2739]: I0911 05:15:46.538887 2739 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Sep 11 05:15:46.538947 kubelet[2739]: I0911 05:15:46.538950 2739 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Sep 11 05:15:46.538947 kubelet[2739]: I0911 05:15:46.538961 2739 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vlxln\" (UniqueName: \"kubernetes.io/projected/ea4dcec3-7636-4d67-a68d-a7ac8de150e8-kube-api-access-vlxln\") on node \"localhost\" DevicePath \"\"" Sep 11 05:15:46.575574 kubelet[2739]: E0911 05:15:46.575529 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:46.576269 containerd[1579]: time="2025-09-11T05:15:46.576176940Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hv2hn,Uid:8901f992-0b72-401d-a222-834b29554fce,Namespace:kube-system,Attempt:0,}" Sep 11 05:15:46.583202 systemd[1]: Removed slice kubepods-besteffort-podea4dcec3_7636_4d67_a68d_a7ac8de150e8.slice - libcontainer container kubepods-besteffort-podea4dcec3_7636_4d67_a68d_a7ac8de150e8.slice. Sep 11 05:15:46.760645 kubelet[2739]: I0911 05:15:46.758280 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-rmm9d" podStartSLOduration=1.294700849 podStartE2EDuration="36.758259796s" podCreationTimestamp="2025-09-11 05:15:10 +0000 UTC" firstStartedPulling="2025-09-11 05:15:10.582036327 +0000 UTC m=+20.132521503" lastFinishedPulling="2025-09-11 05:15:46.045595273 +0000 UTC m=+55.596080450" observedRunningTime="2025-09-11 05:15:46.757259269 +0000 UTC m=+56.307744445" watchObservedRunningTime="2025-09-11 05:15:46.758259796 +0000 UTC m=+56.308744983" Sep 11 05:15:46.821721 systemd[1]: Created slice kubepods-besteffort-podb7799dc1_4b31_42e8_bbc2_e3e6fdd29fb0.slice - libcontainer container kubepods-besteffort-podb7799dc1_4b31_42e8_bbc2_e3e6fdd29fb0.slice. Sep 11 05:15:46.902835 systemd-networkd[1477]: cali6edf2aa00c5: Link UP Sep 11 05:15:46.903399 systemd-networkd[1477]: cali6edf2aa00c5: Gained carrier Sep 11 05:15:46.922792 containerd[1579]: 2025-09-11 05:15:46.643 [INFO][3905] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 05:15:46.922792 containerd[1579]: 2025-09-11 05:15:46.677 [INFO][3905] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0 coredns-674b8bbfcf- kube-system 8901f992-0b72-401d-a222-834b29554fce 917 0 2025-09-11 05:14:55 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-hv2hn eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali6edf2aa00c5 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-" Sep 11 05:15:46.922792 containerd[1579]: 2025-09-11 05:15:46.677 [INFO][3905] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.922792 containerd[1579]: 2025-09-11 05:15:46.856 [INFO][3919] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" HandleID="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Workload="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.857 [INFO][3919] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" HandleID="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Workload="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000e1810), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-hv2hn", "timestamp":"2025-09-11 05:15:46.856875505 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.857 [INFO][3919] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.857 [INFO][3919] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.857 [INFO][3919] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.865 [INFO][3919] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" host="localhost" Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.872 [INFO][3919] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.876 [INFO][3919] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.878 [INFO][3919] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.879 [INFO][3919] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:46.923446 containerd[1579]: 2025-09-11 05:15:46.879 [INFO][3919] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" host="localhost" Sep 11 05:15:46.923675 containerd[1579]: 2025-09-11 05:15:46.881 [INFO][3919] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703 Sep 11 05:15:46.923675 containerd[1579]: 2025-09-11 05:15:46.885 [INFO][3919] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" host="localhost" Sep 11 05:15:46.923675 containerd[1579]: 2025-09-11 05:15:46.890 [INFO][3919] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" host="localhost" Sep 11 05:15:46.923675 containerd[1579]: 2025-09-11 05:15:46.890 [INFO][3919] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" host="localhost" Sep 11 05:15:46.923675 containerd[1579]: 2025-09-11 05:15:46.890 [INFO][3919] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:46.923675 containerd[1579]: 2025-09-11 05:15:46.890 [INFO][3919] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" HandleID="k8s-pod-network.19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Workload="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.923802 containerd[1579]: 2025-09-11 05:15:46.894 [INFO][3905] cni-plugin/k8s.go 418: Populated endpoint ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"8901f992-0b72-401d-a222-834b29554fce", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 14, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-hv2hn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6edf2aa00c5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:46.923893 containerd[1579]: 2025-09-11 05:15:46.894 [INFO][3905] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.923893 containerd[1579]: 2025-09-11 05:15:46.894 [INFO][3905] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6edf2aa00c5 ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.923893 containerd[1579]: 2025-09-11 05:15:46.903 [INFO][3905] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.923991 containerd[1579]: 2025-09-11 05:15:46.903 [INFO][3905] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"8901f992-0b72-401d-a222-834b29554fce", ResourceVersion:"917", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 14, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703", Pod:"coredns-674b8bbfcf-hv2hn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6edf2aa00c5", MAC:"42:22:48:bf:ed:4b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:46.923991 containerd[1579]: 2025-09-11 05:15:46.919 [INFO][3905] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" Namespace="kube-system" Pod="coredns-674b8bbfcf-hv2hn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--hv2hn-eth0" Sep 11 05:15:46.944255 kubelet[2739]: I0911 05:15:46.944167 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlpbw\" (UniqueName: \"kubernetes.io/projected/b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0-kube-api-access-nlpbw\") pod \"whisker-766966cf54-mvzfv\" (UID: \"b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0\") " pod="calico-system/whisker-766966cf54-mvzfv" Sep 11 05:15:46.944255 kubelet[2739]: I0911 05:15:46.944235 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0-whisker-backend-key-pair\") pod \"whisker-766966cf54-mvzfv\" (UID: \"b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0\") " pod="calico-system/whisker-766966cf54-mvzfv" Sep 11 05:15:46.944255 kubelet[2739]: I0911 05:15:46.944251 2739 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0-whisker-ca-bundle\") pod \"whisker-766966cf54-mvzfv\" (UID: \"b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0\") " pod="calico-system/whisker-766966cf54-mvzfv" Sep 11 05:15:47.002737 containerd[1579]: time="2025-09-11T05:15:47.002683123Z" level=info msg="connecting to shim 19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703" address="unix:///run/containerd/s/4f3b5340bbd0e689100a33107711960df430ca7cfea473fa567de2bee9ecce20" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:47.028120 systemd[1]: Started cri-containerd-19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703.scope - libcontainer container 19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703. Sep 11 05:15:47.045467 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:47.125911 containerd[1579]: time="2025-09-11T05:15:47.125847688Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-766966cf54-mvzfv,Uid:b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:47.305059 containerd[1579]: time="2025-09-11T05:15:47.305011777Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hv2hn,Uid:8901f992-0b72-401d-a222-834b29554fce,Namespace:kube-system,Attempt:0,} returns sandbox id \"19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703\"" Sep 11 05:15:47.305865 kubelet[2739]: E0911 05:15:47.305691 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:47.319182 containerd[1579]: time="2025-09-11T05:15:47.319115150Z" level=info msg="CreateContainer within sandbox \"19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 11 05:15:47.340847 containerd[1579]: time="2025-09-11T05:15:47.340780171Z" level=info msg="Container f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:47.348607 containerd[1579]: time="2025-09-11T05:15:47.348564540Z" level=info msg="CreateContainer within sandbox \"19ffcc20a8aff3f0971aeb782c8e18c0fab4d6ac31ecb21c04025cf7a5213703\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6\"" Sep 11 05:15:47.349248 containerd[1579]: time="2025-09-11T05:15:47.349203457Z" level=info msg="StartContainer for \"f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6\"" Sep 11 05:15:47.350542 containerd[1579]: time="2025-09-11T05:15:47.350273617Z" level=info msg="connecting to shim f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6" address="unix:///run/containerd/s/4f3b5340bbd0e689100a33107711960df430ca7cfea473fa567de2bee9ecce20" protocol=ttrpc version=3 Sep 11 05:15:47.370051 systemd[1]: Started cri-containerd-f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6.scope - libcontainer container f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6. Sep 11 05:15:47.415305 containerd[1579]: time="2025-09-11T05:15:47.415160222Z" level=info msg="StartContainer for \"f3e27aa8e0d98897dc57a4f6cd27202b899bac423c58200162566653f34627a6\" returns successfully" Sep 11 05:15:47.432799 systemd-networkd[1477]: cali6935852b3ba: Link UP Sep 11 05:15:47.433084 systemd-networkd[1477]: cali6935852b3ba: Gained carrier Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.340 [INFO][3982] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.354 [INFO][3982] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--766966cf54--mvzfv-eth0 whisker-766966cf54- calico-system b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0 1045 0 2025-09-11 05:15:46 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:766966cf54 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-766966cf54-mvzfv eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali6935852b3ba [] [] }} ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.354 [INFO][3982] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.390 [INFO][4005] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" HandleID="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Workload="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.390 [INFO][4005] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" HandleID="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Workload="localhost-k8s-whisker--766966cf54--mvzfv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000138e30), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-766966cf54-mvzfv", "timestamp":"2025-09-11 05:15:47.390113828 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.390 [INFO][4005] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.390 [INFO][4005] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.390 [INFO][4005] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.400 [INFO][4005] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.405 [INFO][4005] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.409 [INFO][4005] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.411 [INFO][4005] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.413 [INFO][4005] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.413 [INFO][4005] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.415 [INFO][4005] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01 Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.421 [INFO][4005] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.427 [INFO][4005] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.427 [INFO][4005] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" host="localhost" Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.427 [INFO][4005] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:47.451884 containerd[1579]: 2025-09-11 05:15:47.427 [INFO][4005] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" HandleID="k8s-pod-network.a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Workload="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.452533 containerd[1579]: 2025-09-11 05:15:47.430 [INFO][3982] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--766966cf54--mvzfv-eth0", GenerateName:"whisker-766966cf54-", Namespace:"calico-system", SelfLink:"", UID:"b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"766966cf54", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-766966cf54-mvzfv", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali6935852b3ba", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:47.452533 containerd[1579]: 2025-09-11 05:15:47.431 [INFO][3982] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.452533 containerd[1579]: 2025-09-11 05:15:47.431 [INFO][3982] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6935852b3ba ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.452533 containerd[1579]: 2025-09-11 05:15:47.433 [INFO][3982] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.452533 containerd[1579]: 2025-09-11 05:15:47.433 [INFO][3982] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--766966cf54--mvzfv-eth0", GenerateName:"whisker-766966cf54-", Namespace:"calico-system", SelfLink:"", UID:"b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0", ResourceVersion:"1045", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"766966cf54", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01", Pod:"whisker-766966cf54-mvzfv", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali6935852b3ba", MAC:"d2:f0:55:12:da:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:47.452533 containerd[1579]: 2025-09-11 05:15:47.448 [INFO][3982] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" Namespace="calico-system" Pod="whisker-766966cf54-mvzfv" WorkloadEndpoint="localhost-k8s-whisker--766966cf54--mvzfv-eth0" Sep 11 05:15:47.481891 containerd[1579]: time="2025-09-11T05:15:47.481840857Z" level=info msg="connecting to shim a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01" address="unix:///run/containerd/s/27017eecb37f39fe3f606468c269498da9f10fac0c52f05b8a3e1377ad22d3b4" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:47.509146 systemd[1]: Started cri-containerd-a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01.scope - libcontainer container a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01. Sep 11 05:15:47.524476 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:47.577253 containerd[1579]: time="2025-09-11T05:15:47.576208544Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f7ffcc67f-f5zqv,Uid:77c7a7a2-47be-433e-a634-fa7ffa5f9121,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:47.577253 containerd[1579]: time="2025-09-11T05:15:47.576496201Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kqvtp,Uid:01cab480-4fed-4f83-9e86-d02437c18ae7,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:47.578216 containerd[1579]: time="2025-09-11T05:15:47.578124811Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jlz6c,Uid:693f7cea-52aa-44a2-b54f-f6043f50d9a1,Namespace:calico-apiserver,Attempt:0,}" Sep 11 05:15:47.578297 containerd[1579]: time="2025-09-11T05:15:47.578227731Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-ffrdj,Uid:2387d1c6-af43-4e75-8af9-0dc4355b35c4,Namespace:calico-system,Attempt:0,}" Sep 11 05:15:47.743729 containerd[1579]: time="2025-09-11T05:15:47.743657139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-766966cf54-mvzfv,Uid:b7799dc1-4b31-42e8-bbc2-e3e6fdd29fb0,Namespace:calico-system,Attempt:0,} returns sandbox id \"a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01\"" Sep 11 05:15:47.749721 containerd[1579]: time="2025-09-11T05:15:47.749527655Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 11 05:15:47.760333 kubelet[2739]: E0911 05:15:47.760283 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:47.927316 kubelet[2739]: I0911 05:15:47.926632 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-hv2hn" podStartSLOduration=52.926605507 podStartE2EDuration="52.926605507s" podCreationTimestamp="2025-09-11 05:14:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 05:15:47.923706729 +0000 UTC m=+57.474191895" watchObservedRunningTime="2025-09-11 05:15:47.926605507 +0000 UTC m=+57.477090683" Sep 11 05:15:48.177239 systemd-networkd[1477]: cali769e18dd787: Link UP Sep 11 05:15:48.177599 systemd-networkd[1477]: cali769e18dd787: Gained carrier Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.006 [INFO][4186] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.026 [INFO][4186] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0 calico-kube-controllers-5f7ffcc67f- calico-system 77c7a7a2-47be-433e-a634-fa7ffa5f9121 925 0 2025-09-11 05:15:10 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5f7ffcc67f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5f7ffcc67f-f5zqv eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali769e18dd787 [] [] }} ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.026 [INFO][4186] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.118 [INFO][4267] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" HandleID="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Workload="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.118 [INFO][4267] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" HandleID="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Workload="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fae0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5f7ffcc67f-f5zqv", "timestamp":"2025-09-11 05:15:48.118153544 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.118 [INFO][4267] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.118 [INFO][4267] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.118 [INFO][4267] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.130 [INFO][4267] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.135 [INFO][4267] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.141 [INFO][4267] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.144 [INFO][4267] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.147 [INFO][4267] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.147 [INFO][4267] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.149 [INFO][4267] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.154 [INFO][4267] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.165 [INFO][4267] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.165 [INFO][4267] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" host="localhost" Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.165 [INFO][4267] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:48.199552 containerd[1579]: 2025-09-11 05:15:48.165 [INFO][4267] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" HandleID="k8s-pod-network.114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Workload="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.200664 containerd[1579]: 2025-09-11 05:15:48.173 [INFO][4186] cni-plugin/k8s.go 418: Populated endpoint ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0", GenerateName:"calico-kube-controllers-5f7ffcc67f-", Namespace:"calico-system", SelfLink:"", UID:"77c7a7a2-47be-433e-a634-fa7ffa5f9121", ResourceVersion:"925", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5f7ffcc67f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5f7ffcc67f-f5zqv", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali769e18dd787", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.200664 containerd[1579]: 2025-09-11 05:15:48.173 [INFO][4186] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.200664 containerd[1579]: 2025-09-11 05:15:48.173 [INFO][4186] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali769e18dd787 ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.200664 containerd[1579]: 2025-09-11 05:15:48.179 [INFO][4186] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.200664 containerd[1579]: 2025-09-11 05:15:48.181 [INFO][4186] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0", GenerateName:"calico-kube-controllers-5f7ffcc67f-", Namespace:"calico-system", SelfLink:"", UID:"77c7a7a2-47be-433e-a634-fa7ffa5f9121", ResourceVersion:"925", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5f7ffcc67f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af", Pod:"calico-kube-controllers-5f7ffcc67f-f5zqv", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali769e18dd787", MAC:"7e:ef:5b:0f:dd:3e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.200664 containerd[1579]: 2025-09-11 05:15:48.194 [INFO][4186] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" Namespace="calico-system" Pod="calico-kube-controllers-5f7ffcc67f-f5zqv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5f7ffcc67f--f5zqv-eth0" Sep 11 05:15:48.242451 containerd[1579]: time="2025-09-11T05:15:48.242393238Z" level=info msg="connecting to shim 114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af" address="unix:///run/containerd/s/8cdc5e65e3becca6922ff1acda167753ae66eb4767ae782690d8c6900982e2d6" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:48.283480 systemd-networkd[1477]: cali57187c1c470: Link UP Sep 11 05:15:48.284255 systemd-networkd[1477]: cali57187c1c470: Gained carrier Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.032 [INFO][4196] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.057 [INFO][4196] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0 calico-apiserver-7c7f7bddcc- calico-apiserver 693f7cea-52aa-44a2-b54f-f6043f50d9a1 920 0 2025-09-11 05:15:06 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7c7f7bddcc projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7c7f7bddcc-jlz6c eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali57187c1c470 [] [] }} ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.058 [INFO][4196] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.127 [INFO][4276] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" HandleID="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Workload="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.127 [INFO][4276] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" HandleID="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Workload="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a4e30), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7c7f7bddcc-jlz6c", "timestamp":"2025-09-11 05:15:48.127038387 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.127 [INFO][4276] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.165 [INFO][4276] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.165 [INFO][4276] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.234 [INFO][4276] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.245 [INFO][4276] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.251 [INFO][4276] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.253 [INFO][4276] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.256 [INFO][4276] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.256 [INFO][4276] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.257 [INFO][4276] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23 Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.264 [INFO][4276] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.272 [INFO][4276] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.273 [INFO][4276] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" host="localhost" Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.273 [INFO][4276] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:48.305033 containerd[1579]: 2025-09-11 05:15:48.273 [INFO][4276] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" HandleID="k8s-pod-network.6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Workload="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.305863 containerd[1579]: 2025-09-11 05:15:48.277 [INFO][4196] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0", GenerateName:"calico-apiserver-7c7f7bddcc-", Namespace:"calico-apiserver", SelfLink:"", UID:"693f7cea-52aa-44a2-b54f-f6043f50d9a1", ResourceVersion:"920", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c7f7bddcc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7c7f7bddcc-jlz6c", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali57187c1c470", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.305863 containerd[1579]: 2025-09-11 05:15:48.277 [INFO][4196] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.305863 containerd[1579]: 2025-09-11 05:15:48.278 [INFO][4196] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali57187c1c470 ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.305863 containerd[1579]: 2025-09-11 05:15:48.283 [INFO][4196] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.305863 containerd[1579]: 2025-09-11 05:15:48.284 [INFO][4196] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0", GenerateName:"calico-apiserver-7c7f7bddcc-", Namespace:"calico-apiserver", SelfLink:"", UID:"693f7cea-52aa-44a2-b54f-f6043f50d9a1", ResourceVersion:"920", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c7f7bddcc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23", Pod:"calico-apiserver-7c7f7bddcc-jlz6c", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali57187c1c470", MAC:"3a:5f:9e:70:3c:b2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.305863 containerd[1579]: 2025-09-11 05:15:48.301 [INFO][4196] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jlz6c" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jlz6c-eth0" Sep 11 05:15:48.319232 systemd[1]: Started cri-containerd-114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af.scope - libcontainer container 114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af. Sep 11 05:15:48.336787 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:48.406893 containerd[1579]: time="2025-09-11T05:15:48.406846715Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f7ffcc67f-f5zqv,Uid:77c7a7a2-47be-433e-a634-fa7ffa5f9121,Namespace:calico-system,Attempt:0,} returns sandbox id \"114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af\"" Sep 11 05:15:48.408718 systemd-networkd[1477]: cali6edf2aa00c5: Gained IPv6LL Sep 11 05:15:48.424616 containerd[1579]: time="2025-09-11T05:15:48.424449168Z" level=info msg="connecting to shim 6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23" address="unix:///run/containerd/s/517b9d81451056ab7afe0ef6f5e3547be42af48a7bdaff0e955c44c8746bbf13" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:48.471628 systemd-networkd[1477]: calif10c46973a4: Link UP Sep 11 05:15:48.472842 systemd-networkd[1477]: calif10c46973a4: Gained carrier Sep 11 05:15:48.472861 systemd[1]: Started cri-containerd-6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23.scope - libcontainer container 6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23. Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.016 [INFO][4194] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.042 [INFO][4194] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--kqvtp-eth0 csi-node-driver- calico-system 01cab480-4fed-4f83-9e86-d02437c18ae7 781 0 2025-09-11 05:15:10 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-kqvtp eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calif10c46973a4 [] [] }} ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.042 [INFO][4194] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.143 [INFO][4271] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" HandleID="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Workload="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.144 [INFO][4271] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" HandleID="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Workload="localhost-k8s-csi--node--driver--kqvtp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c6fd0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-kqvtp", "timestamp":"2025-09-11 05:15:48.143224173 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.144 [INFO][4271] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.273 [INFO][4271] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.273 [INFO][4271] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.397 [INFO][4271] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.403 [INFO][4271] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.408 [INFO][4271] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.412 [INFO][4271] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.415 [INFO][4271] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.415 [INFO][4271] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.417 [INFO][4271] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0 Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.429 [INFO][4271] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.438 [INFO][4271] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.439 [INFO][4271] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" host="localhost" Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.439 [INFO][4271] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:48.493211 containerd[1579]: 2025-09-11 05:15:48.439 [INFO][4271] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" HandleID="k8s-pod-network.5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Workload="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.494281 containerd[1579]: 2025-09-11 05:15:48.452 [INFO][4194] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--kqvtp-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"01cab480-4fed-4f83-9e86-d02437c18ae7", ResourceVersion:"781", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-kqvtp", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif10c46973a4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.494281 containerd[1579]: 2025-09-11 05:15:48.454 [INFO][4194] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.494281 containerd[1579]: 2025-09-11 05:15:48.454 [INFO][4194] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif10c46973a4 ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.494281 containerd[1579]: 2025-09-11 05:15:48.472 [INFO][4194] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.494281 containerd[1579]: 2025-09-11 05:15:48.472 [INFO][4194] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--kqvtp-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"01cab480-4fed-4f83-9e86-d02437c18ae7", ResourceVersion:"781", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0", Pod:"csi-node-driver-kqvtp", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif10c46973a4", MAC:"46:18:79:1d:28:01", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.494281 containerd[1579]: 2025-09-11 05:15:48.488 [INFO][4194] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" Namespace="calico-system" Pod="csi-node-driver-kqvtp" WorkloadEndpoint="localhost-k8s-csi--node--driver--kqvtp-eth0" Sep 11 05:15:48.527518 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:48.536722 systemd-networkd[1477]: vxlan.calico: Link UP Sep 11 05:15:48.536743 systemd-networkd[1477]: vxlan.calico: Gained carrier Sep 11 05:15:48.539477 containerd[1579]: time="2025-09-11T05:15:48.537870573Z" level=info msg="connecting to shim 5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0" address="unix:///run/containerd/s/541753aef0cdd660312e6e72b8f24d80c7f64f9abe898c97cedb51977075cb31" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:48.573168 systemd-networkd[1477]: cali27a3ad93e7d: Link UP Sep 11 05:15:48.573755 systemd-networkd[1477]: cali27a3ad93e7d: Gained carrier Sep 11 05:15:48.584844 containerd[1579]: time="2025-09-11T05:15:48.584793431Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jtjtb,Uid:76d227b9-ab19-4b27-9577-4e9a40b52e32,Namespace:calico-apiserver,Attempt:0,}" Sep 11 05:15:48.592523 kubelet[2739]: I0911 05:15:48.592330 2739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea4dcec3-7636-4d67-a68d-a7ac8de150e8" path="/var/lib/kubelet/pods/ea4dcec3-7636-4d67-a68d-a7ac8de150e8/volumes" Sep 11 05:15:48.612276 systemd[1]: Started cri-containerd-5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0.scope - libcontainer container 5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0. Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.011 [INFO][4209] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.042 [INFO][4209] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--54d579b49d--ffrdj-eth0 goldmane-54d579b49d- calico-system 2387d1c6-af43-4e75-8af9-0dc4355b35c4 924 0 2025-09-11 05:15:09 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-54d579b49d-ffrdj eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali27a3ad93e7d [] [] }} ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.042 [INFO][4209] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.151 [INFO][4269] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" HandleID="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Workload="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.151 [INFO][4269] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" HandleID="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Workload="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000353e10), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-54d579b49d-ffrdj", "timestamp":"2025-09-11 05:15:48.151141256 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.151 [INFO][4269] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.439 [INFO][4269] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.440 [INFO][4269] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.458 [INFO][4269] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.506 [INFO][4269] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.514 [INFO][4269] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.517 [INFO][4269] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.523 [INFO][4269] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.523 [INFO][4269] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.525 [INFO][4269] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69 Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.534 [INFO][4269] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.547 [INFO][4269] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.548 [INFO][4269] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" host="localhost" Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.548 [INFO][4269] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:48.621959 containerd[1579]: 2025-09-11 05:15:48.548 [INFO][4269] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" HandleID="k8s-pod-network.a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Workload="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.623277 containerd[1579]: 2025-09-11 05:15:48.554 [INFO][4209] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--ffrdj-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"2387d1c6-af43-4e75-8af9-0dc4355b35c4", ResourceVersion:"924", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-54d579b49d-ffrdj", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali27a3ad93e7d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.623277 containerd[1579]: 2025-09-11 05:15:48.555 [INFO][4209] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.623277 containerd[1579]: 2025-09-11 05:15:48.555 [INFO][4209] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali27a3ad93e7d ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.623277 containerd[1579]: 2025-09-11 05:15:48.575 [INFO][4209] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.623277 containerd[1579]: 2025-09-11 05:15:48.579 [INFO][4209] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--ffrdj-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"2387d1c6-af43-4e75-8af9-0dc4355b35c4", ResourceVersion:"924", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69", Pod:"goldmane-54d579b49d-ffrdj", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali27a3ad93e7d", MAC:"b6:58:6c:8a:71:7d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.623277 containerd[1579]: 2025-09-11 05:15:48.596 [INFO][4209] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" Namespace="calico-system" Pod="goldmane-54d579b49d-ffrdj" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--ffrdj-eth0" Sep 11 05:15:48.631972 containerd[1579]: time="2025-09-11T05:15:48.631396831Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jlz6c,Uid:693f7cea-52aa-44a2-b54f-f6043f50d9a1,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23\"" Sep 11 05:15:48.649362 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:48.673670 containerd[1579]: time="2025-09-11T05:15:48.673602201Z" level=info msg="connecting to shim a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69" address="unix:///run/containerd/s/5d65869ad6c294b99230ab41a371228f895cbb6a3b0a64365e6095be448a6d9b" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:48.677402 containerd[1579]: time="2025-09-11T05:15:48.677333143Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kqvtp,Uid:01cab480-4fed-4f83-9e86-d02437c18ae7,Namespace:calico-system,Attempt:0,} returns sandbox id \"5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0\"" Sep 11 05:15:48.708111 systemd[1]: Started cri-containerd-a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69.scope - libcontainer container a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69. Sep 11 05:15:48.731111 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:48.764779 systemd-networkd[1477]: cali59001cead71: Link UP Sep 11 05:15:48.772746 systemd-networkd[1477]: cali59001cead71: Gained carrier Sep 11 05:15:48.774444 kubelet[2739]: E0911 05:15:48.774416 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:48.785319 containerd[1579]: time="2025-09-11T05:15:48.785278361Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-ffrdj,Uid:2387d1c6-af43-4e75-8af9-0dc4355b35c4,Namespace:calico-system,Attempt:0,} returns sandbox id \"a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69\"" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.667 [INFO][4484] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0 calico-apiserver-7c7f7bddcc- calico-apiserver 76d227b9-ab19-4b27-9577-4e9a40b52e32 927 0 2025-09-11 05:15:06 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7c7f7bddcc projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7c7f7bddcc-jtjtb eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali59001cead71 [] [] }} ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.668 [INFO][4484] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.708 [INFO][4543] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" HandleID="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Workload="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.708 [INFO][4543] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" HandleID="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Workload="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c7730), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7c7f7bddcc-jtjtb", "timestamp":"2025-09-11 05:15:48.708323995 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.708 [INFO][4543] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.708 [INFO][4543] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.708 [INFO][4543] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.716 [INFO][4543] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.721 [INFO][4543] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.726 [INFO][4543] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.727 [INFO][4543] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.730 [INFO][4543] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.730 [INFO][4543] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.733 [INFO][4543] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.739 [INFO][4543] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.746 [INFO][4543] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.746 [INFO][4543] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" host="localhost" Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.746 [INFO][4543] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:48.800368 containerd[1579]: 2025-09-11 05:15:48.746 [INFO][4543] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" HandleID="k8s-pod-network.1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Workload="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.801045 containerd[1579]: 2025-09-11 05:15:48.750 [INFO][4484] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0", GenerateName:"calico-apiserver-7c7f7bddcc-", Namespace:"calico-apiserver", SelfLink:"", UID:"76d227b9-ab19-4b27-9577-4e9a40b52e32", ResourceVersion:"927", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c7f7bddcc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7c7f7bddcc-jtjtb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali59001cead71", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.801045 containerd[1579]: 2025-09-11 05:15:48.753 [INFO][4484] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.801045 containerd[1579]: 2025-09-11 05:15:48.753 [INFO][4484] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali59001cead71 ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.801045 containerd[1579]: 2025-09-11 05:15:48.774 [INFO][4484] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.801045 containerd[1579]: 2025-09-11 05:15:48.776 [INFO][4484] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0", GenerateName:"calico-apiserver-7c7f7bddcc-", Namespace:"calico-apiserver", SelfLink:"", UID:"76d227b9-ab19-4b27-9577-4e9a40b52e32", ResourceVersion:"927", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 15, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c7f7bddcc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f", Pod:"calico-apiserver-7c7f7bddcc-jtjtb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali59001cead71", MAC:"de:00:e2:17:0b:7d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:48.801045 containerd[1579]: 2025-09-11 05:15:48.794 [INFO][4484] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" Namespace="calico-apiserver" Pod="calico-apiserver-7c7f7bddcc-jtjtb" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c7f7bddcc--jtjtb-eth0" Sep 11 05:15:48.875336 containerd[1579]: time="2025-09-11T05:15:48.875287502Z" level=info msg="connecting to shim 1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f" address="unix:///run/containerd/s/9fc08af2bd796d184679ed84169bbccbd48c87176cb0ddf0a3e30070ebcf2629" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:48.911210 systemd[1]: Started cri-containerd-1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f.scope - libcontainer container 1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f. Sep 11 05:15:48.933023 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:48.974142 containerd[1579]: time="2025-09-11T05:15:48.974085200Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c7f7bddcc-jtjtb,Uid:76d227b9-ab19-4b27-9577-4e9a40b52e32,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f\"" Sep 11 05:15:49.367182 systemd-networkd[1477]: cali6935852b3ba: Gained IPv6LL Sep 11 05:15:49.432137 systemd-networkd[1477]: cali769e18dd787: Gained IPv6LL Sep 11 05:15:49.523171 containerd[1579]: time="2025-09-11T05:15:49.523093977Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:49.524024 containerd[1579]: time="2025-09-11T05:15:49.523989416Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Sep 11 05:15:49.527090 containerd[1579]: time="2025-09-11T05:15:49.527027113Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:49.529131 containerd[1579]: time="2025-09-11T05:15:49.529080569Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:49.529682 containerd[1579]: time="2025-09-11T05:15:49.529626954Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 1.780059834s" Sep 11 05:15:49.529682 containerd[1579]: time="2025-09-11T05:15:49.529672994Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Sep 11 05:15:49.531131 containerd[1579]: time="2025-09-11T05:15:49.530860878Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 11 05:15:49.535848 containerd[1579]: time="2025-09-11T05:15:49.535803814Z" level=info msg="CreateContainer within sandbox \"a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 11 05:15:49.545826 containerd[1579]: time="2025-09-11T05:15:49.545764853Z" level=info msg="Container cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:49.555643 containerd[1579]: time="2025-09-11T05:15:49.555594858Z" level=info msg="CreateContainer within sandbox \"a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd\"" Sep 11 05:15:49.556404 containerd[1579]: time="2025-09-11T05:15:49.556369734Z" level=info msg="StartContainer for \"cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd\"" Sep 11 05:15:49.557894 containerd[1579]: time="2025-09-11T05:15:49.557852538Z" level=info msg="connecting to shim cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd" address="unix:///run/containerd/s/27017eecb37f39fe3f606468c269498da9f10fac0c52f05b8a3e1377ad22d3b4" protocol=ttrpc version=3 Sep 11 05:15:49.576397 kubelet[2739]: E0911 05:15:49.576363 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:49.576870 containerd[1579]: time="2025-09-11T05:15:49.576839609Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2x5tn,Uid:2f2e53f8-1cc4-4927-8435-c7e54310ce21,Namespace:kube-system,Attempt:0,}" Sep 11 05:15:49.579077 systemd[1]: Started cri-containerd-cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd.scope - libcontainer container cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd. Sep 11 05:15:49.623222 systemd-networkd[1477]: calif10c46973a4: Gained IPv6LL Sep 11 05:15:49.688327 containerd[1579]: time="2025-09-11T05:15:49.687379992Z" level=info msg="StartContainer for \"cfcb409c35ebaf6fc775d298d7dfb87b1a37451f783c3252ce6a5617def470bd\" returns successfully" Sep 11 05:15:49.721768 systemd-networkd[1477]: cali9d9235e9c7b: Link UP Sep 11 05:15:49.722532 systemd-networkd[1477]: cali9d9235e9c7b: Gained carrier Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.633 [INFO][4701] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0 coredns-674b8bbfcf- kube-system 2f2e53f8-1cc4-4927-8435-c7e54310ce21 926 0 2025-09-11 05:14:55 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-2x5tn eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali9d9235e9c7b [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.633 [INFO][4701] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.666 [INFO][4728] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" HandleID="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Workload="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.667 [INFO][4728] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" HandleID="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Workload="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003af0b0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-2x5tn", "timestamp":"2025-09-11 05:15:49.666826606 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.667 [INFO][4728] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.667 [INFO][4728] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.667 [INFO][4728] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.677 [INFO][4728] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.688 [INFO][4728] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.694 [INFO][4728] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.697 [INFO][4728] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.700 [INFO][4728] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.700 [INFO][4728] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.702 [INFO][4728] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16 Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.706 [INFO][4728] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.714 [INFO][4728] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.714 [INFO][4728] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" host="localhost" Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.714 [INFO][4728] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 05:15:49.741962 containerd[1579]: 2025-09-11 05:15:49.714 [INFO][4728] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" HandleID="k8s-pod-network.b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Workload="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.742546 containerd[1579]: 2025-09-11 05:15:49.718 [INFO][4701] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"2f2e53f8-1cc4-4927-8435-c7e54310ce21", ResourceVersion:"926", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 14, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-2x5tn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali9d9235e9c7b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:49.742546 containerd[1579]: 2025-09-11 05:15:49.718 [INFO][4701] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.742546 containerd[1579]: 2025-09-11 05:15:49.718 [INFO][4701] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9d9235e9c7b ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.742546 containerd[1579]: 2025-09-11 05:15:49.723 [INFO][4701] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.742546 containerd[1579]: 2025-09-11 05:15:49.724 [INFO][4701] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"2f2e53f8-1cc4-4927-8435-c7e54310ce21", ResourceVersion:"926", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 5, 14, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16", Pod:"coredns-674b8bbfcf-2x5tn", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali9d9235e9c7b", MAC:"9e:ac:e9:6d:6b:71", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 05:15:49.742546 containerd[1579]: 2025-09-11 05:15:49.734 [INFO][4701] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" Namespace="kube-system" Pod="coredns-674b8bbfcf-2x5tn" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2x5tn-eth0" Sep 11 05:15:49.816662 systemd-networkd[1477]: vxlan.calico: Gained IPv6LL Sep 11 05:15:49.826177 kubelet[2739]: E0911 05:15:49.826142 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:49.835869 containerd[1579]: time="2025-09-11T05:15:49.835803150Z" level=info msg="connecting to shim b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16" address="unix:///run/containerd/s/3baa4221eeacee989382a9c1b15b6c346bdf062154ffa2272f9195735790f52a" namespace=k8s.io protocol=ttrpc version=3 Sep 11 05:15:49.865162 systemd[1]: Started cri-containerd-b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16.scope - libcontainer container b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16. Sep 11 05:15:49.882366 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 05:15:49.922102 containerd[1579]: time="2025-09-11T05:15:49.922033051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2x5tn,Uid:2f2e53f8-1cc4-4927-8435-c7e54310ce21,Namespace:kube-system,Attempt:0,} returns sandbox id \"b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16\"" Sep 11 05:15:49.923039 kubelet[2739]: E0911 05:15:49.922992 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:49.930352 containerd[1579]: time="2025-09-11T05:15:49.930294858Z" level=info msg="CreateContainer within sandbox \"b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 11 05:15:49.941329 containerd[1579]: time="2025-09-11T05:15:49.941244104Z" level=info msg="Container 0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:49.955156 containerd[1579]: time="2025-09-11T05:15:49.955083954Z" level=info msg="CreateContainer within sandbox \"b042f4b02982d3ef6d9e1d8bd0c43d5e3e4afd9d188a75a1cad1adfbfddd1b16\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7\"" Sep 11 05:15:49.955863 containerd[1579]: time="2025-09-11T05:15:49.955810056Z" level=info msg="StartContainer for \"0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7\"" Sep 11 05:15:49.957262 containerd[1579]: time="2025-09-11T05:15:49.957203578Z" level=info msg="connecting to shim 0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7" address="unix:///run/containerd/s/3baa4221eeacee989382a9c1b15b6c346bdf062154ffa2272f9195735790f52a" protocol=ttrpc version=3 Sep 11 05:15:49.985102 systemd[1]: Started cri-containerd-0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7.scope - libcontainer container 0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7. Sep 11 05:15:50.008248 systemd-networkd[1477]: cali59001cead71: Gained IPv6LL Sep 11 05:15:50.022779 containerd[1579]: time="2025-09-11T05:15:50.022719369Z" level=info msg="StartContainer for \"0e8a114e609627365754a6f2c0be77544690f17c6877e7334db3bff4c898abf7\" returns successfully" Sep 11 05:15:50.327145 systemd-networkd[1477]: cali57187c1c470: Gained IPv6LL Sep 11 05:15:50.351723 systemd[1]: Started sshd@8-10.0.0.55:22-10.0.0.1:60274.service - OpenSSH per-connection server daemon (10.0.0.1:60274). Sep 11 05:15:50.443216 sshd[4834]: Accepted publickey for core from 10.0.0.1 port 60274 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:15:50.445419 sshd-session[4834]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:15:50.451657 systemd-logind[1516]: New session 9 of user core. Sep 11 05:15:50.460318 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 11 05:15:50.520137 systemd-networkd[1477]: cali27a3ad93e7d: Gained IPv6LL Sep 11 05:15:50.647445 sshd[4838]: Connection closed by 10.0.0.1 port 60274 Sep 11 05:15:50.647732 sshd-session[4834]: pam_unix(sshd:session): session closed for user core Sep 11 05:15:50.652660 systemd[1]: sshd@8-10.0.0.55:22-10.0.0.1:60274.service: Deactivated successfully. Sep 11 05:15:50.655169 systemd[1]: session-9.scope: Deactivated successfully. Sep 11 05:15:50.656370 systemd-logind[1516]: Session 9 logged out. Waiting for processes to exit. Sep 11 05:15:50.657769 systemd-logind[1516]: Removed session 9. Sep 11 05:15:50.830781 kubelet[2739]: E0911 05:15:50.830552 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:50.831325 kubelet[2739]: E0911 05:15:50.830813 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:50.936883 kubelet[2739]: I0911 05:15:50.936683 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-2x5tn" podStartSLOduration=55.936654336 podStartE2EDuration="55.936654336s" podCreationTimestamp="2025-09-11 05:14:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 05:15:50.899334484 +0000 UTC m=+60.449819660" watchObservedRunningTime="2025-09-11 05:15:50.936654336 +0000 UTC m=+60.487139512" Sep 11 05:15:51.735880 systemd-networkd[1477]: cali9d9235e9c7b: Gained IPv6LL Sep 11 05:15:51.833502 kubelet[2739]: E0911 05:15:51.833446 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:52.835578 kubelet[2739]: E0911 05:15:52.835523 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:53.304001 containerd[1579]: time="2025-09-11T05:15:53.303861123Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:53.306340 containerd[1579]: time="2025-09-11T05:15:53.305745652Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Sep 11 05:15:53.308097 containerd[1579]: time="2025-09-11T05:15:53.308018469Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:53.310775 containerd[1579]: time="2025-09-11T05:15:53.310649975Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:53.311522 containerd[1579]: time="2025-09-11T05:15:53.311461438Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 3.780558821s" Sep 11 05:15:53.311522 containerd[1579]: time="2025-09-11T05:15:53.311502567Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Sep 11 05:15:53.312871 containerd[1579]: time="2025-09-11T05:15:53.312826507Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 11 05:15:53.335642 containerd[1579]: time="2025-09-11T05:15:53.335555524Z" level=info msg="CreateContainer within sandbox \"114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 11 05:15:53.348229 containerd[1579]: time="2025-09-11T05:15:53.348158130Z" level=info msg="Container 08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:53.360449 containerd[1579]: time="2025-09-11T05:15:53.360355819Z" level=info msg="CreateContainer within sandbox \"114c7de8fadd13dcb02ebd6ef7f612040055bec35016660d0beb96af18f668af\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec\"" Sep 11 05:15:53.362956 containerd[1579]: time="2025-09-11T05:15:53.361972512Z" level=info msg="StartContainer for \"08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec\"" Sep 11 05:15:53.365624 containerd[1579]: time="2025-09-11T05:15:53.365573827Z" level=info msg="connecting to shim 08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec" address="unix:///run/containerd/s/8cdc5e65e3becca6922ff1acda167753ae66eb4767ae782690d8c6900982e2d6" protocol=ttrpc version=3 Sep 11 05:15:53.419561 systemd[1]: Started cri-containerd-08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec.scope - libcontainer container 08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec. Sep 11 05:15:53.689372 containerd[1579]: time="2025-09-11T05:15:53.689214728Z" level=info msg="StartContainer for \"08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec\" returns successfully" Sep 11 05:15:53.856999 kubelet[2739]: I0911 05:15:53.856901 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5f7ffcc67f-f5zqv" podStartSLOduration=38.956831144 podStartE2EDuration="43.856881406s" podCreationTimestamp="2025-09-11 05:15:10 +0000 UTC" firstStartedPulling="2025-09-11 05:15:48.41248333 +0000 UTC m=+57.962968506" lastFinishedPulling="2025-09-11 05:15:53.312533582 +0000 UTC m=+62.863018768" observedRunningTime="2025-09-11 05:15:53.856437281 +0000 UTC m=+63.406922457" watchObservedRunningTime="2025-09-11 05:15:53.856881406 +0000 UTC m=+63.407366582" Sep 11 05:15:53.938391 containerd[1579]: time="2025-09-11T05:15:53.938326466Z" level=info msg="TaskExit event in podsandbox handler container_id:\"08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec\" id:\"87db280f5ee2cc98425ff94a8db60c1c504a8d2e8ee1b2d20604565fd63ea2d4\" pid:4924 exited_at:{seconds:1757567753 nanos:937139050}" Sep 11 05:15:55.663259 systemd[1]: Started sshd@9-10.0.0.55:22-10.0.0.1:60290.service - OpenSSH per-connection server daemon (10.0.0.1:60290). Sep 11 05:15:55.746568 sshd[4943]: Accepted publickey for core from 10.0.0.1 port 60290 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:15:55.748546 sshd-session[4943]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:15:55.754094 systemd-logind[1516]: New session 10 of user core. Sep 11 05:15:55.764173 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 11 05:15:55.944397 sshd[4946]: Connection closed by 10.0.0.1 port 60290 Sep 11 05:15:55.944664 sshd-session[4943]: pam_unix(sshd:session): session closed for user core Sep 11 05:15:55.949270 systemd[1]: sshd@9-10.0.0.55:22-10.0.0.1:60290.service: Deactivated successfully. Sep 11 05:15:55.951278 systemd[1]: session-10.scope: Deactivated successfully. Sep 11 05:15:55.952173 systemd-logind[1516]: Session 10 logged out. Waiting for processes to exit. Sep 11 05:15:55.953536 systemd-logind[1516]: Removed session 10. Sep 11 05:15:57.459042 containerd[1579]: time="2025-09-11T05:15:57.458880276Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:57.460017 containerd[1579]: time="2025-09-11T05:15:57.459939269Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Sep 11 05:15:57.461743 containerd[1579]: time="2025-09-11T05:15:57.461672097Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:57.465197 containerd[1579]: time="2025-09-11T05:15:57.465080884Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:57.466483 containerd[1579]: time="2025-09-11T05:15:57.466416759Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 4.153551378s" Sep 11 05:15:57.466581 containerd[1579]: time="2025-09-11T05:15:57.466498747Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 11 05:15:57.468580 containerd[1579]: time="2025-09-11T05:15:57.468534707Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 11 05:15:57.473809 containerd[1579]: time="2025-09-11T05:15:57.473746336Z" level=info msg="CreateContainer within sandbox \"6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 11 05:15:57.487153 containerd[1579]: time="2025-09-11T05:15:57.487065476Z" level=info msg="Container 635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:57.502418 containerd[1579]: time="2025-09-11T05:15:57.502287951Z" level=info msg="CreateContainer within sandbox \"6416f4004d17e064717f4754a43e6d7f20c88d29425f828f1a397076f7bb6f23\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84\"" Sep 11 05:15:57.503532 containerd[1579]: time="2025-09-11T05:15:57.503428131Z" level=info msg="StartContainer for \"635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84\"" Sep 11 05:15:57.505894 containerd[1579]: time="2025-09-11T05:15:57.505787753Z" level=info msg="connecting to shim 635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84" address="unix:///run/containerd/s/517b9d81451056ab7afe0ef6f5e3547be42af48a7bdaff0e955c44c8746bbf13" protocol=ttrpc version=3 Sep 11 05:15:57.540269 systemd[1]: Started cri-containerd-635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84.scope - libcontainer container 635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84. Sep 11 05:15:57.575822 kubelet[2739]: E0911 05:15:57.575663 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:15:57.644198 containerd[1579]: time="2025-09-11T05:15:57.644136264Z" level=info msg="StartContainer for \"635e862d46107628dbd62889d2677e42acfaf879d997bec8b5656416bb99da84\" returns successfully" Sep 11 05:15:58.851754 kubelet[2739]: I0911 05:15:58.851700 2739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 05:15:59.399211 containerd[1579]: time="2025-09-11T05:15:59.399003979Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:59.401401 containerd[1579]: time="2025-09-11T05:15:59.401355149Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Sep 11 05:15:59.403479 containerd[1579]: time="2025-09-11T05:15:59.403349665Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:59.407264 containerd[1579]: time="2025-09-11T05:15:59.407206443Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:15:59.408154 containerd[1579]: time="2025-09-11T05:15:59.408103173Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 1.939529561s" Sep 11 05:15:59.408154 containerd[1579]: time="2025-09-11T05:15:59.408137509Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Sep 11 05:15:59.410235 containerd[1579]: time="2025-09-11T05:15:59.410176139Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 11 05:15:59.416305 containerd[1579]: time="2025-09-11T05:15:59.416244500Z" level=info msg="CreateContainer within sandbox \"5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 11 05:15:59.445232 containerd[1579]: time="2025-09-11T05:15:59.443516112Z" level=info msg="Container a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:15:59.449456 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1436555874.mount: Deactivated successfully. Sep 11 05:15:59.460638 containerd[1579]: time="2025-09-11T05:15:59.460567150Z" level=info msg="CreateContainer within sandbox \"5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b\"" Sep 11 05:15:59.461519 containerd[1579]: time="2025-09-11T05:15:59.461468268Z" level=info msg="StartContainer for \"a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b\"" Sep 11 05:15:59.463324 containerd[1579]: time="2025-09-11T05:15:59.463287338Z" level=info msg="connecting to shim a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b" address="unix:///run/containerd/s/541753aef0cdd660312e6e72b8f24d80c7f64f9abe898c97cedb51977075cb31" protocol=ttrpc version=3 Sep 11 05:15:59.499273 systemd[1]: Started cri-containerd-a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b.scope - libcontainer container a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b. Sep 11 05:15:59.565764 containerd[1579]: time="2025-09-11T05:15:59.565694556Z" level=info msg="StartContainer for \"a2e8e9987212cda3b0f836874de58637496cd66232524a1a677bc1d480a3040b\" returns successfully" Sep 11 05:16:00.964488 systemd[1]: Started sshd@10-10.0.0.55:22-10.0.0.1:45924.service - OpenSSH per-connection server daemon (10.0.0.1:45924). Sep 11 05:16:01.048649 sshd[5052]: Accepted publickey for core from 10.0.0.1 port 45924 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:01.051023 sshd-session[5052]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:01.057100 systemd-logind[1516]: New session 11 of user core. Sep 11 05:16:01.065285 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 11 05:16:01.254014 sshd[5055]: Connection closed by 10.0.0.1 port 45924 Sep 11 05:16:01.256253 sshd-session[5052]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:01.269618 systemd[1]: sshd@10-10.0.0.55:22-10.0.0.1:45924.service: Deactivated successfully. Sep 11 05:16:01.273162 systemd[1]: session-11.scope: Deactivated successfully. Sep 11 05:16:01.275182 systemd-logind[1516]: Session 11 logged out. Waiting for processes to exit. Sep 11 05:16:01.281412 systemd[1]: Started sshd@11-10.0.0.55:22-10.0.0.1:45936.service - OpenSSH per-connection server daemon (10.0.0.1:45936). Sep 11 05:16:01.282321 systemd-logind[1516]: Removed session 11. Sep 11 05:16:01.357590 sshd[5070]: Accepted publickey for core from 10.0.0.1 port 45936 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:01.359350 sshd-session[5070]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:01.364963 systemd-logind[1516]: New session 12 of user core. Sep 11 05:16:01.372247 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 11 05:16:01.634625 sshd[5073]: Connection closed by 10.0.0.1 port 45936 Sep 11 05:16:01.635026 sshd-session[5070]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:01.643575 systemd[1]: sshd@11-10.0.0.55:22-10.0.0.1:45936.service: Deactivated successfully. Sep 11 05:16:01.647232 systemd[1]: session-12.scope: Deactivated successfully. Sep 11 05:16:01.649106 systemd-logind[1516]: Session 12 logged out. Waiting for processes to exit. Sep 11 05:16:01.653118 systemd[1]: Started sshd@12-10.0.0.55:22-10.0.0.1:45950.service - OpenSSH per-connection server daemon (10.0.0.1:45950). Sep 11 05:16:01.654679 systemd-logind[1516]: Removed session 12. Sep 11 05:16:01.718575 sshd[5086]: Accepted publickey for core from 10.0.0.1 port 45950 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:01.720485 sshd-session[5086]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:01.727360 systemd-logind[1516]: New session 13 of user core. Sep 11 05:16:01.737857 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 11 05:16:01.889351 sshd[5089]: Connection closed by 10.0.0.1 port 45950 Sep 11 05:16:01.889692 sshd-session[5086]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:01.897152 systemd[1]: sshd@12-10.0.0.55:22-10.0.0.1:45950.service: Deactivated successfully. Sep 11 05:16:01.899787 systemd[1]: session-13.scope: Deactivated successfully. Sep 11 05:16:01.900978 systemd-logind[1516]: Session 13 logged out. Waiting for processes to exit. Sep 11 05:16:01.903264 systemd-logind[1516]: Removed session 13. Sep 11 05:16:03.422602 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2466803531.mount: Deactivated successfully. Sep 11 05:16:04.209073 containerd[1579]: time="2025-09-11T05:16:04.209005296Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:04.216189 containerd[1579]: time="2025-09-11T05:16:04.216092294Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Sep 11 05:16:04.217612 containerd[1579]: time="2025-09-11T05:16:04.217583257Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:04.219701 containerd[1579]: time="2025-09-11T05:16:04.219639660Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:04.220256 containerd[1579]: time="2025-09-11T05:16:04.220200032Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 4.809982713s" Sep 11 05:16:04.220256 containerd[1579]: time="2025-09-11T05:16:04.220248074Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Sep 11 05:16:04.221322 containerd[1579]: time="2025-09-11T05:16:04.221291600Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 11 05:16:04.226192 containerd[1579]: time="2025-09-11T05:16:04.226163049Z" level=info msg="CreateContainer within sandbox \"a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 11 05:16:04.237569 containerd[1579]: time="2025-09-11T05:16:04.237495449Z" level=info msg="Container d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:16:04.284363 containerd[1579]: time="2025-09-11T05:16:04.284307212Z" level=info msg="CreateContainer within sandbox \"a281daeb287a31a45736d8de3eb8290faa150419c60bc8aaadcc53669eac6d69\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\"" Sep 11 05:16:04.284884 containerd[1579]: time="2025-09-11T05:16:04.284856774Z" level=info msg="StartContainer for \"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\"" Sep 11 05:16:04.286119 containerd[1579]: time="2025-09-11T05:16:04.286084372Z" level=info msg="connecting to shim d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f" address="unix:///run/containerd/s/5d65869ad6c294b99230ab41a371228f895cbb6a3b0a64365e6095be448a6d9b" protocol=ttrpc version=3 Sep 11 05:16:04.312097 systemd[1]: Started cri-containerd-d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f.scope - libcontainer container d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f. Sep 11 05:16:04.576405 kubelet[2739]: E0911 05:16:04.576348 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:16:04.675741 containerd[1579]: time="2025-09-11T05:16:04.675687621Z" level=info msg="StartContainer for \"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\" returns successfully" Sep 11 05:16:04.954802 containerd[1579]: time="2025-09-11T05:16:04.954651943Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\" id:\"9da8fcb23b6b7a1807f575a05f592a1ba2cc0629464bb0f3df149ace53174355\" pid:5161 exit_status:1 exited_at:{seconds:1757567764 nanos:954258119}" Sep 11 05:16:05.041193 kubelet[2739]: I0911 05:16:05.041102 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jlz6c" podStartSLOduration=50.206307808 podStartE2EDuration="59.041076059s" podCreationTimestamp="2025-09-11 05:15:06 +0000 UTC" firstStartedPulling="2025-09-11 05:15:48.633338544 +0000 UTC m=+58.183823720" lastFinishedPulling="2025-09-11 05:15:57.468106795 +0000 UTC m=+67.018591971" observedRunningTime="2025-09-11 05:15:57.868395807 +0000 UTC m=+67.418880983" watchObservedRunningTime="2025-09-11 05:16:05.041076059 +0000 UTC m=+74.591561255" Sep 11 05:16:05.041415 kubelet[2739]: I0911 05:16:05.041294 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-ffrdj" podStartSLOduration=40.614116831 podStartE2EDuration="56.041289046s" podCreationTimestamp="2025-09-11 05:15:09 +0000 UTC" firstStartedPulling="2025-09-11 05:15:48.793916057 +0000 UTC m=+58.344401233" lastFinishedPulling="2025-09-11 05:16:04.221088272 +0000 UTC m=+73.771573448" observedRunningTime="2025-09-11 05:16:05.040541717 +0000 UTC m=+74.591026883" watchObservedRunningTime="2025-09-11 05:16:05.041289046 +0000 UTC m=+74.591774222" Sep 11 05:16:05.705198 containerd[1579]: time="2025-09-11T05:16:05.705127948Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:05.857950 containerd[1579]: time="2025-09-11T05:16:05.857859788Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Sep 11 05:16:05.860369 containerd[1579]: time="2025-09-11T05:16:05.860326615Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 1.639002643s" Sep 11 05:16:05.860442 containerd[1579]: time="2025-09-11T05:16:05.860376040Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 11 05:16:05.861325 containerd[1579]: time="2025-09-11T05:16:05.861293974Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 11 05:16:06.160854 containerd[1579]: time="2025-09-11T05:16:06.160796170Z" level=info msg="CreateContainer within sandbox \"1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 11 05:16:06.258445 containerd[1579]: time="2025-09-11T05:16:06.258392696Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\" id:\"980d75ff1a3020bc11d3f8662d78a04da647930d9fbc448f035632001c214d35\" pid:5188 exit_status:1 exited_at:{seconds:1757567766 nanos:256464361}" Sep 11 05:16:06.904256 systemd[1]: Started sshd@13-10.0.0.55:22-10.0.0.1:45952.service - OpenSSH per-connection server daemon (10.0.0.1:45952). Sep 11 05:16:06.981702 sshd[5201]: Accepted publickey for core from 10.0.0.1 port 45952 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:06.983890 sshd-session[5201]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:06.990463 systemd-logind[1516]: New session 14 of user core. Sep 11 05:16:06.998313 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 11 05:16:07.152439 sshd[5204]: Connection closed by 10.0.0.1 port 45952 Sep 11 05:16:07.152760 sshd-session[5201]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:07.157859 systemd-logind[1516]: Session 14 logged out. Waiting for processes to exit. Sep 11 05:16:07.158287 systemd[1]: sshd@13-10.0.0.55:22-10.0.0.1:45952.service: Deactivated successfully. Sep 11 05:16:07.160776 systemd[1]: session-14.scope: Deactivated successfully. Sep 11 05:16:07.162752 systemd-logind[1516]: Removed session 14. Sep 11 05:16:07.528129 containerd[1579]: time="2025-09-11T05:16:07.527833067Z" level=info msg="Container a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:16:07.563174 containerd[1579]: time="2025-09-11T05:16:07.563114136Z" level=info msg="CreateContainer within sandbox \"1d69beb6d9f5d7177f337bca1ca368fe38a0630cf387adc5b262a2caa9b9690f\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a\"" Sep 11 05:16:07.565650 containerd[1579]: time="2025-09-11T05:16:07.563854529Z" level=info msg="StartContainer for \"a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a\"" Sep 11 05:16:07.565650 containerd[1579]: time="2025-09-11T05:16:07.565345989Z" level=info msg="connecting to shim a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a" address="unix:///run/containerd/s/9fc08af2bd796d184679ed84169bbccbd48c87176cb0ddf0a3e30070ebcf2629" protocol=ttrpc version=3 Sep 11 05:16:07.576517 kubelet[2739]: E0911 05:16:07.576478 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:16:07.599205 systemd[1]: Started cri-containerd-a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a.scope - libcontainer container a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a. Sep 11 05:16:07.835317 containerd[1579]: time="2025-09-11T05:16:07.835264970Z" level=info msg="StartContainer for \"a140ce0bff77b3e49e3bdfb4d9e7d0231963325931c1d05ce06a312fa726ba0a\" returns successfully" Sep 11 05:16:08.053931 kubelet[2739]: I0911 05:16:08.053830 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7c7f7bddcc-jtjtb" podStartSLOduration=45.168640492 podStartE2EDuration="1m2.053810066s" podCreationTimestamp="2025-09-11 05:15:06 +0000 UTC" firstStartedPulling="2025-09-11 05:15:48.9760354 +0000 UTC m=+58.526520576" lastFinishedPulling="2025-09-11 05:16:05.861204974 +0000 UTC m=+75.411690150" observedRunningTime="2025-09-11 05:16:08.05229869 +0000 UTC m=+77.602783866" watchObservedRunningTime="2025-09-11 05:16:08.053810066 +0000 UTC m=+77.604295262" Sep 11 05:16:09.884738 kubelet[2739]: I0911 05:16:09.884688 2739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 05:16:10.501559 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2262705151.mount: Deactivated successfully. Sep 11 05:16:11.323696 containerd[1579]: time="2025-09-11T05:16:11.323536831Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:11.325334 containerd[1579]: time="2025-09-11T05:16:11.325198761Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Sep 11 05:16:11.329434 containerd[1579]: time="2025-09-11T05:16:11.329384613Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:11.335112 containerd[1579]: time="2025-09-11T05:16:11.335028376Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:11.338400 containerd[1579]: time="2025-09-11T05:16:11.338361362Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 5.477041318s" Sep 11 05:16:11.338400 containerd[1579]: time="2025-09-11T05:16:11.338399405Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Sep 11 05:16:11.341225 containerd[1579]: time="2025-09-11T05:16:11.341191860Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 11 05:16:11.351063 containerd[1579]: time="2025-09-11T05:16:11.350995086Z" level=info msg="CreateContainer within sandbox \"a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 11 05:16:11.371305 containerd[1579]: time="2025-09-11T05:16:11.371214115Z" level=info msg="Container 04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:16:11.397914 containerd[1579]: time="2025-09-11T05:16:11.397824261Z" level=info msg="CreateContainer within sandbox \"a1f3c1aad09cc5d7527b87127bf6473ea0a363e327a8626d73a2708b9cb48e01\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496\"" Sep 11 05:16:11.399368 containerd[1579]: time="2025-09-11T05:16:11.399259880Z" level=info msg="StartContainer for \"04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496\"" Sep 11 05:16:11.401206 containerd[1579]: time="2025-09-11T05:16:11.401161055Z" level=info msg="connecting to shim 04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496" address="unix:///run/containerd/s/27017eecb37f39fe3f606468c269498da9f10fac0c52f05b8a3e1377ad22d3b4" protocol=ttrpc version=3 Sep 11 05:16:11.445433 systemd[1]: Started cri-containerd-04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496.scope - libcontainer container 04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496. Sep 11 05:16:11.520014 containerd[1579]: time="2025-09-11T05:16:11.519946527Z" level=info msg="StartContainer for \"04bb94b1b1f9c725394d3d3c8a4a88f36766cc0d583d0b9b1373a719e505b496\" returns successfully" Sep 11 05:16:12.174321 systemd[1]: Started sshd@14-10.0.0.55:22-10.0.0.1:52342.service - OpenSSH per-connection server daemon (10.0.0.1:52342). Sep 11 05:16:12.549366 sshd[5312]: Accepted publickey for core from 10.0.0.1 port 52342 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:12.550888 sshd-session[5312]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:12.556325 systemd-logind[1516]: New session 15 of user core. Sep 11 05:16:12.569169 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 11 05:16:12.778839 sshd[5316]: Connection closed by 10.0.0.1 port 52342 Sep 11 05:16:12.779335 sshd-session[5312]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:12.784612 systemd[1]: sshd@14-10.0.0.55:22-10.0.0.1:52342.service: Deactivated successfully. Sep 11 05:16:12.788191 systemd[1]: session-15.scope: Deactivated successfully. Sep 11 05:16:12.790075 systemd-logind[1516]: Session 15 logged out. Waiting for processes to exit. Sep 11 05:16:12.792321 systemd-logind[1516]: Removed session 15. Sep 11 05:16:14.381961 containerd[1579]: time="2025-09-11T05:16:14.381858767Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:14.383863 containerd[1579]: time="2025-09-11T05:16:14.383788573Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Sep 11 05:16:14.386613 containerd[1579]: time="2025-09-11T05:16:14.386539514Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:14.402242 containerd[1579]: time="2025-09-11T05:16:14.402125112Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 05:16:14.403180 containerd[1579]: time="2025-09-11T05:16:14.403113835Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 3.061786096s" Sep 11 05:16:14.403180 containerd[1579]: time="2025-09-11T05:16:14.403156977Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Sep 11 05:16:14.414242 containerd[1579]: time="2025-09-11T05:16:14.414008849Z" level=info msg="CreateContainer within sandbox \"5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 11 05:16:14.431746 containerd[1579]: time="2025-09-11T05:16:14.431275069Z" level=info msg="Container a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e: CDI devices from CRI Config.CDIDevices: []" Sep 11 05:16:14.451951 containerd[1579]: time="2025-09-11T05:16:14.451860091Z" level=info msg="CreateContainer within sandbox \"5c9b9f55c91c738de098594d2d23c93785f299c4898d3d5415f3708b0e5024a0\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e\"" Sep 11 05:16:14.452681 containerd[1579]: time="2025-09-11T05:16:14.452629908Z" level=info msg="StartContainer for \"a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e\"" Sep 11 05:16:14.454635 containerd[1579]: time="2025-09-11T05:16:14.454595952Z" level=info msg="connecting to shim a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e" address="unix:///run/containerd/s/541753aef0cdd660312e6e72b8f24d80c7f64f9abe898c97cedb51977075cb31" protocol=ttrpc version=3 Sep 11 05:16:14.491247 systemd[1]: Started cri-containerd-a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e.scope - libcontainer container a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e. Sep 11 05:16:14.551742 containerd[1579]: time="2025-09-11T05:16:14.551665980Z" level=info msg="StartContainer for \"a082aff55226675f76e2f91aaa3fd4f3c84bcde456111be8cef99c73b4c57a6e\" returns successfully" Sep 11 05:16:14.665646 kubelet[2739]: I0911 05:16:14.665509 2739 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 11 05:16:14.667277 kubelet[2739]: I0911 05:16:14.667248 2739 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 11 05:16:14.922383 kubelet[2739]: I0911 05:16:14.922217 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-766966cf54-mvzfv" podStartSLOduration=5.329746272 podStartE2EDuration="28.922182239s" podCreationTimestamp="2025-09-11 05:15:46 +0000 UTC" firstStartedPulling="2025-09-11 05:15:47.747007179 +0000 UTC m=+57.297492355" lastFinishedPulling="2025-09-11 05:16:11.339443145 +0000 UTC m=+80.889928322" observedRunningTime="2025-09-11 05:16:11.917656886 +0000 UTC m=+81.468142052" watchObservedRunningTime="2025-09-11 05:16:14.922182239 +0000 UTC m=+84.472667415" Sep 11 05:16:14.923315 kubelet[2739]: I0911 05:16:14.923244 2739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-kqvtp" podStartSLOduration=39.199051807 podStartE2EDuration="1m4.923236437s" podCreationTimestamp="2025-09-11 05:15:10 +0000 UTC" firstStartedPulling="2025-09-11 05:15:48.681024086 +0000 UTC m=+58.231509262" lastFinishedPulling="2025-09-11 05:16:14.405208716 +0000 UTC m=+83.955693892" observedRunningTime="2025-09-11 05:16:14.921241697 +0000 UTC m=+84.471726883" watchObservedRunningTime="2025-09-11 05:16:14.923236437 +0000 UTC m=+84.473721613" Sep 11 05:16:16.948726 containerd[1579]: time="2025-09-11T05:16:16.948675317Z" level=info msg="TaskExit event in podsandbox handler container_id:\"688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0\" id:\"2a19fc54359f255d7d55dce8d0894109c8f7422cb1e312ad88c457a87c31b7d2\" pid:5379 exited_at:{seconds:1757567776 nanos:948277801}" Sep 11 05:16:17.037569 containerd[1579]: time="2025-09-11T05:16:17.037496868Z" level=info msg="TaskExit event in podsandbox handler container_id:\"688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0\" id:\"2441d2a9b8238165a41df061421d9cf7f0329763b14fa0881e41f1ec19489ea7\" pid:5404 exited_at:{seconds:1757567777 nanos:37162633}" Sep 11 05:16:17.362414 kubelet[2739]: I0911 05:16:17.362339 2739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 05:16:17.791984 systemd[1]: Started sshd@15-10.0.0.55:22-10.0.0.1:52344.service - OpenSSH per-connection server daemon (10.0.0.1:52344). Sep 11 05:16:17.874298 sshd[5422]: Accepted publickey for core from 10.0.0.1 port 52344 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:17.876469 sshd-session[5422]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:17.882278 systemd-logind[1516]: New session 16 of user core. Sep 11 05:16:17.901293 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 11 05:16:18.339727 sshd[5426]: Connection closed by 10.0.0.1 port 52344 Sep 11 05:16:18.340341 sshd-session[5422]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:18.346656 systemd[1]: sshd@15-10.0.0.55:22-10.0.0.1:52344.service: Deactivated successfully. Sep 11 05:16:18.349621 systemd[1]: session-16.scope: Deactivated successfully. Sep 11 05:16:18.350936 systemd-logind[1516]: Session 16 logged out. Waiting for processes to exit. Sep 11 05:16:18.352648 systemd-logind[1516]: Removed session 16. Sep 11 05:16:20.511241 containerd[1579]: time="2025-09-11T05:16:20.511190489Z" level=info msg="TaskExit event in podsandbox handler container_id:\"08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec\" id:\"cad940aabd6d00c027d8baf3346fffafeb625270167e2ac33c9be31bf98f5fba\" pid:5451 exited_at:{seconds:1757567780 nanos:510890698}" Sep 11 05:16:23.354949 systemd[1]: Started sshd@16-10.0.0.55:22-10.0.0.1:46948.service - OpenSSH per-connection server daemon (10.0.0.1:46948). Sep 11 05:16:23.408990 sshd[5462]: Accepted publickey for core from 10.0.0.1 port 46948 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:23.411319 sshd-session[5462]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:23.417312 systemd-logind[1516]: New session 17 of user core. Sep 11 05:16:23.427210 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 11 05:16:23.546144 sshd[5465]: Connection closed by 10.0.0.1 port 46948 Sep 11 05:16:23.546553 sshd-session[5462]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:23.551830 systemd[1]: sshd@16-10.0.0.55:22-10.0.0.1:46948.service: Deactivated successfully. Sep 11 05:16:23.554285 systemd[1]: session-17.scope: Deactivated successfully. Sep 11 05:16:23.555790 systemd-logind[1516]: Session 17 logged out. Waiting for processes to exit. Sep 11 05:16:23.557667 systemd-logind[1516]: Removed session 17. Sep 11 05:16:23.890237 containerd[1579]: time="2025-09-11T05:16:23.890173711Z" level=info msg="TaskExit event in podsandbox handler container_id:\"08fb003ca81a6b1542e514f0c76b0f81b215fb5b0c9d3af7b96b88b8dc8564ec\" id:\"6926aee12b5709b8178fcf762c71c8d4d5c78d3c237f483404f87b6bf81cb97d\" pid:5490 exited_at:{seconds:1757567783 nanos:889902877}" Sep 11 05:16:24.579732 kubelet[2739]: E0911 05:16:24.579688 2739 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 05:16:28.559750 systemd[1]: Started sshd@17-10.0.0.55:22-10.0.0.1:46960.service - OpenSSH per-connection server daemon (10.0.0.1:46960). Sep 11 05:16:28.625118 sshd[5503]: Accepted publickey for core from 10.0.0.1 port 46960 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:28.626946 sshd-session[5503]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:28.631643 systemd-logind[1516]: New session 18 of user core. Sep 11 05:16:28.644213 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 11 05:16:28.767246 sshd[5506]: Connection closed by 10.0.0.1 port 46960 Sep 11 05:16:28.767669 sshd-session[5503]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:28.777192 systemd[1]: sshd@17-10.0.0.55:22-10.0.0.1:46960.service: Deactivated successfully. Sep 11 05:16:28.779410 systemd[1]: session-18.scope: Deactivated successfully. Sep 11 05:16:28.780367 systemd-logind[1516]: Session 18 logged out. Waiting for processes to exit. Sep 11 05:16:28.783383 systemd[1]: Started sshd@18-10.0.0.55:22-10.0.0.1:46976.service - OpenSSH per-connection server daemon (10.0.0.1:46976). Sep 11 05:16:28.784432 systemd-logind[1516]: Removed session 18. Sep 11 05:16:28.846321 sshd[5519]: Accepted publickey for core from 10.0.0.1 port 46976 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:28.847738 sshd-session[5519]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:28.853986 systemd-logind[1516]: New session 19 of user core. Sep 11 05:16:28.869153 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 11 05:16:29.312585 sshd[5522]: Connection closed by 10.0.0.1 port 46976 Sep 11 05:16:29.313059 sshd-session[5519]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:29.328037 systemd[1]: sshd@18-10.0.0.55:22-10.0.0.1:46976.service: Deactivated successfully. Sep 11 05:16:29.330248 systemd[1]: session-19.scope: Deactivated successfully. Sep 11 05:16:29.331010 systemd-logind[1516]: Session 19 logged out. Waiting for processes to exit. Sep 11 05:16:29.334100 systemd[1]: Started sshd@19-10.0.0.55:22-10.0.0.1:46978.service - OpenSSH per-connection server daemon (10.0.0.1:46978). Sep 11 05:16:29.335033 systemd-logind[1516]: Removed session 19. Sep 11 05:16:29.410291 sshd[5539]: Accepted publickey for core from 10.0.0.1 port 46978 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:29.412354 sshd-session[5539]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:29.416810 systemd-logind[1516]: New session 20 of user core. Sep 11 05:16:29.424209 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 11 05:16:30.124971 sshd[5542]: Connection closed by 10.0.0.1 port 46978 Sep 11 05:16:30.127724 sshd-session[5539]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:30.139024 systemd[1]: sshd@19-10.0.0.55:22-10.0.0.1:46978.service: Deactivated successfully. Sep 11 05:16:30.142779 systemd[1]: session-20.scope: Deactivated successfully. Sep 11 05:16:30.146338 systemd-logind[1516]: Session 20 logged out. Waiting for processes to exit. Sep 11 05:16:30.155443 systemd[1]: Started sshd@20-10.0.0.55:22-10.0.0.1:37400.service - OpenSSH per-connection server daemon (10.0.0.1:37400). Sep 11 05:16:30.158227 systemd-logind[1516]: Removed session 20. Sep 11 05:16:30.217744 sshd[5563]: Accepted publickey for core from 10.0.0.1 port 37400 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:30.219691 sshd-session[5563]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:30.226530 systemd-logind[1516]: New session 21 of user core. Sep 11 05:16:30.239168 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 11 05:16:30.549575 sshd[5566]: Connection closed by 10.0.0.1 port 37400 Sep 11 05:16:30.550022 sshd-session[5563]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:30.561847 systemd[1]: sshd@20-10.0.0.55:22-10.0.0.1:37400.service: Deactivated successfully. Sep 11 05:16:30.565181 systemd[1]: session-21.scope: Deactivated successfully. Sep 11 05:16:30.567006 systemd-logind[1516]: Session 21 logged out. Waiting for processes to exit. Sep 11 05:16:30.571642 systemd[1]: Started sshd@21-10.0.0.55:22-10.0.0.1:37402.service - OpenSSH per-connection server daemon (10.0.0.1:37402). Sep 11 05:16:30.573078 systemd-logind[1516]: Removed session 21. Sep 11 05:16:30.632048 sshd[5577]: Accepted publickey for core from 10.0.0.1 port 37402 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:30.635701 sshd-session[5577]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:30.643690 systemd-logind[1516]: New session 22 of user core. Sep 11 05:16:30.651078 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 11 05:16:30.775053 sshd[5580]: Connection closed by 10.0.0.1 port 37402 Sep 11 05:16:30.775447 sshd-session[5577]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:30.779845 systemd[1]: sshd@21-10.0.0.55:22-10.0.0.1:37402.service: Deactivated successfully. Sep 11 05:16:30.782051 systemd[1]: session-22.scope: Deactivated successfully. Sep 11 05:16:30.782812 systemd-logind[1516]: Session 22 logged out. Waiting for processes to exit. Sep 11 05:16:30.784128 systemd-logind[1516]: Removed session 22. Sep 11 05:16:35.791177 systemd[1]: Started sshd@22-10.0.0.55:22-10.0.0.1:37410.service - OpenSSH per-connection server daemon (10.0.0.1:37410). Sep 11 05:16:35.850342 sshd[5596]: Accepted publickey for core from 10.0.0.1 port 37410 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:35.851992 sshd-session[5596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:35.856459 systemd-logind[1516]: New session 23 of user core. Sep 11 05:16:35.867061 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 11 05:16:35.988411 sshd[5599]: Connection closed by 10.0.0.1 port 37410 Sep 11 05:16:35.988742 sshd-session[5596]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:35.993025 systemd[1]: sshd@22-10.0.0.55:22-10.0.0.1:37410.service: Deactivated successfully. Sep 11 05:16:35.995295 systemd[1]: session-23.scope: Deactivated successfully. Sep 11 05:16:35.996171 systemd-logind[1516]: Session 23 logged out. Waiting for processes to exit. Sep 11 05:16:35.998006 systemd-logind[1516]: Removed session 23. Sep 11 05:16:36.257377 containerd[1579]: time="2025-09-11T05:16:36.257228561Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\" id:\"b9664375c5a9c2d454e021592804c1ae0ac3a1fbc6fbaa870b36bcc59252d7d4\" pid:5622 exited_at:{seconds:1757567796 nanos:246191158}" Sep 11 05:16:40.595262 containerd[1579]: time="2025-09-11T05:16:40.595192430Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d602504f870fb752b207fa8b780c1bcfd3e233cf8d0ccbcb86ad3b724f6b7a4f\" id:\"a64a1ae8ae3d010d9ba5b4198b598d50df3657cf43bcffd27b1b11b378ee83f5\" pid:5649 exited_at:{seconds:1757567800 nanos:594740856}" Sep 11 05:16:41.004697 systemd[1]: Started sshd@23-10.0.0.55:22-10.0.0.1:51638.service - OpenSSH per-connection server daemon (10.0.0.1:51638). Sep 11 05:16:41.089191 sshd[5662]: Accepted publickey for core from 10.0.0.1 port 51638 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:41.091291 sshd-session[5662]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:41.098452 systemd-logind[1516]: New session 24 of user core. Sep 11 05:16:41.105227 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 11 05:16:41.242309 sshd[5665]: Connection closed by 10.0.0.1 port 51638 Sep 11 05:16:41.242678 sshd-session[5662]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:41.248461 systemd[1]: sshd@23-10.0.0.55:22-10.0.0.1:51638.service: Deactivated successfully. Sep 11 05:16:41.250988 systemd[1]: session-24.scope: Deactivated successfully. Sep 11 05:16:41.252013 systemd-logind[1516]: Session 24 logged out. Waiting for processes to exit. Sep 11 05:16:41.254193 systemd-logind[1516]: Removed session 24. Sep 11 05:16:46.256579 systemd[1]: Started sshd@24-10.0.0.55:22-10.0.0.1:51648.service - OpenSSH per-connection server daemon (10.0.0.1:51648). Sep 11 05:16:46.337987 sshd[5678]: Accepted publickey for core from 10.0.0.1 port 51648 ssh2: RSA SHA256:1pZGs1ipgSiR49h35TYw6JQUcu6QRzqFxKKJKg1EdYM Sep 11 05:16:46.339702 sshd-session[5678]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 05:16:46.345960 systemd-logind[1516]: New session 25 of user core. Sep 11 05:16:46.353143 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 11 05:16:46.641054 sshd[5681]: Connection closed by 10.0.0.1 port 51648 Sep 11 05:16:46.641977 sshd-session[5678]: pam_unix(sshd:session): session closed for user core Sep 11 05:16:46.649244 systemd-logind[1516]: Session 25 logged out. Waiting for processes to exit. Sep 11 05:16:46.649450 systemd[1]: sshd@24-10.0.0.55:22-10.0.0.1:51648.service: Deactivated successfully. Sep 11 05:16:46.652085 systemd[1]: session-25.scope: Deactivated successfully. Sep 11 05:16:46.654297 systemd-logind[1516]: Removed session 25. Sep 11 05:16:47.091837 containerd[1579]: time="2025-09-11T05:16:47.070573735Z" level=info msg="TaskExit event in podsandbox handler container_id:\"688457730ac0a78e34fc99090d6c361871d426f7deffb5e2ca3a91e671227ee0\" id:\"a8d93390f8e09bae595ed0a82512c32fcb6cc8717849ba03c3d258ddca1ba633\" pid:5704 exited_at:{seconds:1757567807 nanos:70123904}"