Sep 11 00:18:50.865032 kernel: Linux version 6.12.46-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed Sep 10 22:15:45 -00 2025 Sep 11 00:18:50.865065 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=20820f07706ad5590d38fe5324b9055d59a89dc1109fdc449cad1a53209b9dbd Sep 11 00:18:50.865078 kernel: BIOS-provided physical RAM map: Sep 11 00:18:50.865084 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 11 00:18:50.865091 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 11 00:18:50.865100 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 11 00:18:50.865108 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Sep 11 00:18:50.865114 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 11 00:18:50.865121 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Sep 11 00:18:50.865130 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Sep 11 00:18:50.865136 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Sep 11 00:18:50.865143 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Sep 11 00:18:50.865149 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Sep 11 00:18:50.865156 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Sep 11 00:18:50.865170 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Sep 11 00:18:50.865181 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 11 00:18:50.865190 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Sep 11 00:18:50.865204 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Sep 11 00:18:50.865214 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Sep 11 00:18:50.865223 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Sep 11 00:18:50.865233 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Sep 11 00:18:50.865242 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 11 00:18:50.865256 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 11 00:18:50.865265 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 11 00:18:50.865273 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Sep 11 00:18:50.865283 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 11 00:18:50.865290 kernel: NX (Execute Disable) protection: active Sep 11 00:18:50.865297 kernel: APIC: Static calls initialized Sep 11 00:18:50.865304 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Sep 11 00:18:50.865313 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Sep 11 00:18:50.865323 kernel: extended physical RAM map: Sep 11 00:18:50.865332 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 11 00:18:50.865342 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 11 00:18:50.865351 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 11 00:18:50.865360 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Sep 11 00:18:50.865370 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 11 00:18:50.865382 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Sep 11 00:18:50.865389 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Sep 11 00:18:50.865396 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Sep 11 00:18:50.865404 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Sep 11 00:18:50.865415 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Sep 11 00:18:50.865428 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Sep 11 00:18:50.865446 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Sep 11 00:18:50.865454 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Sep 11 00:18:50.865461 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Sep 11 00:18:50.865469 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Sep 11 00:18:50.865480 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Sep 11 00:18:50.865494 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 11 00:18:50.865504 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Sep 11 00:18:50.865514 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Sep 11 00:18:50.865524 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Sep 11 00:18:50.865534 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Sep 11 00:18:50.865581 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Sep 11 00:18:50.865592 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 11 00:18:50.865602 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 11 00:18:50.865611 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 11 00:18:50.865620 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Sep 11 00:18:50.865629 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 11 00:18:50.865638 kernel: efi: EFI v2.7 by EDK II Sep 11 00:18:50.865648 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Sep 11 00:18:50.865658 kernel: random: crng init done Sep 11 00:18:50.865673 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Sep 11 00:18:50.865683 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Sep 11 00:18:50.865699 kernel: secureboot: Secure boot disabled Sep 11 00:18:50.865707 kernel: SMBIOS 2.8 present. Sep 11 00:18:50.865714 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Sep 11 00:18:50.865723 kernel: DMI: Memory slots populated: 1/1 Sep 11 00:18:50.865732 kernel: Hypervisor detected: KVM Sep 11 00:18:50.865741 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 11 00:18:50.865750 kernel: kvm-clock: using sched offset of 5404855263 cycles Sep 11 00:18:50.865761 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 11 00:18:50.865771 kernel: tsc: Detected 2794.748 MHz processor Sep 11 00:18:50.865781 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 11 00:18:50.865794 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 11 00:18:50.865812 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Sep 11 00:18:50.865822 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 11 00:18:50.865843 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 11 00:18:50.865853 kernel: Using GB pages for direct mapping Sep 11 00:18:50.865872 kernel: ACPI: Early table checksum verification disabled Sep 11 00:18:50.865882 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Sep 11 00:18:50.865894 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Sep 11 00:18:50.865904 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 00:18:50.865914 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 00:18:50.865928 kernel: ACPI: FACS 0x000000009CBDD000 000040 Sep 11 00:18:50.865938 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 00:18:50.865948 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 00:18:50.865967 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 00:18:50.865978 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 11 00:18:50.866001 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Sep 11 00:18:50.866018 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Sep 11 00:18:50.866028 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Sep 11 00:18:50.866042 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Sep 11 00:18:50.866063 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Sep 11 00:18:50.866073 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Sep 11 00:18:50.866086 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Sep 11 00:18:50.866095 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Sep 11 00:18:50.866105 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Sep 11 00:18:50.866115 kernel: No NUMA configuration found Sep 11 00:18:50.866125 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Sep 11 00:18:50.866134 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Sep 11 00:18:50.866142 kernel: Zone ranges: Sep 11 00:18:50.866154 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 11 00:18:50.866162 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Sep 11 00:18:50.866169 kernel: Normal empty Sep 11 00:18:50.866177 kernel: Device empty Sep 11 00:18:50.866185 kernel: Movable zone start for each node Sep 11 00:18:50.866192 kernel: Early memory node ranges Sep 11 00:18:50.866200 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 11 00:18:50.866208 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Sep 11 00:18:50.866223 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Sep 11 00:18:50.866236 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Sep 11 00:18:50.866253 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Sep 11 00:18:50.866261 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Sep 11 00:18:50.866272 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Sep 11 00:18:50.866284 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Sep 11 00:18:50.866300 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Sep 11 00:18:50.866311 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 11 00:18:50.866322 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 11 00:18:50.866347 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Sep 11 00:18:50.866379 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 11 00:18:50.866400 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Sep 11 00:18:50.866419 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Sep 11 00:18:50.866430 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Sep 11 00:18:50.866469 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Sep 11 00:18:50.866492 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Sep 11 00:18:50.866503 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 11 00:18:50.866511 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 11 00:18:50.866524 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 11 00:18:50.866532 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 11 00:18:50.866540 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 11 00:18:50.866571 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 11 00:18:50.866582 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 11 00:18:50.866599 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 11 00:18:50.866620 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 11 00:18:50.866642 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 11 00:18:50.866654 kernel: TSC deadline timer available Sep 11 00:18:50.866669 kernel: CPU topo: Max. logical packages: 1 Sep 11 00:18:50.866679 kernel: CPU topo: Max. logical dies: 1 Sep 11 00:18:50.866688 kernel: CPU topo: Max. dies per package: 1 Sep 11 00:18:50.866698 kernel: CPU topo: Max. threads per core: 1 Sep 11 00:18:50.866707 kernel: CPU topo: Num. cores per package: 4 Sep 11 00:18:50.866717 kernel: CPU topo: Num. threads per package: 4 Sep 11 00:18:50.866727 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Sep 11 00:18:50.866746 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 11 00:18:50.866757 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 11 00:18:50.866767 kernel: kvm-guest: setup PV sched yield Sep 11 00:18:50.866784 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Sep 11 00:18:50.866794 kernel: Booting paravirtualized kernel on KVM Sep 11 00:18:50.866805 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 11 00:18:50.866815 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 11 00:18:50.866825 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Sep 11 00:18:50.866836 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Sep 11 00:18:50.866846 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 11 00:18:50.866864 kernel: kvm-guest: PV spinlocks enabled Sep 11 00:18:50.866875 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 11 00:18:50.866894 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=20820f07706ad5590d38fe5324b9055d59a89dc1109fdc449cad1a53209b9dbd Sep 11 00:18:50.866909 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 11 00:18:50.866918 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 11 00:18:50.866926 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 11 00:18:50.866943 kernel: Fallback order for Node 0: 0 Sep 11 00:18:50.866952 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Sep 11 00:18:50.866967 kernel: Policy zone: DMA32 Sep 11 00:18:50.866991 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 11 00:18:50.867003 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 11 00:18:50.867013 kernel: ftrace: allocating 40106 entries in 157 pages Sep 11 00:18:50.867021 kernel: ftrace: allocated 157 pages with 5 groups Sep 11 00:18:50.867029 kernel: Dynamic Preempt: voluntary Sep 11 00:18:50.867037 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 11 00:18:50.867047 kernel: rcu: RCU event tracing is enabled. Sep 11 00:18:50.867055 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 11 00:18:50.867063 kernel: Trampoline variant of Tasks RCU enabled. Sep 11 00:18:50.867071 kernel: Rude variant of Tasks RCU enabled. Sep 11 00:18:50.867087 kernel: Tracing variant of Tasks RCU enabled. Sep 11 00:18:50.867102 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 11 00:18:50.867120 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 11 00:18:50.867134 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 00:18:50.867154 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 00:18:50.867169 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 11 00:18:50.867188 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 11 00:18:50.867197 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 11 00:18:50.867205 kernel: Console: colour dummy device 80x25 Sep 11 00:18:50.867216 kernel: printk: legacy console [ttyS0] enabled Sep 11 00:18:50.867225 kernel: ACPI: Core revision 20240827 Sep 11 00:18:50.867233 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 11 00:18:50.867242 kernel: APIC: Switch to symmetric I/O mode setup Sep 11 00:18:50.867252 kernel: x2apic enabled Sep 11 00:18:50.867263 kernel: APIC: Switched APIC routing to: physical x2apic Sep 11 00:18:50.867273 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 11 00:18:50.867281 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 11 00:18:50.867289 kernel: kvm-guest: setup PV IPIs Sep 11 00:18:50.867306 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 11 00:18:50.867314 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Sep 11 00:18:50.867323 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Sep 11 00:18:50.867339 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 11 00:18:50.867355 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 11 00:18:50.867365 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 11 00:18:50.867373 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 11 00:18:50.867381 kernel: Spectre V2 : Mitigation: Retpolines Sep 11 00:18:50.867389 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 11 00:18:50.867401 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 11 00:18:50.867408 kernel: active return thunk: retbleed_return_thunk Sep 11 00:18:50.867423 kernel: RETBleed: Mitigation: untrained return thunk Sep 11 00:18:50.867441 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 11 00:18:50.867452 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 11 00:18:50.867470 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 11 00:18:50.867485 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 11 00:18:50.867495 kernel: active return thunk: srso_return_thunk Sep 11 00:18:50.867518 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 11 00:18:50.867530 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 11 00:18:50.867566 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 11 00:18:50.867577 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 11 00:18:50.867585 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 11 00:18:50.867594 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 11 00:18:50.867605 kernel: Freeing SMP alternatives memory: 32K Sep 11 00:18:50.867615 kernel: pid_max: default: 32768 minimum: 301 Sep 11 00:18:50.867626 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 11 00:18:50.867641 kernel: landlock: Up and running. Sep 11 00:18:50.867649 kernel: SELinux: Initializing. Sep 11 00:18:50.867657 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 11 00:18:50.867666 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 11 00:18:50.867674 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 11 00:18:50.867681 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 11 00:18:50.867689 kernel: ... version: 0 Sep 11 00:18:50.867697 kernel: ... bit width: 48 Sep 11 00:18:50.867705 kernel: ... generic registers: 6 Sep 11 00:18:50.867715 kernel: ... value mask: 0000ffffffffffff Sep 11 00:18:50.867723 kernel: ... max period: 00007fffffffffff Sep 11 00:18:50.867730 kernel: ... fixed-purpose events: 0 Sep 11 00:18:50.867738 kernel: ... event mask: 000000000000003f Sep 11 00:18:50.867746 kernel: signal: max sigframe size: 1776 Sep 11 00:18:50.867753 kernel: rcu: Hierarchical SRCU implementation. Sep 11 00:18:50.867762 kernel: rcu: Max phase no-delay instances is 400. Sep 11 00:18:50.867770 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 11 00:18:50.867777 kernel: smp: Bringing up secondary CPUs ... Sep 11 00:18:50.867787 kernel: smpboot: x86: Booting SMP configuration: Sep 11 00:18:50.867795 kernel: .... node #0, CPUs: #1 #2 #3 Sep 11 00:18:50.867803 kernel: smp: Brought up 1 node, 4 CPUs Sep 11 00:18:50.867811 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Sep 11 00:18:50.867819 kernel: Memory: 2422672K/2565800K available (14336K kernel code, 2429K rwdata, 9960K rodata, 54036K init, 2932K bss, 137196K reserved, 0K cma-reserved) Sep 11 00:18:50.867827 kernel: devtmpfs: initialized Sep 11 00:18:50.867835 kernel: x86/mm: Memory block size: 128MB Sep 11 00:18:50.867844 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Sep 11 00:18:50.867854 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Sep 11 00:18:50.867868 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Sep 11 00:18:50.867878 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Sep 11 00:18:50.867889 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Sep 11 00:18:50.867899 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Sep 11 00:18:50.867910 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 11 00:18:50.867920 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 11 00:18:50.867930 kernel: pinctrl core: initialized pinctrl subsystem Sep 11 00:18:50.867941 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 11 00:18:50.867951 kernel: audit: initializing netlink subsys (disabled) Sep 11 00:18:50.867965 kernel: audit: type=2000 audit(1757549926.474:1): state=initialized audit_enabled=0 res=1 Sep 11 00:18:50.867975 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 11 00:18:50.867996 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 11 00:18:50.868007 kernel: cpuidle: using governor menu Sep 11 00:18:50.868017 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 11 00:18:50.868028 kernel: dca service started, version 1.12.1 Sep 11 00:18:50.868037 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Sep 11 00:18:50.868046 kernel: PCI: Using configuration type 1 for base access Sep 11 00:18:50.868054 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 11 00:18:50.868065 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 11 00:18:50.868073 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 11 00:18:50.868081 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 11 00:18:50.868089 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 11 00:18:50.868097 kernel: ACPI: Added _OSI(Module Device) Sep 11 00:18:50.868105 kernel: ACPI: Added _OSI(Processor Device) Sep 11 00:18:50.868113 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 11 00:18:50.868121 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 11 00:18:50.868129 kernel: ACPI: Interpreter enabled Sep 11 00:18:50.868139 kernel: ACPI: PM: (supports S0 S3 S5) Sep 11 00:18:50.868147 kernel: ACPI: Using IOAPIC for interrupt routing Sep 11 00:18:50.868155 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 11 00:18:50.868163 kernel: PCI: Using E820 reservations for host bridge windows Sep 11 00:18:50.868171 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 11 00:18:50.868179 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 11 00:18:50.868434 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 11 00:18:50.868610 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 11 00:18:50.868743 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 11 00:18:50.868753 kernel: PCI host bridge to bus 0000:00 Sep 11 00:18:50.868901 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 11 00:18:50.869031 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 11 00:18:50.869161 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 11 00:18:50.869309 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Sep 11 00:18:50.869475 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Sep 11 00:18:50.869657 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Sep 11 00:18:50.869810 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 11 00:18:50.870017 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Sep 11 00:18:50.870197 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Sep 11 00:18:50.870391 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Sep 11 00:18:50.870591 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Sep 11 00:18:50.870746 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Sep 11 00:18:50.871409 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 11 00:18:50.871572 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 11 00:18:50.871701 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Sep 11 00:18:50.871849 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Sep 11 00:18:50.872041 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Sep 11 00:18:50.872192 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Sep 11 00:18:50.872362 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Sep 11 00:18:50.872499 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Sep 11 00:18:50.872645 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Sep 11 00:18:50.872779 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Sep 11 00:18:50.872902 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Sep 11 00:18:50.873040 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Sep 11 00:18:50.873172 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Sep 11 00:18:50.873294 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Sep 11 00:18:50.873428 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Sep 11 00:18:50.873569 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 11 00:18:50.873702 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Sep 11 00:18:50.873825 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Sep 11 00:18:50.873945 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Sep 11 00:18:50.874153 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Sep 11 00:18:50.874313 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Sep 11 00:18:50.874331 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 11 00:18:50.874343 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 11 00:18:50.874353 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 11 00:18:50.874364 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 11 00:18:50.874375 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 11 00:18:50.874384 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 11 00:18:50.874399 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 11 00:18:50.874410 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 11 00:18:50.874421 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 11 00:18:50.874432 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 11 00:18:50.874443 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 11 00:18:50.874453 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 11 00:18:50.874464 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 11 00:18:50.874474 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 11 00:18:50.874485 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 11 00:18:50.874499 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 11 00:18:50.874509 kernel: iommu: Default domain type: Translated Sep 11 00:18:50.874520 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 11 00:18:50.874531 kernel: efivars: Registered efivars operations Sep 11 00:18:50.874560 kernel: PCI: Using ACPI for IRQ routing Sep 11 00:18:50.874572 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 11 00:18:50.874582 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Sep 11 00:18:50.874593 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Sep 11 00:18:50.874604 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Sep 11 00:18:50.874618 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Sep 11 00:18:50.874629 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Sep 11 00:18:50.874639 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Sep 11 00:18:50.874650 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Sep 11 00:18:50.874659 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Sep 11 00:18:50.874809 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 11 00:18:50.874993 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 11 00:18:50.875149 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 11 00:18:50.875167 kernel: vgaarb: loaded Sep 11 00:18:50.875179 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 11 00:18:50.875190 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 11 00:18:50.875201 kernel: clocksource: Switched to clocksource kvm-clock Sep 11 00:18:50.875212 kernel: VFS: Disk quotas dquot_6.6.0 Sep 11 00:18:50.875223 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 11 00:18:50.875234 kernel: pnp: PnP ACPI init Sep 11 00:18:50.875537 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Sep 11 00:18:50.875573 kernel: pnp: PnP ACPI: found 6 devices Sep 11 00:18:50.875582 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 11 00:18:50.875591 kernel: NET: Registered PF_INET protocol family Sep 11 00:18:50.875599 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 11 00:18:50.875608 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 11 00:18:50.875616 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 11 00:18:50.875625 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 11 00:18:50.875633 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 11 00:18:50.875641 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 11 00:18:50.875652 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 11 00:18:50.875660 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 11 00:18:50.875669 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 11 00:18:50.875677 kernel: NET: Registered PF_XDP protocol family Sep 11 00:18:50.875810 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Sep 11 00:18:50.876738 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Sep 11 00:18:50.876914 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 11 00:18:50.877091 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 11 00:18:50.877248 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 11 00:18:50.877366 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Sep 11 00:18:50.877503 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Sep 11 00:18:50.877651 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Sep 11 00:18:50.877665 kernel: PCI: CLS 0 bytes, default 64 Sep 11 00:18:50.877675 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Sep 11 00:18:50.877684 kernel: Initialise system trusted keyrings Sep 11 00:18:50.877698 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 11 00:18:50.877707 kernel: Key type asymmetric registered Sep 11 00:18:50.877716 kernel: Asymmetric key parser 'x509' registered Sep 11 00:18:50.877727 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 11 00:18:50.877736 kernel: io scheduler mq-deadline registered Sep 11 00:18:50.877747 kernel: io scheduler kyber registered Sep 11 00:18:50.877758 kernel: io scheduler bfq registered Sep 11 00:18:50.877773 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 11 00:18:50.877786 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 11 00:18:50.877797 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 11 00:18:50.877808 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 11 00:18:50.877818 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 11 00:18:50.877828 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 11 00:18:50.877838 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 11 00:18:50.877848 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 11 00:18:50.877858 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 11 00:18:50.878029 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 11 00:18:50.878170 kernel: rtc_cmos 00:04: registered as rtc0 Sep 11 00:18:50.878186 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 11 00:18:50.878328 kernel: rtc_cmos 00:04: setting system clock to 2025-09-11T00:18:50 UTC (1757549930) Sep 11 00:18:50.878473 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Sep 11 00:18:50.878489 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 11 00:18:50.878501 kernel: efifb: probing for efifb Sep 11 00:18:50.878512 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Sep 11 00:18:50.878529 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Sep 11 00:18:50.878553 kernel: efifb: scrolling: redraw Sep 11 00:18:50.878563 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 11 00:18:50.878572 kernel: Console: switching to colour frame buffer device 160x50 Sep 11 00:18:50.878581 kernel: fb0: EFI VGA frame buffer device Sep 11 00:18:50.878589 kernel: pstore: Using crash dump compression: deflate Sep 11 00:18:50.878598 kernel: pstore: Registered efi_pstore as persistent store backend Sep 11 00:18:50.878607 kernel: NET: Registered PF_INET6 protocol family Sep 11 00:18:50.878615 kernel: Segment Routing with IPv6 Sep 11 00:18:50.878628 kernel: In-situ OAM (IOAM) with IPv6 Sep 11 00:18:50.878637 kernel: NET: Registered PF_PACKET protocol family Sep 11 00:18:50.878646 kernel: Key type dns_resolver registered Sep 11 00:18:50.878654 kernel: IPI shorthand broadcast: enabled Sep 11 00:18:50.878663 kernel: sched_clock: Marking stable (4863003063, 190784156)->(5198185751, -144398532) Sep 11 00:18:50.878672 kernel: registered taskstats version 1 Sep 11 00:18:50.878681 kernel: Loading compiled-in X.509 certificates Sep 11 00:18:50.878690 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.46-flatcar: 941433bdd955e1c3aa4064827516bddd510466ee' Sep 11 00:18:50.878699 kernel: Demotion targets for Node 0: null Sep 11 00:18:50.878710 kernel: Key type .fscrypt registered Sep 11 00:18:50.878719 kernel: Key type fscrypt-provisioning registered Sep 11 00:18:50.878728 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 11 00:18:50.878737 kernel: ima: Allocated hash algorithm: sha1 Sep 11 00:18:50.878745 kernel: ima: No architecture policies found Sep 11 00:18:50.878754 kernel: clk: Disabling unused clocks Sep 11 00:18:50.878763 kernel: Warning: unable to open an initial console. Sep 11 00:18:50.878772 kernel: Freeing unused kernel image (initmem) memory: 54036K Sep 11 00:18:50.878781 kernel: Write protecting the kernel read-only data: 24576k Sep 11 00:18:50.878793 kernel: Freeing unused kernel image (rodata/data gap) memory: 280K Sep 11 00:18:50.878801 kernel: Run /init as init process Sep 11 00:18:50.878810 kernel: with arguments: Sep 11 00:18:50.878819 kernel: /init Sep 11 00:18:50.878828 kernel: with environment: Sep 11 00:18:50.878837 kernel: HOME=/ Sep 11 00:18:50.878845 kernel: TERM=linux Sep 11 00:18:50.878854 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 11 00:18:50.878869 systemd[1]: Successfully made /usr/ read-only. Sep 11 00:18:50.878885 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 11 00:18:50.878895 systemd[1]: Detected virtualization kvm. Sep 11 00:18:50.878904 systemd[1]: Detected architecture x86-64. Sep 11 00:18:50.878913 systemd[1]: Running in initrd. Sep 11 00:18:50.878922 systemd[1]: No hostname configured, using default hostname. Sep 11 00:18:50.878931 systemd[1]: Hostname set to . Sep 11 00:18:50.878940 systemd[1]: Initializing machine ID from VM UUID. Sep 11 00:18:50.878951 systemd[1]: Queued start job for default target initrd.target. Sep 11 00:18:50.878960 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 00:18:50.878969 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 00:18:50.878979 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 11 00:18:50.879002 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 11 00:18:50.879014 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 11 00:18:50.879026 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 11 00:18:50.879042 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 11 00:18:50.879054 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 11 00:18:50.879066 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 00:18:50.880198 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 11 00:18:50.880207 systemd[1]: Reached target paths.target - Path Units. Sep 11 00:18:50.880216 systemd[1]: Reached target slices.target - Slice Units. Sep 11 00:18:50.880225 systemd[1]: Reached target swap.target - Swaps. Sep 11 00:18:50.880234 systemd[1]: Reached target timers.target - Timer Units. Sep 11 00:18:50.880247 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 11 00:18:50.880256 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 11 00:18:50.880266 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 11 00:18:50.880275 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 11 00:18:50.880284 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 11 00:18:50.880293 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 11 00:18:50.880302 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 00:18:50.880311 systemd[1]: Reached target sockets.target - Socket Units. Sep 11 00:18:50.880320 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 11 00:18:50.880331 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 11 00:18:50.880340 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 11 00:18:50.880349 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 11 00:18:50.880358 systemd[1]: Starting systemd-fsck-usr.service... Sep 11 00:18:50.880368 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 11 00:18:50.880377 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 11 00:18:50.880385 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 00:18:50.880394 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 11 00:18:50.880406 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 00:18:50.880415 systemd[1]: Finished systemd-fsck-usr.service. Sep 11 00:18:50.880425 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 11 00:18:50.880434 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 11 00:18:50.880480 systemd-journald[221]: Collecting audit messages is disabled. Sep 11 00:18:50.880508 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 11 00:18:50.880518 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:18:50.880527 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 11 00:18:50.880537 systemd-journald[221]: Journal started Sep 11 00:18:50.880576 systemd-journald[221]: Runtime Journal (/run/log/journal/837ce06a809f4e869b3fbd9a7e8872bd) is 6M, max 48.4M, 42.4M free. Sep 11 00:18:50.851537 systemd-modules-load[223]: Inserted module 'overlay' Sep 11 00:18:50.883315 systemd[1]: Started systemd-journald.service - Journal Service. Sep 11 00:18:50.885069 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 00:18:50.889752 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 11 00:18:50.891290 systemd-modules-load[223]: Inserted module 'br_netfilter' Sep 11 00:18:50.892277 kernel: Bridge firewalling registered Sep 11 00:18:50.897823 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 11 00:18:50.898316 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 11 00:18:50.902805 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 11 00:18:50.913856 systemd-tmpfiles[246]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 11 00:18:50.917192 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 11 00:18:50.919832 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 11 00:18:50.920177 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 00:18:50.923188 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 11 00:18:50.948561 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 11 00:18:50.963357 dracut-cmdline[261]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=20820f07706ad5590d38fe5324b9055d59a89dc1109fdc449cad1a53209b9dbd Sep 11 00:18:51.008934 systemd-resolved[264]: Positive Trust Anchors: Sep 11 00:18:51.008962 systemd-resolved[264]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 11 00:18:51.009021 systemd-resolved[264]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 11 00:18:51.011990 systemd-resolved[264]: Defaulting to hostname 'linux'. Sep 11 00:18:51.013481 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 11 00:18:51.019710 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 11 00:18:51.088620 kernel: SCSI subsystem initialized Sep 11 00:18:51.097590 kernel: Loading iSCSI transport class v2.0-870. Sep 11 00:18:51.108589 kernel: iscsi: registered transport (tcp) Sep 11 00:18:51.131871 kernel: iscsi: registered transport (qla4xxx) Sep 11 00:18:51.131952 kernel: QLogic iSCSI HBA Driver Sep 11 00:18:51.156725 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 11 00:18:51.182386 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 00:18:51.186449 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 11 00:18:51.259035 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 11 00:18:51.262121 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 11 00:18:51.385615 kernel: raid6: avx2x4 gen() 28803 MB/s Sep 11 00:18:51.402588 kernel: raid6: avx2x2 gen() 29746 MB/s Sep 11 00:18:51.419689 kernel: raid6: avx2x1 gen() 19952 MB/s Sep 11 00:18:51.419740 kernel: raid6: using algorithm avx2x2 gen() 29746 MB/s Sep 11 00:18:51.499041 kernel: raid6: .... xor() 19683 MB/s, rmw enabled Sep 11 00:18:51.499168 kernel: raid6: using avx2x2 recovery algorithm Sep 11 00:18:51.527592 kernel: xor: automatically using best checksumming function avx Sep 11 00:18:51.741633 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 11 00:18:51.752673 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 11 00:18:51.755232 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 00:18:51.790634 systemd-udevd[473]: Using default interface naming scheme 'v255'. Sep 11 00:18:51.798090 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 00:18:51.802536 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 11 00:18:51.830290 dracut-pre-trigger[480]: rd.md=0: removing MD RAID activation Sep 11 00:18:51.865310 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 11 00:18:51.868191 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 11 00:18:52.006076 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 00:18:52.010930 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 11 00:18:52.068287 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 11 00:18:52.068732 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 11 00:18:52.075574 kernel: cryptd: max_cpu_qlen set to 1000 Sep 11 00:18:52.133583 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 11 00:18:52.133674 kernel: libata version 3.00 loaded. Sep 11 00:18:52.133687 kernel: GPT:9289727 != 19775487 Sep 11 00:18:52.134968 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 11 00:18:52.134998 kernel: GPT:9289727 != 19775487 Sep 11 00:18:52.137024 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 11 00:18:52.137105 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 00:18:52.137563 kernel: AES CTR mode by8 optimization enabled Sep 11 00:18:52.143900 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 00:18:52.146655 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Sep 11 00:18:52.144045 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:18:52.151863 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 00:18:52.157321 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 00:18:52.166863 kernel: ahci 0000:00:1f.2: version 3.0 Sep 11 00:18:52.167173 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 11 00:18:52.169722 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Sep 11 00:18:52.169923 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Sep 11 00:18:52.170080 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 11 00:18:52.172305 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 11 00:18:52.192682 kernel: scsi host0: ahci Sep 11 00:18:52.192955 kernel: scsi host1: ahci Sep 11 00:18:52.193278 kernel: scsi host2: ahci Sep 11 00:18:52.194336 kernel: scsi host3: ahci Sep 11 00:18:52.194517 kernel: scsi host4: ahci Sep 11 00:18:52.195011 kernel: scsi host5: ahci Sep 11 00:18:52.196010 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 00:18:52.196325 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:18:52.202802 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 1 Sep 11 00:18:52.202852 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 1 Sep 11 00:18:52.202880 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 1 Sep 11 00:18:52.202895 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 1 Sep 11 00:18:52.202909 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 1 Sep 11 00:18:52.206113 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 1 Sep 11 00:18:52.513044 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 11 00:18:52.513132 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 11 00:18:52.513915 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 11 00:18:52.517855 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 11 00:18:52.517908 kernel: ata3.00: LPM support broken, forcing max_power Sep 11 00:18:52.517960 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 11 00:18:52.517977 kernel: ata3.00: applying bridge limits Sep 11 00:18:52.517991 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 11 00:18:52.519767 kernel: ata3.00: LPM support broken, forcing max_power Sep 11 00:18:52.519791 kernel: ata3.00: configured for UDMA/100 Sep 11 00:18:52.521591 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 11 00:18:52.539593 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 11 00:18:52.618621 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 11 00:18:52.655180 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 11 00:18:52.664782 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 11 00:18:52.664923 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 11 00:18:52.704698 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 11 00:18:52.705071 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 11 00:18:52.707169 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 11 00:18:52.723192 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 11 00:18:52.739720 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 11 00:18:52.744016 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 11 00:18:52.742823 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 00:18:52.790942 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:18:52.956634 disk-uuid[635]: Primary Header is updated. Sep 11 00:18:52.956634 disk-uuid[635]: Secondary Entries is updated. Sep 11 00:18:52.956634 disk-uuid[635]: Secondary Header is updated. Sep 11 00:18:52.962573 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 00:18:52.967563 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 00:18:53.189158 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 11 00:18:53.191020 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 11 00:18:53.192959 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 00:18:53.193209 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 11 00:18:53.194729 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 11 00:18:53.228079 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 11 00:18:54.004766 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 11 00:18:54.005246 disk-uuid[642]: The operation has completed successfully. Sep 11 00:18:54.038799 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 11 00:18:54.038979 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 11 00:18:54.086269 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 11 00:18:54.116824 sh[668]: Success Sep 11 00:18:54.137619 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 11 00:18:54.137713 kernel: device-mapper: uevent: version 1.0.3 Sep 11 00:18:54.139498 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 11 00:18:54.150590 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Sep 11 00:18:54.188223 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 11 00:18:54.193037 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 11 00:18:54.222011 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 11 00:18:54.230535 kernel: BTRFS: device fsid 1d23f222-37c7-4ff5-813e-235ce83bed46 devid 1 transid 40 /dev/mapper/usr (253:0) scanned by mount (680) Sep 11 00:18:54.230631 kernel: BTRFS info (device dm-0): first mount of filesystem 1d23f222-37c7-4ff5-813e-235ce83bed46 Sep 11 00:18:54.230643 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 11 00:18:54.240060 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 11 00:18:54.240137 kernel: BTRFS info (device dm-0): enabling free space tree Sep 11 00:18:54.242224 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 11 00:18:54.243101 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 11 00:18:54.244365 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 11 00:18:54.245427 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 11 00:18:54.251217 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 11 00:18:54.282580 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (713) Sep 11 00:18:54.285074 kernel: BTRFS info (device vda6): first mount of filesystem dfd585e5-5346-4151-8d09-25f0fad7f81c Sep 11 00:18:54.285108 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 11 00:18:54.288606 kernel: BTRFS info (device vda6): turning on async discard Sep 11 00:18:54.288655 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 00:18:54.294584 kernel: BTRFS info (device vda6): last unmount of filesystem dfd585e5-5346-4151-8d09-25f0fad7f81c Sep 11 00:18:54.295853 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 11 00:18:54.299194 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 11 00:18:54.556281 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 11 00:18:54.560863 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 11 00:18:54.568583 ignition[758]: Ignition 2.21.0 Sep 11 00:18:54.568606 ignition[758]: Stage: fetch-offline Sep 11 00:18:54.568668 ignition[758]: no configs at "/usr/lib/ignition/base.d" Sep 11 00:18:54.568682 ignition[758]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 00:18:54.568821 ignition[758]: parsed url from cmdline: "" Sep 11 00:18:54.568827 ignition[758]: no config URL provided Sep 11 00:18:54.568834 ignition[758]: reading system config file "/usr/lib/ignition/user.ign" Sep 11 00:18:54.568849 ignition[758]: no config at "/usr/lib/ignition/user.ign" Sep 11 00:18:54.568898 ignition[758]: op(1): [started] loading QEMU firmware config module Sep 11 00:18:54.568906 ignition[758]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 11 00:18:54.581278 ignition[758]: op(1): [finished] loading QEMU firmware config module Sep 11 00:18:54.618508 systemd-networkd[858]: lo: Link UP Sep 11 00:18:54.618518 systemd-networkd[858]: lo: Gained carrier Sep 11 00:18:54.620693 systemd-networkd[858]: Enumeration completed Sep 11 00:18:54.621494 systemd-networkd[858]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 00:18:54.621501 systemd-networkd[858]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 11 00:18:54.621666 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 11 00:18:54.622262 systemd-networkd[858]: eth0: Link UP Sep 11 00:18:54.622892 systemd-networkd[858]: eth0: Gained carrier Sep 11 00:18:54.622904 systemd-networkd[858]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 00:18:54.625533 systemd[1]: Reached target network.target - Network. Sep 11 00:18:54.639250 ignition[758]: parsing config with SHA512: ef92d29ac6fb8fbebde52556343031470e295f8ec80fb45bac9ef577abaa1e308406cd29bdf698b1ee424094778b12c82a096eacbbe466d52f4455e9267ca88f Sep 11 00:18:54.641089 systemd-networkd[858]: eth0: DHCPv4 address 10.0.0.82/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 11 00:18:54.644864 unknown[758]: fetched base config from "system" Sep 11 00:18:54.644885 unknown[758]: fetched user config from "qemu" Sep 11 00:18:54.645384 ignition[758]: fetch-offline: fetch-offline passed Sep 11 00:18:54.645448 ignition[758]: Ignition finished successfully Sep 11 00:18:54.649855 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 11 00:18:54.652024 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 11 00:18:54.653099 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 11 00:18:54.758828 ignition[865]: Ignition 2.21.0 Sep 11 00:18:54.758847 ignition[865]: Stage: kargs Sep 11 00:18:54.759016 ignition[865]: no configs at "/usr/lib/ignition/base.d" Sep 11 00:18:54.759031 ignition[865]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 00:18:54.762339 ignition[865]: kargs: kargs passed Sep 11 00:18:54.762495 ignition[865]: Ignition finished successfully Sep 11 00:18:54.771004 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 11 00:18:54.774269 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 11 00:18:54.830261 ignition[873]: Ignition 2.21.0 Sep 11 00:18:54.830278 ignition[873]: Stage: disks Sep 11 00:18:54.830510 ignition[873]: no configs at "/usr/lib/ignition/base.d" Sep 11 00:18:54.830523 ignition[873]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 00:18:54.835256 ignition[873]: disks: disks passed Sep 11 00:18:54.838751 ignition[873]: Ignition finished successfully Sep 11 00:18:54.841940 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 11 00:18:54.843413 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 11 00:18:54.845631 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 11 00:18:54.848150 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 11 00:18:54.848388 systemd[1]: Reached target sysinit.target - System Initialization. Sep 11 00:18:54.848923 systemd[1]: Reached target basic.target - Basic System. Sep 11 00:18:54.850444 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 11 00:18:54.886054 systemd-fsck[883]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 11 00:18:55.578536 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 11 00:18:55.584507 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 11 00:18:55.759597 kernel: EXT4-fs (vda9): mounted filesystem 8ebc908f-0860-41e2-beed-287b778bd592 r/w with ordered data mode. Quota mode: none. Sep 11 00:18:55.761085 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 11 00:18:55.762055 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 11 00:18:55.766187 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 11 00:18:55.770198 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 11 00:18:55.771677 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 11 00:18:55.771745 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 11 00:18:55.771783 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 11 00:18:55.787532 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 11 00:18:55.789888 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 11 00:18:55.796615 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (891) Sep 11 00:18:55.801788 kernel: BTRFS info (device vda6): first mount of filesystem dfd585e5-5346-4151-8d09-25f0fad7f81c Sep 11 00:18:55.802181 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 11 00:18:55.807258 kernel: BTRFS info (device vda6): turning on async discard Sep 11 00:18:55.807364 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 00:18:55.809433 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 11 00:18:55.819437 systemd-networkd[858]: eth0: Gained IPv6LL Sep 11 00:18:55.852154 initrd-setup-root[915]: cut: /sysroot/etc/passwd: No such file or directory Sep 11 00:18:55.858861 initrd-setup-root[922]: cut: /sysroot/etc/group: No such file or directory Sep 11 00:18:55.864828 initrd-setup-root[929]: cut: /sysroot/etc/shadow: No such file or directory Sep 11 00:18:55.871538 initrd-setup-root[936]: cut: /sysroot/etc/gshadow: No such file or directory Sep 11 00:18:56.006261 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 11 00:18:56.010893 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 11 00:18:56.013415 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 11 00:18:56.049698 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 11 00:18:56.050879 kernel: BTRFS info (device vda6): last unmount of filesystem dfd585e5-5346-4151-8d09-25f0fad7f81c Sep 11 00:18:56.086095 ignition[1005]: INFO : Ignition 2.21.0 Sep 11 00:18:56.086095 ignition[1005]: INFO : Stage: mount Sep 11 00:18:56.090350 ignition[1005]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 00:18:56.090350 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 00:18:56.090350 ignition[1005]: INFO : mount: mount passed Sep 11 00:18:56.090350 ignition[1005]: INFO : Ignition finished successfully Sep 11 00:18:56.089101 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 11 00:18:56.091207 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 11 00:18:56.094301 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 11 00:18:56.763872 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 11 00:18:56.805677 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1018) Sep 11 00:18:56.805754 kernel: BTRFS info (device vda6): first mount of filesystem dfd585e5-5346-4151-8d09-25f0fad7f81c Sep 11 00:18:56.805771 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 11 00:18:56.811360 kernel: BTRFS info (device vda6): turning on async discard Sep 11 00:18:56.811443 kernel: BTRFS info (device vda6): enabling free space tree Sep 11 00:18:56.813753 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 11 00:18:56.858951 ignition[1035]: INFO : Ignition 2.21.0 Sep 11 00:18:56.858951 ignition[1035]: INFO : Stage: files Sep 11 00:18:56.861254 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 00:18:56.861254 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 00:18:56.863724 ignition[1035]: DEBUG : files: compiled without relabeling support, skipping Sep 11 00:18:56.864904 ignition[1035]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 11 00:18:56.864904 ignition[1035]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 11 00:18:56.868051 ignition[1035]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 11 00:18:56.868051 ignition[1035]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 11 00:18:56.868051 ignition[1035]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 11 00:18:56.867487 unknown[1035]: wrote ssh authorized keys file for user: core Sep 11 00:18:56.873827 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 11 00:18:56.873827 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Sep 11 00:18:56.952621 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 11 00:18:58.929356 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 11 00:18:58.929356 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 11 00:18:58.933921 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 11 00:18:59.008536 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 11 00:18:59.010839 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 11 00:18:59.010839 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 00:18:59.015966 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 00:18:59.015966 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 00:18:59.015966 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Sep 11 00:18:59.433658 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 11 00:18:59.872158 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 11 00:18:59.872158 ignition[1035]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 11 00:18:59.876114 ignition[1035]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 11 00:19:00.058436 ignition[1035]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 11 00:19:00.058436 ignition[1035]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 11 00:19:00.058436 ignition[1035]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 11 00:19:00.058436 ignition[1035]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 11 00:19:00.067704 ignition[1035]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 11 00:19:00.067704 ignition[1035]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 11 00:19:00.067704 ignition[1035]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 11 00:19:00.080037 ignition[1035]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 11 00:19:00.084718 ignition[1035]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 11 00:19:00.086319 ignition[1035]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 11 00:19:00.086319 ignition[1035]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 11 00:19:00.086319 ignition[1035]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 11 00:19:00.086319 ignition[1035]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 11 00:19:00.086319 ignition[1035]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 11 00:19:00.086319 ignition[1035]: INFO : files: files passed Sep 11 00:19:00.086319 ignition[1035]: INFO : Ignition finished successfully Sep 11 00:19:00.097111 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 11 00:19:00.103218 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 11 00:19:00.104113 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 11 00:19:00.123065 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 11 00:19:00.123238 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 11 00:19:00.128364 initrd-setup-root-after-ignition[1063]: grep: /sysroot/oem/oem-release: No such file or directory Sep 11 00:19:00.133141 initrd-setup-root-after-ignition[1066]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 11 00:19:00.133141 initrd-setup-root-after-ignition[1066]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 11 00:19:00.137011 initrd-setup-root-after-ignition[1070]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 11 00:19:00.140613 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 11 00:19:00.142364 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 11 00:19:00.146309 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 11 00:19:00.226314 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 11 00:19:00.226451 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 11 00:19:00.246525 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 11 00:19:00.249189 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 11 00:19:00.251472 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 11 00:19:00.252865 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 11 00:19:00.280637 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 11 00:19:00.283602 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 11 00:19:00.304626 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 11 00:19:00.305026 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 00:19:00.305536 systemd[1]: Stopped target timers.target - Timer Units. Sep 11 00:19:00.306033 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 11 00:19:00.306155 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 11 00:19:00.312873 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 11 00:19:00.314053 systemd[1]: Stopped target basic.target - Basic System. Sep 11 00:19:00.316790 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 11 00:19:00.317054 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 11 00:19:00.317381 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 11 00:19:00.317887 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 11 00:19:00.318216 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 11 00:19:00.318536 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 11 00:19:00.319046 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 11 00:19:00.319364 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 11 00:19:00.319854 systemd[1]: Stopped target swap.target - Swaps. Sep 11 00:19:00.320152 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 11 00:19:00.320263 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 11 00:19:00.337830 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 11 00:19:00.337975 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 00:19:00.338271 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 11 00:19:00.338375 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 00:19:00.342090 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 11 00:19:00.342214 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 11 00:19:00.344618 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 11 00:19:00.344782 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 11 00:19:00.345191 systemd[1]: Stopped target paths.target - Path Units. Sep 11 00:19:00.349758 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 11 00:19:00.355628 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 00:19:00.355817 systemd[1]: Stopped target slices.target - Slice Units. Sep 11 00:19:00.358407 systemd[1]: Stopped target sockets.target - Socket Units. Sep 11 00:19:00.358881 systemd[1]: iscsid.socket: Deactivated successfully. Sep 11 00:19:00.359008 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 11 00:19:00.362534 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 11 00:19:00.362647 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 11 00:19:00.363519 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 11 00:19:00.363655 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 11 00:19:00.365290 systemd[1]: ignition-files.service: Deactivated successfully. Sep 11 00:19:00.365396 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 11 00:19:00.368095 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 11 00:19:00.369754 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 11 00:19:00.369879 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 00:19:00.371963 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 11 00:19:00.374979 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 11 00:19:00.375137 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 00:19:00.376043 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 11 00:19:00.376177 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 11 00:19:00.385957 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 11 00:19:00.396811 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 11 00:19:00.413931 ignition[1090]: INFO : Ignition 2.21.0 Sep 11 00:19:00.413931 ignition[1090]: INFO : Stage: umount Sep 11 00:19:00.415742 ignition[1090]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 11 00:19:00.415742 ignition[1090]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 11 00:19:00.415742 ignition[1090]: INFO : umount: umount passed Sep 11 00:19:00.415742 ignition[1090]: INFO : Ignition finished successfully Sep 11 00:19:00.419258 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 11 00:19:00.421002 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 11 00:19:00.421157 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 11 00:19:00.423360 systemd[1]: Stopped target network.target - Network. Sep 11 00:19:00.424506 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 11 00:19:00.424608 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 11 00:19:00.435267 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 11 00:19:00.435331 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 11 00:19:00.437347 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 11 00:19:00.437437 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 11 00:19:00.438321 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 11 00:19:00.438397 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 11 00:19:00.441365 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 11 00:19:00.441874 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 11 00:19:00.442489 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 11 00:19:00.442667 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 11 00:19:00.447997 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 11 00:19:00.448116 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 11 00:19:00.454106 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 11 00:19:00.454254 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 11 00:19:00.460428 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 11 00:19:00.460942 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 11 00:19:00.461013 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 00:19:00.467347 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 11 00:19:00.467642 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 11 00:19:00.467787 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 11 00:19:00.472773 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 11 00:19:00.474379 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 11 00:19:00.475510 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 11 00:19:00.475594 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 11 00:19:00.481830 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 11 00:19:00.484817 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 11 00:19:00.484910 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 11 00:19:00.487164 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 11 00:19:00.487241 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 11 00:19:00.492335 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 11 00:19:00.492457 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 11 00:19:00.495915 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 00:19:00.497436 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 11 00:19:00.519850 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 11 00:19:00.520857 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 00:19:00.523014 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 11 00:19:00.523077 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 11 00:19:00.525154 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 11 00:19:00.525205 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 00:19:00.527378 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 11 00:19:00.527451 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 11 00:19:00.530438 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 11 00:19:00.530652 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 11 00:19:00.536659 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 11 00:19:00.536758 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 11 00:19:00.540613 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 11 00:19:00.543948 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 11 00:19:00.545260 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 00:19:00.548458 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 11 00:19:00.548560 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 00:19:00.552897 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 00:19:00.554116 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:19:00.557685 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 11 00:19:00.562731 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 11 00:19:00.572365 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 11 00:19:00.572532 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 11 00:19:00.575590 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 11 00:19:00.578786 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 11 00:19:00.607899 systemd[1]: Switching root. Sep 11 00:19:00.644562 systemd-journald[221]: Received SIGTERM from PID 1 (systemd). Sep 11 00:19:00.644675 systemd-journald[221]: Journal stopped Sep 11 00:19:03.168206 kernel: SELinux: policy capability network_peer_controls=1 Sep 11 00:19:03.168303 kernel: SELinux: policy capability open_perms=1 Sep 11 00:19:03.168337 kernel: SELinux: policy capability extended_socket_class=1 Sep 11 00:19:03.168358 kernel: SELinux: policy capability always_check_network=0 Sep 11 00:19:03.168378 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 11 00:19:03.168398 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 11 00:19:03.168416 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 11 00:19:03.168431 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 11 00:19:03.168447 kernel: SELinux: policy capability userspace_initial_context=0 Sep 11 00:19:03.168463 kernel: audit: type=1403 audit(1757549942.088:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 11 00:19:03.168487 systemd[1]: Successfully loaded SELinux policy in 66.824ms. Sep 11 00:19:03.168526 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 9.370ms. Sep 11 00:19:03.168567 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 11 00:19:03.168585 systemd[1]: Detected virtualization kvm. Sep 11 00:19:03.168603 systemd[1]: Detected architecture x86-64. Sep 11 00:19:03.168619 systemd[1]: Detected first boot. Sep 11 00:19:03.168634 systemd[1]: Initializing machine ID from VM UUID. Sep 11 00:19:03.168660 kernel: Guest personality initialized and is inactive Sep 11 00:19:03.168677 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 11 00:19:03.168692 kernel: Initialized host personality Sep 11 00:19:03.168716 kernel: NET: Registered PF_VSOCK protocol family Sep 11 00:19:03.168734 zram_generator::config[1139]: No configuration found. Sep 11 00:19:03.168752 systemd[1]: Populated /etc with preset unit settings. Sep 11 00:19:03.168769 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 11 00:19:03.168785 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 11 00:19:03.168801 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 11 00:19:03.168817 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 11 00:19:03.168834 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 11 00:19:03.168860 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 11 00:19:03.168876 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 11 00:19:03.168892 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 11 00:19:03.168909 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 11 00:19:03.168926 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 11 00:19:03.168943 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 11 00:19:03.168959 systemd[1]: Created slice user.slice - User and Session Slice. Sep 11 00:19:03.168978 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 11 00:19:03.168996 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 11 00:19:03.169019 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 11 00:19:03.169036 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 11 00:19:03.169053 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 11 00:19:03.169070 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 11 00:19:03.169087 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 11 00:19:03.169103 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 11 00:19:03.169119 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 11 00:19:03.169148 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 11 00:19:03.169165 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 11 00:19:03.169180 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 11 00:19:03.169196 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 11 00:19:03.169212 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 11 00:19:03.169228 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 11 00:19:03.169244 systemd[1]: Reached target slices.target - Slice Units. Sep 11 00:19:03.169260 systemd[1]: Reached target swap.target - Swaps. Sep 11 00:19:03.169276 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 11 00:19:03.169299 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 11 00:19:03.169316 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 11 00:19:03.169332 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 11 00:19:03.169351 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 11 00:19:03.169367 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 11 00:19:03.169383 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 11 00:19:03.169398 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 11 00:19:03.169414 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 11 00:19:03.169430 systemd[1]: Mounting media.mount - External Media Directory... Sep 11 00:19:03.169447 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 00:19:03.169469 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 11 00:19:03.169485 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 11 00:19:03.169501 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 11 00:19:03.169518 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 11 00:19:03.169534 systemd[1]: Reached target machines.target - Containers. Sep 11 00:19:03.169568 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 11 00:19:03.169585 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 00:19:03.169601 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 11 00:19:03.169625 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 11 00:19:03.169641 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 00:19:03.169667 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 11 00:19:03.169682 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 00:19:03.169698 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 11 00:19:03.169714 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 00:19:03.169733 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 11 00:19:03.169749 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 11 00:19:03.169778 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 11 00:19:03.169794 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 11 00:19:03.169810 systemd[1]: Stopped systemd-fsck-usr.service. Sep 11 00:19:03.169827 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 00:19:03.169844 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 11 00:19:03.169860 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 11 00:19:03.169882 kernel: ACPI: bus type drm_connector registered Sep 11 00:19:03.169898 kernel: fuse: init (API version 7.41) Sep 11 00:19:03.169914 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 11 00:19:03.169929 kernel: loop: module loaded Sep 11 00:19:03.169945 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 11 00:19:03.169961 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 11 00:19:03.169977 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 11 00:19:03.169993 systemd[1]: verity-setup.service: Deactivated successfully. Sep 11 00:19:03.170017 systemd[1]: Stopped verity-setup.service. Sep 11 00:19:03.170034 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 00:19:03.170051 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 11 00:19:03.170096 systemd-journald[1213]: Collecting audit messages is disabled. Sep 11 00:19:03.170137 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 11 00:19:03.170153 systemd[1]: Mounted media.mount - External Media Directory. Sep 11 00:19:03.170166 systemd-journald[1213]: Journal started Sep 11 00:19:03.170188 systemd-journald[1213]: Runtime Journal (/run/log/journal/837ce06a809f4e869b3fbd9a7e8872bd) is 6M, max 48.4M, 42.4M free. Sep 11 00:19:02.886213 systemd[1]: Queued start job for default target multi-user.target. Sep 11 00:19:02.912124 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 11 00:19:02.912822 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 11 00:19:03.172575 systemd[1]: Started systemd-journald.service - Journal Service. Sep 11 00:19:03.173782 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 11 00:19:03.174935 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 11 00:19:03.176135 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 11 00:19:03.177373 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 11 00:19:03.178942 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 11 00:19:03.180403 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 11 00:19:03.180783 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 11 00:19:03.182198 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 00:19:03.182407 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 00:19:03.183816 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 11 00:19:03.184022 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 11 00:19:03.185331 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 00:19:03.185536 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 00:19:03.187145 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 11 00:19:03.187354 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 11 00:19:03.188838 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 00:19:03.189125 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 00:19:03.190658 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 11 00:19:03.192267 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 11 00:19:03.194082 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 11 00:19:03.195922 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 11 00:19:03.212212 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 11 00:19:03.215167 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 11 00:19:03.219658 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 11 00:19:03.221093 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 11 00:19:03.221134 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 11 00:19:03.223259 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 11 00:19:03.231653 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 11 00:19:03.233844 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 00:19:03.235489 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 11 00:19:03.238307 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 11 00:19:03.239800 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 11 00:19:03.241162 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 11 00:19:03.243480 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 11 00:19:03.246047 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 11 00:19:03.248881 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 11 00:19:03.254225 systemd-journald[1213]: Time spent on flushing to /var/log/journal/837ce06a809f4e869b3fbd9a7e8872bd is 25.454ms for 1068 entries. Sep 11 00:19:03.254225 systemd-journald[1213]: System Journal (/var/log/journal/837ce06a809f4e869b3fbd9a7e8872bd) is 8M, max 195.6M, 187.6M free. Sep 11 00:19:03.747113 systemd-journald[1213]: Received client request to flush runtime journal. Sep 11 00:19:03.747176 kernel: loop0: detected capacity change from 0 to 128016 Sep 11 00:19:03.747209 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 11 00:19:03.747227 kernel: loop1: detected capacity change from 0 to 229808 Sep 11 00:19:03.747245 kernel: loop2: detected capacity change from 0 to 111000 Sep 11 00:19:03.747269 kernel: loop3: detected capacity change from 0 to 128016 Sep 11 00:19:03.747287 kernel: loop4: detected capacity change from 0 to 229808 Sep 11 00:19:03.747305 kernel: loop5: detected capacity change from 0 to 111000 Sep 11 00:19:03.747324 zram_generator::config[1300]: No configuration found. Sep 11 00:19:03.255369 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 11 00:19:03.259678 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 11 00:19:03.262722 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 11 00:19:03.272286 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 11 00:19:03.398610 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 11 00:19:03.430996 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 11 00:19:03.435688 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 11 00:19:03.439610 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 11 00:19:03.441420 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 11 00:19:03.450742 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 11 00:19:03.590698 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Sep 11 00:19:03.590711 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Sep 11 00:19:03.594922 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 11 00:19:03.667208 (sd-merge)[1275]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 11 00:19:03.667798 (sd-merge)[1275]: Merged extensions into '/usr'. Sep 11 00:19:03.672648 systemd[1]: Reload requested from client PID 1257 ('systemd-sysext') (unit systemd-sysext.service)... Sep 11 00:19:03.672660 systemd[1]: Reloading... Sep 11 00:19:03.952729 systemd[1]: Reloading finished in 279 ms. Sep 11 00:19:03.990799 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 11 00:19:04.000478 ldconfig[1252]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 11 00:19:04.010067 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 11 00:19:04.011961 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 11 00:19:04.013985 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 11 00:19:04.016017 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 11 00:19:04.032684 systemd[1]: Starting ensure-sysext.service... Sep 11 00:19:04.035355 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 11 00:19:04.060924 systemd[1]: Reload requested from client PID 1344 ('systemctl') (unit ensure-sysext.service)... Sep 11 00:19:04.060948 systemd[1]: Reloading... Sep 11 00:19:04.073271 systemd-tmpfiles[1345]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 11 00:19:04.073317 systemd-tmpfiles[1345]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 11 00:19:04.073741 systemd-tmpfiles[1345]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 11 00:19:04.074102 systemd-tmpfiles[1345]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 11 00:19:04.075355 systemd-tmpfiles[1345]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 11 00:19:04.075759 systemd-tmpfiles[1345]: ACLs are not supported, ignoring. Sep 11 00:19:04.076049 systemd-tmpfiles[1345]: ACLs are not supported, ignoring. Sep 11 00:19:04.082771 systemd-tmpfiles[1345]: Detected autofs mount point /boot during canonicalization of boot. Sep 11 00:19:04.082786 systemd-tmpfiles[1345]: Skipping /boot Sep 11 00:19:04.095843 systemd-tmpfiles[1345]: Detected autofs mount point /boot during canonicalization of boot. Sep 11 00:19:04.095986 systemd-tmpfiles[1345]: Skipping /boot Sep 11 00:19:04.133580 zram_generator::config[1369]: No configuration found. Sep 11 00:19:04.370378 systemd[1]: Reloading finished in 309 ms. Sep 11 00:19:04.425689 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 11 00:19:04.435011 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 11 00:19:04.500564 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 11 00:19:04.506918 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 11 00:19:04.515474 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 11 00:19:04.518931 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 11 00:19:04.539807 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 11 00:19:04.548046 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 00:19:04.548323 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 11 00:19:04.549918 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 11 00:19:04.552146 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 11 00:19:04.554271 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 11 00:19:04.561533 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 11 00:19:04.564512 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 11 00:19:04.564683 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 11 00:19:04.568813 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 11 00:19:04.573793 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 11 00:19:04.574497 augenrules[1441]: No rules Sep 11 00:19:04.575632 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 11 00:19:04.578699 systemd[1]: audit-rules.service: Deactivated successfully. Sep 11 00:19:04.581872 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 11 00:19:04.584055 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 11 00:19:04.586289 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 11 00:19:04.586588 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 11 00:19:04.589519 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 11 00:19:04.590137 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 11 00:19:04.591996 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 11 00:19:04.592397 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 11 00:19:04.594387 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 11 00:19:04.594705 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 11 00:19:04.601241 systemd[1]: Finished ensure-sysext.service. Sep 11 00:19:04.602958 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 11 00:19:04.613634 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 11 00:19:04.613847 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 11 00:19:04.616156 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 11 00:19:04.618966 systemd-udevd[1438]: Using default interface naming scheme 'v255'. Sep 11 00:19:04.620817 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 11 00:19:04.622131 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 11 00:19:04.627352 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 11 00:19:04.639781 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 11 00:19:04.646778 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 11 00:19:04.649248 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 11 00:19:04.657806 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 11 00:19:04.755456 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 11 00:19:04.801525 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 11 00:19:04.807580 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 11 00:19:04.811566 kernel: mousedev: PS/2 mouse device common for all mice Sep 11 00:19:04.835608 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 11 00:19:04.842919 kernel: ACPI: button: Power Button [PWRF] Sep 11 00:19:04.841080 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 11 00:19:04.880862 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Sep 11 00:19:04.884882 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 11 00:19:04.885089 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 11 00:19:04.879001 systemd-networkd[1474]: lo: Link UP Sep 11 00:19:04.879013 systemd-networkd[1474]: lo: Gained carrier Sep 11 00:19:04.880836 systemd-networkd[1474]: Enumeration completed Sep 11 00:19:04.880965 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 11 00:19:04.881238 systemd-networkd[1474]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 00:19:04.881242 systemd-networkd[1474]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 11 00:19:04.882589 systemd-networkd[1474]: eth0: Link UP Sep 11 00:19:04.882772 systemd-networkd[1474]: eth0: Gained carrier Sep 11 00:19:04.882786 systemd-networkd[1474]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 11 00:19:04.885876 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 11 00:19:04.890350 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 11 00:19:04.896623 systemd-networkd[1474]: eth0: DHCPv4 address 10.0.0.82/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 11 00:19:04.919945 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 11 00:19:04.921862 systemd[1]: Reached target time-set.target - System Time Set. Sep 11 00:19:05.895911 systemd-timesyncd[1454]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 11 00:19:05.896239 systemd-timesyncd[1454]: Initial clock synchronization to Thu 2025-09-11 00:19:05.895813 UTC. Sep 11 00:19:05.897277 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 11 00:19:05.928288 systemd-resolved[1418]: Positive Trust Anchors: Sep 11 00:19:05.928321 systemd-resolved[1418]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 11 00:19:05.928363 systemd-resolved[1418]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 11 00:19:05.936140 systemd-resolved[1418]: Defaulting to hostname 'linux'. Sep 11 00:19:05.936669 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 00:19:05.939565 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 11 00:19:05.941796 systemd[1]: Reached target network.target - Network. Sep 11 00:19:05.942882 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 11 00:19:05.993775 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 11 00:19:05.994184 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:19:05.997400 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 11 00:19:06.054436 kernel: kvm_amd: TSC scaling supported Sep 11 00:19:06.054536 kernel: kvm_amd: Nested Virtualization enabled Sep 11 00:19:06.054571 kernel: kvm_amd: Nested Paging enabled Sep 11 00:19:06.055857 kernel: kvm_amd: LBR virtualization supported Sep 11 00:19:06.055890 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 11 00:19:06.057130 kernel: kvm_amd: Virtual GIF supported Sep 11 00:19:06.093629 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 11 00:19:06.095737 systemd[1]: Reached target sysinit.target - System Initialization. Sep 11 00:19:06.097316 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 11 00:19:06.098855 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 11 00:19:06.100491 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Sep 11 00:19:06.102069 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 11 00:19:06.103576 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 11 00:19:06.105658 kernel: EDAC MC: Ver: 3.0.0 Sep 11 00:19:06.105729 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 11 00:19:06.107333 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 11 00:19:06.107381 systemd[1]: Reached target paths.target - Path Units. Sep 11 00:19:06.108675 systemd[1]: Reached target timers.target - Timer Units. Sep 11 00:19:06.111027 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 11 00:19:06.114493 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 11 00:19:06.118832 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 11 00:19:06.120390 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 11 00:19:06.121945 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 11 00:19:06.131028 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 11 00:19:06.133116 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 11 00:19:06.135298 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 11 00:19:06.137280 systemd[1]: Reached target sockets.target - Socket Units. Sep 11 00:19:06.138399 systemd[1]: Reached target basic.target - Basic System. Sep 11 00:19:06.139514 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 11 00:19:06.139582 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 11 00:19:06.140811 systemd[1]: Starting containerd.service - containerd container runtime... Sep 11 00:19:06.143432 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 11 00:19:06.145604 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 11 00:19:06.147930 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 11 00:19:06.155926 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 11 00:19:06.157026 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 11 00:19:06.158607 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Sep 11 00:19:06.161126 jq[1541]: false Sep 11 00:19:06.161866 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 11 00:19:06.163938 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 11 00:19:06.170683 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 11 00:19:06.174037 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 11 00:19:06.180534 google_oslogin_nss_cache[1543]: oslogin_cache_refresh[1543]: Refreshing passwd entry cache Sep 11 00:19:06.180568 oslogin_cache_refresh[1543]: Refreshing passwd entry cache Sep 11 00:19:06.180737 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 11 00:19:06.183075 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 11 00:19:06.186940 extend-filesystems[1542]: Found /dev/vda6 Sep 11 00:19:06.190467 extend-filesystems[1542]: Found /dev/vda9 Sep 11 00:19:06.191680 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 11 00:19:06.192924 systemd[1]: Starting update-engine.service - Update Engine... Sep 11 00:19:06.193947 google_oslogin_nss_cache[1543]: oslogin_cache_refresh[1543]: Failure getting users, quitting Sep 11 00:19:06.193947 google_oslogin_nss_cache[1543]: oslogin_cache_refresh[1543]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 11 00:19:06.193947 google_oslogin_nss_cache[1543]: oslogin_cache_refresh[1543]: Refreshing group entry cache Sep 11 00:19:06.193827 oslogin_cache_refresh[1543]: Failure getting users, quitting Sep 11 00:19:06.193855 oslogin_cache_refresh[1543]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 11 00:19:06.193934 oslogin_cache_refresh[1543]: Refreshing group entry cache Sep 11 00:19:06.194112 extend-filesystems[1542]: Checking size of /dev/vda9 Sep 11 00:19:06.196721 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 11 00:19:06.202150 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 11 00:19:06.204201 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 11 00:19:06.204873 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 11 00:19:06.206066 oslogin_cache_refresh[1543]: Failure getting groups, quitting Sep 11 00:19:06.207766 google_oslogin_nss_cache[1543]: oslogin_cache_refresh[1543]: Failure getting groups, quitting Sep 11 00:19:06.207766 google_oslogin_nss_cache[1543]: oslogin_cache_refresh[1543]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 11 00:19:06.205260 systemd[1]: motdgen.service: Deactivated successfully. Sep 11 00:19:06.206085 oslogin_cache_refresh[1543]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 11 00:19:06.205507 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 11 00:19:06.210207 extend-filesystems[1542]: Resized partition /dev/vda9 Sep 11 00:19:06.214217 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Sep 11 00:19:06.215855 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Sep 11 00:19:06.217186 extend-filesystems[1569]: resize2fs 1.47.2 (1-Jan-2025) Sep 11 00:19:06.217786 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 11 00:19:06.218044 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 11 00:19:06.220933 jq[1561]: true Sep 11 00:19:06.225478 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 11 00:19:06.235567 jq[1570]: true Sep 11 00:19:06.252080 (ntainerd)[1571]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 11 00:19:06.258380 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 11 00:19:06.259953 update_engine[1559]: I20250911 00:19:06.253013 1559 main.cc:92] Flatcar Update Engine starting Sep 11 00:19:06.268667 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 11 00:19:06.274853 tar[1565]: linux-amd64/LICENSE Sep 11 00:19:06.293232 tar[1565]: linux-amd64/helm Sep 11 00:19:06.293404 extend-filesystems[1569]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 11 00:19:06.293404 extend-filesystems[1569]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 11 00:19:06.293404 extend-filesystems[1569]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 11 00:19:06.297204 extend-filesystems[1542]: Resized filesystem in /dev/vda9 Sep 11 00:19:06.297024 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 11 00:19:06.297382 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 11 00:19:06.320493 dbus-daemon[1539]: [system] SELinux support is enabled Sep 11 00:19:06.321298 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 11 00:19:06.324612 bash[1602]: Updated "/home/core/.ssh/authorized_keys" Sep 11 00:19:06.326101 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 11 00:19:06.326378 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 11 00:19:06.328745 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 11 00:19:06.328768 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 11 00:19:06.330452 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 11 00:19:06.332440 systemd-logind[1552]: Watching system buttons on /dev/input/event2 (Power Button) Sep 11 00:19:06.334528 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 11 00:19:06.334784 systemd-logind[1552]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 11 00:19:06.335982 systemd-logind[1552]: New seat seat0. Sep 11 00:19:06.336080 update_engine[1559]: I20250911 00:19:06.335829 1559 update_check_scheduler.cc:74] Next update check in 8m14s Sep 11 00:19:06.339631 systemd[1]: Started update-engine.service - Update Engine. Sep 11 00:19:06.341017 systemd[1]: Started systemd-logind.service - User Login Management. Sep 11 00:19:06.347826 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 11 00:19:06.404286 sshd_keygen[1575]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 11 00:19:06.411944 locksmithd[1605]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 11 00:19:06.443116 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 11 00:19:06.446539 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 11 00:19:06.449718 systemd[1]: Started sshd@0-10.0.0.82:22-10.0.0.1:51584.service - OpenSSH per-connection server daemon (10.0.0.1:51584). Sep 11 00:19:06.476056 systemd[1]: issuegen.service: Deactivated successfully. Sep 11 00:19:06.476439 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 11 00:19:06.480174 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 11 00:19:06.510319 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 11 00:19:06.645648 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 11 00:19:06.648892 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 11 00:19:06.650536 systemd[1]: Reached target getty.target - Login Prompts. Sep 11 00:19:06.687618 containerd[1571]: time="2025-09-11T00:19:06Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 11 00:19:06.688687 containerd[1571]: time="2025-09-11T00:19:06.688653755Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 11 00:19:06.702589 containerd[1571]: time="2025-09-11T00:19:06.702493154Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="10.42µs" Sep 11 00:19:06.702757 containerd[1571]: time="2025-09-11T00:19:06.702736641Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 11 00:19:06.702813 containerd[1571]: time="2025-09-11T00:19:06.702800871Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 11 00:19:06.703132 containerd[1571]: time="2025-09-11T00:19:06.703112506Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 11 00:19:06.703191 containerd[1571]: time="2025-09-11T00:19:06.703179572Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 11 00:19:06.703253 containerd[1571]: time="2025-09-11T00:19:06.703242169Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 11 00:19:06.703364 containerd[1571]: time="2025-09-11T00:19:06.703347366Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 11 00:19:06.703411 containerd[1571]: time="2025-09-11T00:19:06.703400115Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 11 00:19:06.703800 containerd[1571]: time="2025-09-11T00:19:06.703778936Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 11 00:19:06.703853 containerd[1571]: time="2025-09-11T00:19:06.703842034Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 11 00:19:06.703900 containerd[1571]: time="2025-09-11T00:19:06.703887981Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 11 00:19:06.703942 containerd[1571]: time="2025-09-11T00:19:06.703931953Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 11 00:19:06.704113 containerd[1571]: time="2025-09-11T00:19:06.704092404Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 11 00:19:06.704436 containerd[1571]: time="2025-09-11T00:19:06.704417484Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 11 00:19:06.704511 containerd[1571]: time="2025-09-11T00:19:06.704497153Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 11 00:19:06.704598 containerd[1571]: time="2025-09-11T00:19:06.704574408Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 11 00:19:06.704684 containerd[1571]: time="2025-09-11T00:19:06.704669376Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 11 00:19:06.705133 containerd[1571]: time="2025-09-11T00:19:06.705048457Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 11 00:19:06.705343 containerd[1571]: time="2025-09-11T00:19:06.705306141Z" level=info msg="metadata content store policy set" policy=shared Sep 11 00:19:06.723015 tar[1565]: linux-amd64/README.md Sep 11 00:19:06.758947 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 11 00:19:06.952337 sshd[1625]: Accepted publickey for core from 10.0.0.1 port 51584 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:06.955389 sshd-session[1625]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:06.963483 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 11 00:19:06.966471 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 11 00:19:06.976046 systemd-logind[1552]: New session 1 of user core. Sep 11 00:19:07.014417 containerd[1571]: time="2025-09-11T00:19:07.014335431Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 11 00:19:07.014417 containerd[1571]: time="2025-09-11T00:19:07.014412315Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 11 00:19:07.014644 containerd[1571]: time="2025-09-11T00:19:07.014450737Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 11 00:19:07.014644 containerd[1571]: time="2025-09-11T00:19:07.014500761Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 11 00:19:07.014644 containerd[1571]: time="2025-09-11T00:19:07.014566455Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 11 00:19:07.014644 containerd[1571]: time="2025-09-11T00:19:07.014595259Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 11 00:19:07.014644 containerd[1571]: time="2025-09-11T00:19:07.014628230Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 11 00:19:07.014775 containerd[1571]: time="2025-09-11T00:19:07.014646134Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 11 00:19:07.014775 containerd[1571]: time="2025-09-11T00:19:07.014661543Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 11 00:19:07.014775 containerd[1571]: time="2025-09-11T00:19:07.014683654Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 11 00:19:07.014775 containerd[1571]: time="2025-09-11T00:19:07.014696368Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 11 00:19:07.014775 containerd[1571]: time="2025-09-11T00:19:07.014713530Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 11 00:19:07.014982 containerd[1571]: time="2025-09-11T00:19:07.014940135Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 11 00:19:07.014982 containerd[1571]: time="2025-09-11T00:19:07.014977115Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 11 00:19:07.015055 containerd[1571]: time="2025-09-11T00:19:07.014998755Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 11 00:19:07.015055 containerd[1571]: time="2025-09-11T00:19:07.015014886Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 11 00:19:07.015055 containerd[1571]: time="2025-09-11T00:19:07.015046535Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 11 00:19:07.015131 containerd[1571]: time="2025-09-11T00:19:07.015061984Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 11 00:19:07.015131 containerd[1571]: time="2025-09-11T00:19:07.015077643Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 11 00:19:07.015131 containerd[1571]: time="2025-09-11T00:19:07.015090728Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 11 00:19:07.015131 containerd[1571]: time="2025-09-11T00:19:07.015107419Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 11 00:19:07.015131 containerd[1571]: time="2025-09-11T00:19:07.015121546Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 11 00:19:07.015283 containerd[1571]: time="2025-09-11T00:19:07.015135001Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 11 00:19:07.015283 containerd[1571]: time="2025-09-11T00:19:07.015236441Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 11 00:19:07.015283 containerd[1571]: time="2025-09-11T00:19:07.015258082Z" level=info msg="Start snapshots syncer" Sep 11 00:19:07.015370 containerd[1571]: time="2025-09-11T00:19:07.015311883Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 11 00:19:07.015720 containerd[1571]: time="2025-09-11T00:19:07.015654195Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 11 00:19:07.015872 containerd[1571]: time="2025-09-11T00:19:07.015722433Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 11 00:19:07.015872 containerd[1571]: time="2025-09-11T00:19:07.015803325Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 11 00:19:07.015969 containerd[1571]: time="2025-09-11T00:19:07.015941524Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 11 00:19:07.016045 containerd[1571]: time="2025-09-11T00:19:07.015972342Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 11 00:19:07.016045 containerd[1571]: time="2025-09-11T00:19:07.015994152Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 11 00:19:07.016045 containerd[1571]: time="2025-09-11T00:19:07.016007738Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 11 00:19:07.016045 containerd[1571]: time="2025-09-11T00:19:07.016023578Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 11 00:19:07.016045 containerd[1571]: time="2025-09-11T00:19:07.016036833Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 11 00:19:07.016140 containerd[1571]: time="2025-09-11T00:19:07.016048805Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 11 00:19:07.016140 containerd[1571]: time="2025-09-11T00:19:07.016073331Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 11 00:19:07.016140 containerd[1571]: time="2025-09-11T00:19:07.016085734Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 11 00:19:07.016140 containerd[1571]: time="2025-09-11T00:19:07.016098328Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016139435Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016158731Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016170062Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016182325Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016192464Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016204497Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016217501Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 11 00:19:07.016235 containerd[1571]: time="2025-09-11T00:19:07.016239413Z" level=info msg="runtime interface created" Sep 11 00:19:07.016427 containerd[1571]: time="2025-09-11T00:19:07.016247037Z" level=info msg="created NRI interface" Sep 11 00:19:07.016427 containerd[1571]: time="2025-09-11T00:19:07.016271944Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 11 00:19:07.016427 containerd[1571]: time="2025-09-11T00:19:07.016287042Z" level=info msg="Connect containerd service" Sep 11 00:19:07.016427 containerd[1571]: time="2025-09-11T00:19:07.016313592Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 11 00:19:07.018957 containerd[1571]: time="2025-09-11T00:19:07.018904251Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 11 00:19:07.032720 systemd-networkd[1474]: eth0: Gained IPv6LL Sep 11 00:19:07.043728 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 11 00:19:07.048464 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 11 00:19:07.051331 systemd[1]: Reached target network-online.target - Network is Online. Sep 11 00:19:07.055428 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 11 00:19:07.058340 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:07.079973 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 11 00:19:07.084015 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 11 00:19:07.466705 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 11 00:19:07.479992 (systemd)[1649]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 11 00:19:07.484844 systemd-logind[1552]: New session c1 of user core. Sep 11 00:19:07.493123 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 11 00:19:07.494463 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 11 00:19:07.496504 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 11 00:19:08.162439 containerd[1571]: time="2025-09-11T00:19:08.162338447Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 11 00:19:08.162439 containerd[1571]: time="2025-09-11T00:19:08.162443293Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162473089Z" level=info msg="Start subscribing containerd event" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162501553Z" level=info msg="Start recovering state" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162634452Z" level=info msg="Start event monitor" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162647997Z" level=info msg="Start cni network conf syncer for default" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162654219Z" level=info msg="Start streaming server" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162662284Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162669438Z" level=info msg="runtime interface starting up..." Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162675379Z" level=info msg="starting plugins..." Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162694334Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 11 00:19:08.163044 containerd[1571]: time="2025-09-11T00:19:08.162832433Z" level=info msg="containerd successfully booted in 1.477143s" Sep 11 00:19:08.163678 systemd[1]: Started containerd.service - containerd container runtime. Sep 11 00:19:08.263796 systemd[1649]: Queued start job for default target default.target. Sep 11 00:19:08.300180 systemd[1649]: Created slice app.slice - User Application Slice. Sep 11 00:19:08.300213 systemd[1649]: Reached target paths.target - Paths. Sep 11 00:19:08.300273 systemd[1649]: Reached target timers.target - Timers. Sep 11 00:19:08.302134 systemd[1649]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 11 00:19:08.316745 systemd[1649]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 11 00:19:08.316935 systemd[1649]: Reached target sockets.target - Sockets. Sep 11 00:19:08.317011 systemd[1649]: Reached target basic.target - Basic System. Sep 11 00:19:08.317065 systemd[1649]: Reached target default.target - Main User Target. Sep 11 00:19:08.317127 systemd[1649]: Startup finished in 821ms. Sep 11 00:19:08.317217 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 11 00:19:08.320438 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 11 00:19:08.393730 systemd[1]: Started sshd@1-10.0.0.82:22-10.0.0.1:51594.service - OpenSSH per-connection server daemon (10.0.0.1:51594). Sep 11 00:19:08.462075 sshd[1685]: Accepted publickey for core from 10.0.0.1 port 51594 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:08.464045 sshd-session[1685]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:08.469882 systemd-logind[1552]: New session 2 of user core. Sep 11 00:19:08.541843 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 11 00:19:08.624822 sshd[1688]: Connection closed by 10.0.0.1 port 51594 Sep 11 00:19:08.626616 sshd-session[1685]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:08.635746 systemd[1]: sshd@1-10.0.0.82:22-10.0.0.1:51594.service: Deactivated successfully. Sep 11 00:19:08.637818 systemd[1]: session-2.scope: Deactivated successfully. Sep 11 00:19:08.638812 systemd-logind[1552]: Session 2 logged out. Waiting for processes to exit. Sep 11 00:19:08.642052 systemd[1]: Started sshd@2-10.0.0.82:22-10.0.0.1:51598.service - OpenSSH per-connection server daemon (10.0.0.1:51598). Sep 11 00:19:08.644597 systemd-logind[1552]: Removed session 2. Sep 11 00:19:08.732680 sshd[1694]: Accepted publickey for core from 10.0.0.1 port 51598 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:08.734729 sshd-session[1694]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:08.740307 systemd-logind[1552]: New session 3 of user core. Sep 11 00:19:08.765845 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 11 00:19:08.841994 sshd[1697]: Connection closed by 10.0.0.1 port 51598 Sep 11 00:19:08.842916 sshd-session[1694]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:08.848933 systemd[1]: sshd@2-10.0.0.82:22-10.0.0.1:51598.service: Deactivated successfully. Sep 11 00:19:08.851645 systemd[1]: session-3.scope: Deactivated successfully. Sep 11 00:19:08.852529 systemd-logind[1552]: Session 3 logged out. Waiting for processes to exit. Sep 11 00:19:08.854254 systemd-logind[1552]: Removed session 3. Sep 11 00:19:09.935615 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:09.937631 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 11 00:19:09.939246 systemd[1]: Startup finished in 4.927s (kernel) + 11.435s (initrd) + 6.944s (userspace) = 23.307s. Sep 11 00:19:09.956358 (kubelet)[1707]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 00:19:10.691798 kubelet[1707]: E0911 00:19:10.691707 1707 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 00:19:10.696405 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 00:19:10.696692 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 00:19:10.697218 systemd[1]: kubelet.service: Consumed 3.317s CPU time, 267.3M memory peak. Sep 11 00:19:18.867740 systemd[1]: Started sshd@3-10.0.0.82:22-10.0.0.1:33826.service - OpenSSH per-connection server daemon (10.0.0.1:33826). Sep 11 00:19:18.936249 sshd[1720]: Accepted publickey for core from 10.0.0.1 port 33826 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:18.938787 sshd-session[1720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:18.944271 systemd-logind[1552]: New session 4 of user core. Sep 11 00:19:18.965795 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 11 00:19:19.023322 sshd[1723]: Connection closed by 10.0.0.1 port 33826 Sep 11 00:19:19.023908 sshd-session[1720]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:19.043177 systemd[1]: sshd@3-10.0.0.82:22-10.0.0.1:33826.service: Deactivated successfully. Sep 11 00:19:19.045423 systemd[1]: session-4.scope: Deactivated successfully. Sep 11 00:19:19.046398 systemd-logind[1552]: Session 4 logged out. Waiting for processes to exit. Sep 11 00:19:19.050652 systemd[1]: Started sshd@4-10.0.0.82:22-10.0.0.1:33836.service - OpenSSH per-connection server daemon (10.0.0.1:33836). Sep 11 00:19:19.051424 systemd-logind[1552]: Removed session 4. Sep 11 00:19:19.117448 sshd[1729]: Accepted publickey for core from 10.0.0.1 port 33836 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:19.119324 sshd-session[1729]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:19.125047 systemd-logind[1552]: New session 5 of user core. Sep 11 00:19:19.146984 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 11 00:19:19.201147 sshd[1732]: Connection closed by 10.0.0.1 port 33836 Sep 11 00:19:19.201579 sshd-session[1729]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:19.215172 systemd[1]: sshd@4-10.0.0.82:22-10.0.0.1:33836.service: Deactivated successfully. Sep 11 00:19:19.217795 systemd[1]: session-5.scope: Deactivated successfully. Sep 11 00:19:19.218866 systemd-logind[1552]: Session 5 logged out. Waiting for processes to exit. Sep 11 00:19:19.223059 systemd[1]: Started sshd@5-10.0.0.82:22-10.0.0.1:33850.service - OpenSSH per-connection server daemon (10.0.0.1:33850). Sep 11 00:19:19.223983 systemd-logind[1552]: Removed session 5. Sep 11 00:19:19.287265 sshd[1738]: Accepted publickey for core from 10.0.0.1 port 33850 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:19.289161 sshd-session[1738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:19.295093 systemd-logind[1552]: New session 6 of user core. Sep 11 00:19:19.304910 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 11 00:19:19.362466 sshd[1741]: Connection closed by 10.0.0.1 port 33850 Sep 11 00:19:19.362911 sshd-session[1738]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:19.376595 systemd[1]: sshd@5-10.0.0.82:22-10.0.0.1:33850.service: Deactivated successfully. Sep 11 00:19:19.378406 systemd[1]: session-6.scope: Deactivated successfully. Sep 11 00:19:19.379177 systemd-logind[1552]: Session 6 logged out. Waiting for processes to exit. Sep 11 00:19:19.382022 systemd[1]: Started sshd@6-10.0.0.82:22-10.0.0.1:33860.service - OpenSSH per-connection server daemon (10.0.0.1:33860). Sep 11 00:19:19.383004 systemd-logind[1552]: Removed session 6. Sep 11 00:19:19.431155 sshd[1747]: Accepted publickey for core from 10.0.0.1 port 33860 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:19.433166 sshd-session[1747]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:19.439253 systemd-logind[1552]: New session 7 of user core. Sep 11 00:19:19.453890 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 11 00:19:19.515950 sudo[1751]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 11 00:19:19.516292 sudo[1751]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 00:19:19.538957 sudo[1751]: pam_unix(sudo:session): session closed for user root Sep 11 00:19:19.540923 sshd[1750]: Connection closed by 10.0.0.1 port 33860 Sep 11 00:19:19.541342 sshd-session[1747]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:19.557936 systemd[1]: sshd@6-10.0.0.82:22-10.0.0.1:33860.service: Deactivated successfully. Sep 11 00:19:19.561752 systemd[1]: session-7.scope: Deactivated successfully. Sep 11 00:19:19.562732 systemd-logind[1552]: Session 7 logged out. Waiting for processes to exit. Sep 11 00:19:19.566065 systemd[1]: Started sshd@7-10.0.0.82:22-10.0.0.1:33874.service - OpenSSH per-connection server daemon (10.0.0.1:33874). Sep 11 00:19:19.566965 systemd-logind[1552]: Removed session 7. Sep 11 00:19:19.630689 sshd[1757]: Accepted publickey for core from 10.0.0.1 port 33874 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:19.632703 sshd-session[1757]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:19.638031 systemd-logind[1552]: New session 8 of user core. Sep 11 00:19:19.647942 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 11 00:19:19.705156 sudo[1762]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 11 00:19:19.705519 sudo[1762]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 00:19:19.713567 sudo[1762]: pam_unix(sudo:session): session closed for user root Sep 11 00:19:19.721695 sudo[1761]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 11 00:19:19.722075 sudo[1761]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 00:19:19.733620 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 11 00:19:19.791948 augenrules[1784]: No rules Sep 11 00:19:19.793705 systemd[1]: audit-rules.service: Deactivated successfully. Sep 11 00:19:19.794043 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 11 00:19:19.795455 sudo[1761]: pam_unix(sudo:session): session closed for user root Sep 11 00:19:19.797294 sshd[1760]: Connection closed by 10.0.0.1 port 33874 Sep 11 00:19:19.797747 sshd-session[1757]: pam_unix(sshd:session): session closed for user core Sep 11 00:19:19.815008 systemd[1]: sshd@7-10.0.0.82:22-10.0.0.1:33874.service: Deactivated successfully. Sep 11 00:19:19.817158 systemd[1]: session-8.scope: Deactivated successfully. Sep 11 00:19:19.818074 systemd-logind[1552]: Session 8 logged out. Waiting for processes to exit. Sep 11 00:19:19.821409 systemd[1]: Started sshd@8-10.0.0.82:22-10.0.0.1:33884.service - OpenSSH per-connection server daemon (10.0.0.1:33884). Sep 11 00:19:19.822088 systemd-logind[1552]: Removed session 8. Sep 11 00:19:19.885114 sshd[1793]: Accepted publickey for core from 10.0.0.1 port 33884 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:19:19.887649 sshd-session[1793]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:19:19.893662 systemd-logind[1552]: New session 9 of user core. Sep 11 00:19:19.903757 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 11 00:19:19.962302 sudo[1797]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 11 00:19:19.962777 sudo[1797]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 11 00:19:20.751152 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 11 00:19:20.753352 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:21.160338 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:21.325363 (kubelet)[1824]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 00:19:21.518928 kubelet[1824]: E0911 00:19:21.517674 1824 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 00:19:21.524996 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 00:19:21.525237 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 00:19:21.525718 systemd[1]: kubelet.service: Consumed 491ms CPU time, 109.1M memory peak. Sep 11 00:19:21.550724 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 11 00:19:21.569045 (dockerd)[1834]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 11 00:19:22.270942 dockerd[1834]: time="2025-09-11T00:19:22.270846004Z" level=info msg="Starting up" Sep 11 00:19:22.271954 dockerd[1834]: time="2025-09-11T00:19:22.271909729Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 11 00:19:22.287529 dockerd[1834]: time="2025-09-11T00:19:22.287466199Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 11 00:19:23.502359 dockerd[1834]: time="2025-09-11T00:19:23.502234758Z" level=info msg="Loading containers: start." Sep 11 00:19:23.722605 kernel: Initializing XFRM netlink socket Sep 11 00:19:24.146600 systemd-networkd[1474]: docker0: Link UP Sep 11 00:19:24.152589 dockerd[1834]: time="2025-09-11T00:19:24.152476956Z" level=info msg="Loading containers: done." Sep 11 00:19:24.244749 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1502111682-merged.mount: Deactivated successfully. Sep 11 00:19:24.399822 dockerd[1834]: time="2025-09-11T00:19:24.399632890Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 11 00:19:24.399822 dockerd[1834]: time="2025-09-11T00:19:24.399757975Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 11 00:19:24.400008 dockerd[1834]: time="2025-09-11T00:19:24.399893138Z" level=info msg="Initializing buildkit" Sep 11 00:19:24.701947 dockerd[1834]: time="2025-09-11T00:19:24.701748382Z" level=info msg="Completed buildkit initialization" Sep 11 00:19:24.709577 dockerd[1834]: time="2025-09-11T00:19:24.709475075Z" level=info msg="Daemon has completed initialization" Sep 11 00:19:24.709748 dockerd[1834]: time="2025-09-11T00:19:24.709597816Z" level=info msg="API listen on /run/docker.sock" Sep 11 00:19:24.709865 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 11 00:19:26.424745 containerd[1571]: time="2025-09-11T00:19:26.424692313Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\"" Sep 11 00:19:28.726899 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2685574236.mount: Deactivated successfully. Sep 11 00:19:30.762748 containerd[1571]: time="2025-09-11T00:19:30.762673906Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:30.955050 containerd[1571]: time="2025-09-11T00:19:30.954942677Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.5: active requests=0, bytes read=30114893" Sep 11 00:19:30.964927 containerd[1571]: time="2025-09-11T00:19:30.964774920Z" level=info msg="ImageCreate event name:\"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:30.970633 containerd[1571]: time="2025-09-11T00:19:30.970568557Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:30.971757 containerd[1571]: time="2025-09-11T00:19:30.971711411Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.5\" with image id \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\", size \"30111492\" in 4.546977049s" Sep 11 00:19:30.971813 containerd[1571]: time="2025-09-11T00:19:30.971758619Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\" returns image reference \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\"" Sep 11 00:19:30.972867 containerd[1571]: time="2025-09-11T00:19:30.972832544Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\"" Sep 11 00:19:31.751186 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 11 00:19:31.753152 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:32.028000 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:32.046180 (kubelet)[2121]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 00:19:32.237321 kubelet[2121]: E0911 00:19:32.237231 2121 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 00:19:32.243446 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 00:19:32.243749 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 00:19:32.244287 systemd[1]: kubelet.service: Consumed 412ms CPU time, 109.1M memory peak. Sep 11 00:19:36.110494 containerd[1571]: time="2025-09-11T00:19:36.110406364Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:36.111220 containerd[1571]: time="2025-09-11T00:19:36.111172851Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.5: active requests=0, bytes read=26020844" Sep 11 00:19:36.112499 containerd[1571]: time="2025-09-11T00:19:36.112436331Z" level=info msg="ImageCreate event name:\"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:36.115846 containerd[1571]: time="2025-09-11T00:19:36.115764153Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:36.116726 containerd[1571]: time="2025-09-11T00:19:36.116685962Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.5\" with image id \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\", size \"27681301\" in 5.143822009s" Sep 11 00:19:36.116726 containerd[1571]: time="2025-09-11T00:19:36.116720988Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\" returns image reference \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\"" Sep 11 00:19:36.117263 containerd[1571]: time="2025-09-11T00:19:36.117224503Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\"" Sep 11 00:19:38.247752 containerd[1571]: time="2025-09-11T00:19:38.247671983Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:38.248808 containerd[1571]: time="2025-09-11T00:19:38.248769509Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.5: active requests=0, bytes read=20155568" Sep 11 00:19:38.250073 containerd[1571]: time="2025-09-11T00:19:38.250032712Z" level=info msg="ImageCreate event name:\"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:38.254951 containerd[1571]: time="2025-09-11T00:19:38.254897165Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:38.255780 containerd[1571]: time="2025-09-11T00:19:38.255708121Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.5\" with image id \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\", size \"21816043\" in 2.138436299s" Sep 11 00:19:38.255780 containerd[1571]: time="2025-09-11T00:19:38.255772674Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\" returns image reference \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\"" Sep 11 00:19:38.257027 containerd[1571]: time="2025-09-11T00:19:38.257001823Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\"" Sep 11 00:19:39.308831 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1961337201.mount: Deactivated successfully. Sep 11 00:19:40.165291 containerd[1571]: time="2025-09-11T00:19:40.165207594Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:40.166356 containerd[1571]: time="2025-09-11T00:19:40.166306607Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.5: active requests=0, bytes read=31929469" Sep 11 00:19:40.167909 containerd[1571]: time="2025-09-11T00:19:40.167860540Z" level=info msg="ImageCreate event name:\"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:40.170438 containerd[1571]: time="2025-09-11T00:19:40.170374590Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:40.171137 containerd[1571]: time="2025-09-11T00:19:40.171071143Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.5\" with image id \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\", repo tag \"registry.k8s.io/kube-proxy:v1.33.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\", size \"31928488\" in 1.914003765s" Sep 11 00:19:40.171137 containerd[1571]: time="2025-09-11T00:19:40.171124816Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\" returns image reference \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\"" Sep 11 00:19:40.171683 containerd[1571]: time="2025-09-11T00:19:40.171650913Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 11 00:19:40.740516 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount318588428.mount: Deactivated successfully. Sep 11 00:19:42.088187 containerd[1571]: time="2025-09-11T00:19:42.088036088Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:42.088902 containerd[1571]: time="2025-09-11T00:19:42.088838661Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Sep 11 00:19:42.090525 containerd[1571]: time="2025-09-11T00:19:42.090466678Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:42.095041 containerd[1571]: time="2025-09-11T00:19:42.094964403Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:42.096751 containerd[1571]: time="2025-09-11T00:19:42.096705997Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.925027723s" Sep 11 00:19:42.096751 containerd[1571]: time="2025-09-11T00:19:42.096743790Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Sep 11 00:19:42.097604 containerd[1571]: time="2025-09-11T00:19:42.097477400Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 11 00:19:42.251311 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Sep 11 00:19:42.253282 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:42.483804 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:42.489085 (kubelet)[2205]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 11 00:19:42.535708 kubelet[2205]: E0911 00:19:42.535607 2205 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 11 00:19:42.540492 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 11 00:19:42.540831 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 11 00:19:42.541329 systemd[1]: kubelet.service: Consumed 251ms CPU time, 110.3M memory peak. Sep 11 00:19:42.885385 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount602393253.mount: Deactivated successfully. Sep 11 00:19:42.892627 containerd[1571]: time="2025-09-11T00:19:42.892573434Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 00:19:42.893664 containerd[1571]: time="2025-09-11T00:19:42.893633468Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 11 00:19:42.895049 containerd[1571]: time="2025-09-11T00:19:42.894997671Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 00:19:42.898804 containerd[1571]: time="2025-09-11T00:19:42.898723292Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 11 00:19:42.899689 containerd[1571]: time="2025-09-11T00:19:42.899573395Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 802.033475ms" Sep 11 00:19:42.899689 containerd[1571]: time="2025-09-11T00:19:42.899605526Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 11 00:19:42.900253 containerd[1571]: time="2025-09-11T00:19:42.900174332Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 11 00:19:43.348192 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3638816926.mount: Deactivated successfully. Sep 11 00:19:46.048389 containerd[1571]: time="2025-09-11T00:19:46.048311073Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:46.051252 containerd[1571]: time="2025-09-11T00:19:46.051198753Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58378433" Sep 11 00:19:46.051767 containerd[1571]: time="2025-09-11T00:19:46.051731405Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:46.055967 containerd[1571]: time="2025-09-11T00:19:46.055889492Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:19:46.057144 containerd[1571]: time="2025-09-11T00:19:46.057094242Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 3.156869903s" Sep 11 00:19:46.057144 containerd[1571]: time="2025-09-11T00:19:46.057123036Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Sep 11 00:19:50.712974 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:50.713201 systemd[1]: kubelet.service: Consumed 251ms CPU time, 110.3M memory peak. Sep 11 00:19:50.716460 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:50.750391 systemd[1]: Reload requested from client PID 2304 ('systemctl') (unit session-9.scope)... Sep 11 00:19:50.750418 systemd[1]: Reloading... Sep 11 00:19:50.856614 zram_generator::config[2350]: No configuration found. Sep 11 00:19:51.279053 systemd[1]: Reloading finished in 528 ms. Sep 11 00:19:51.357733 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 11 00:19:51.357868 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 11 00:19:51.358199 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:51.358247 systemd[1]: kubelet.service: Consumed 175ms CPU time, 98.2M memory peak. Sep 11 00:19:51.360112 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:51.659842 update_engine[1559]: I20250911 00:19:51.659608 1559 update_attempter.cc:509] Updating boot flags... Sep 11 00:19:52.062622 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:52.080148 (kubelet)[2411]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 11 00:19:52.254591 kubelet[2411]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 00:19:52.254591 kubelet[2411]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 11 00:19:52.254591 kubelet[2411]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 00:19:52.255321 kubelet[2411]: I0911 00:19:52.254645 2411 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 11 00:19:52.935295 kubelet[2411]: I0911 00:19:52.935206 2411 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 11 00:19:52.935295 kubelet[2411]: I0911 00:19:52.935265 2411 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 11 00:19:52.935773 kubelet[2411]: I0911 00:19:52.935743 2411 server.go:956] "Client rotation is on, will bootstrap in background" Sep 11 00:19:52.988300 kubelet[2411]: E0911 00:19:52.988231 2411 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.82:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 11 00:19:52.988697 kubelet[2411]: I0911 00:19:52.988674 2411 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 00:19:52.996622 kubelet[2411]: I0911 00:19:52.996585 2411 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 11 00:19:53.003807 kubelet[2411]: I0911 00:19:53.003771 2411 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 11 00:19:53.004171 kubelet[2411]: I0911 00:19:53.004126 2411 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 11 00:19:53.004593 kubelet[2411]: I0911 00:19:53.004167 2411 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 11 00:19:53.004815 kubelet[2411]: I0911 00:19:53.004627 2411 topology_manager.go:138] "Creating topology manager with none policy" Sep 11 00:19:53.004815 kubelet[2411]: I0911 00:19:53.004644 2411 container_manager_linux.go:303] "Creating device plugin manager" Sep 11 00:19:53.004885 kubelet[2411]: I0911 00:19:53.004855 2411 state_mem.go:36] "Initialized new in-memory state store" Sep 11 00:19:53.011358 kubelet[2411]: I0911 00:19:53.011295 2411 kubelet.go:480] "Attempting to sync node with API server" Sep 11 00:19:53.011358 kubelet[2411]: I0911 00:19:53.011364 2411 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 11 00:19:53.011498 kubelet[2411]: I0911 00:19:53.011418 2411 kubelet.go:386] "Adding apiserver pod source" Sep 11 00:19:53.011498 kubelet[2411]: I0911 00:19:53.011446 2411 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 11 00:19:53.014847 kubelet[2411]: E0911 00:19:53.014820 2411 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.82:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 11 00:19:53.014939 kubelet[2411]: E0911 00:19:53.014820 2411 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 11 00:19:53.056931 kubelet[2411]: I0911 00:19:53.056813 2411 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 11 00:19:53.057443 kubelet[2411]: I0911 00:19:53.057407 2411 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 11 00:19:53.058958 kubelet[2411]: W0911 00:19:53.058911 2411 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 11 00:19:53.062255 kubelet[2411]: I0911 00:19:53.062224 2411 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 11 00:19:53.062303 kubelet[2411]: I0911 00:19:53.062280 2411 server.go:1289] "Started kubelet" Sep 11 00:19:53.063292 kubelet[2411]: I0911 00:19:53.062685 2411 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 11 00:19:53.063957 kubelet[2411]: I0911 00:19:53.063927 2411 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 11 00:19:53.064074 kubelet[2411]: I0911 00:19:53.063924 2411 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 11 00:19:53.064532 kubelet[2411]: I0911 00:19:53.063930 2411 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 11 00:19:53.066505 kubelet[2411]: I0911 00:19:53.066457 2411 server.go:317] "Adding debug handlers to kubelet server" Sep 11 00:19:53.067424 kubelet[2411]: I0911 00:19:53.067303 2411 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 11 00:19:53.069333 kubelet[2411]: E0911 00:19:53.069283 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:53.069416 kubelet[2411]: I0911 00:19:53.069361 2411 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 11 00:19:53.069480 kubelet[2411]: E0911 00:19:53.069459 2411 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 11 00:19:53.070010 kubelet[2411]: I0911 00:19:53.069644 2411 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 11 00:19:53.070010 kubelet[2411]: I0911 00:19:53.069723 2411 reconciler.go:26] "Reconciler: start to sync state" Sep 11 00:19:53.070388 kubelet[2411]: E0911 00:19:53.070363 2411 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="200ms" Sep 11 00:19:53.070584 kubelet[2411]: E0911 00:19:53.070514 2411 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.82:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 11 00:19:53.070584 kubelet[2411]: I0911 00:19:53.070573 2411 factory.go:223] Registration of the systemd container factory successfully Sep 11 00:19:53.070820 kubelet[2411]: I0911 00:19:53.070654 2411 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 11 00:19:53.071811 kubelet[2411]: E0911 00:19:53.069774 2411 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.82:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.82:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18641268237727aa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-11 00:19:53.062246314 +0000 UTC m=+0.973826175,LastTimestamp:2025-09-11 00:19:53.062246314 +0000 UTC m=+0.973826175,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 11 00:19:53.073120 kubelet[2411]: I0911 00:19:53.073069 2411 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 11 00:19:53.075053 kubelet[2411]: I0911 00:19:53.073657 2411 factory.go:223] Registration of the containerd container factory successfully Sep 11 00:19:53.093092 kubelet[2411]: I0911 00:19:53.093051 2411 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 11 00:19:53.093293 kubelet[2411]: I0911 00:19:53.093277 2411 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 11 00:19:53.093407 kubelet[2411]: I0911 00:19:53.093371 2411 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 11 00:19:53.093407 kubelet[2411]: I0911 00:19:53.093394 2411 kubelet.go:2436] "Starting kubelet main sync loop" Sep 11 00:19:53.093598 kubelet[2411]: E0911 00:19:53.093451 2411 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 11 00:19:53.096838 kubelet[2411]: I0911 00:19:53.096497 2411 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 11 00:19:53.096838 kubelet[2411]: I0911 00:19:53.096517 2411 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 11 00:19:53.096838 kubelet[2411]: I0911 00:19:53.096570 2411 state_mem.go:36] "Initialized new in-memory state store" Sep 11 00:19:53.097121 kubelet[2411]: E0911 00:19:53.096945 2411 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 11 00:19:53.169570 kubelet[2411]: E0911 00:19:53.169495 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:53.194601 kubelet[2411]: E0911 00:19:53.194490 2411 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 11 00:19:53.269846 kubelet[2411]: E0911 00:19:53.269786 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:53.271569 kubelet[2411]: E0911 00:19:53.271504 2411 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="400ms" Sep 11 00:19:53.336883 kubelet[2411]: I0911 00:19:53.336804 2411 policy_none.go:49] "None policy: Start" Sep 11 00:19:53.336883 kubelet[2411]: I0911 00:19:53.336874 2411 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 11 00:19:53.336883 kubelet[2411]: I0911 00:19:53.336902 2411 state_mem.go:35] "Initializing new in-memory state store" Sep 11 00:19:53.344173 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 11 00:19:53.361510 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 11 00:19:53.365314 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 11 00:19:53.370630 kubelet[2411]: E0911 00:19:53.370579 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:53.378034 kubelet[2411]: E0911 00:19:53.377987 2411 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 11 00:19:53.378391 kubelet[2411]: I0911 00:19:53.378266 2411 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 11 00:19:53.378391 kubelet[2411]: I0911 00:19:53.378287 2411 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 11 00:19:53.378640 kubelet[2411]: I0911 00:19:53.378614 2411 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 11 00:19:53.380873 kubelet[2411]: E0911 00:19:53.380825 2411 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 11 00:19:53.381003 kubelet[2411]: E0911 00:19:53.380909 2411 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 11 00:19:53.407590 systemd[1]: Created slice kubepods-burstable-pod074b928878c1347f0b60bd8241148072.slice - libcontainer container kubepods-burstable-pod074b928878c1347f0b60bd8241148072.slice. Sep 11 00:19:53.418087 kubelet[2411]: E0911 00:19:53.418020 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:53.422669 systemd[1]: Created slice kubepods-burstable-podb678d5c6713e936e66aa5bb73166297e.slice - libcontainer container kubepods-burstable-podb678d5c6713e936e66aa5bb73166297e.slice. Sep 11 00:19:53.439428 kubelet[2411]: E0911 00:19:53.439370 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:53.441571 systemd[1]: Created slice kubepods-burstable-pod7b968cf906b2d9d713a362c43868bef2.slice - libcontainer container kubepods-burstable-pod7b968cf906b2d9d713a362c43868bef2.slice. Sep 11 00:19:53.443640 kubelet[2411]: E0911 00:19:53.443614 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:53.472291 kubelet[2411]: I0911 00:19:53.472083 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:53.472291 kubelet[2411]: I0911 00:19:53.472151 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:53.472291 kubelet[2411]: I0911 00:19:53.472181 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:53.472291 kubelet[2411]: I0911 00:19:53.472276 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7b968cf906b2d9d713a362c43868bef2-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7b968cf906b2d9d713a362c43868bef2\") " pod="kube-system/kube-scheduler-localhost" Sep 11 00:19:53.472539 kubelet[2411]: I0911 00:19:53.472308 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/074b928878c1347f0b60bd8241148072-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"074b928878c1347f0b60bd8241148072\") " pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:53.472539 kubelet[2411]: I0911 00:19:53.472328 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/074b928878c1347f0b60bd8241148072-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"074b928878c1347f0b60bd8241148072\") " pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:53.472539 kubelet[2411]: I0911 00:19:53.472348 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/074b928878c1347f0b60bd8241148072-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"074b928878c1347f0b60bd8241148072\") " pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:53.472539 kubelet[2411]: I0911 00:19:53.472366 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:53.472539 kubelet[2411]: I0911 00:19:53.472392 2411 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:53.480491 kubelet[2411]: I0911 00:19:53.480450 2411 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 00:19:53.480978 kubelet[2411]: E0911 00:19:53.480931 2411 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" Sep 11 00:19:53.672258 kubelet[2411]: E0911 00:19:53.672188 2411 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="800ms" Sep 11 00:19:53.682851 kubelet[2411]: I0911 00:19:53.682800 2411 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 00:19:53.683223 kubelet[2411]: E0911 00:19:53.683180 2411 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" Sep 11 00:19:53.719798 kubelet[2411]: E0911 00:19:53.719703 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:53.720636 containerd[1571]: time="2025-09-11T00:19:53.720577474Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:074b928878c1347f0b60bd8241148072,Namespace:kube-system,Attempt:0,}" Sep 11 00:19:53.741042 kubelet[2411]: E0911 00:19:53.740915 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:53.741596 containerd[1571]: time="2025-09-11T00:19:53.741443833Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b678d5c6713e936e66aa5bb73166297e,Namespace:kube-system,Attempt:0,}" Sep 11 00:19:53.744945 kubelet[2411]: E0911 00:19:53.744891 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:53.745512 containerd[1571]: time="2025-09-11T00:19:53.745471263Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7b968cf906b2d9d713a362c43868bef2,Namespace:kube-system,Attempt:0,}" Sep 11 00:19:53.754174 containerd[1571]: time="2025-09-11T00:19:53.754105929Z" level=info msg="connecting to shim 29594caeae105787dbd058cc3cfe42ece73874f0cfcc01be15ef05e86c4caf61" address="unix:///run/containerd/s/305b94a721ee6ea95a74cdcf97d1c5210149dfcf637731aaf18609531017d8d9" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:19:53.815732 containerd[1571]: time="2025-09-11T00:19:53.815637695Z" level=info msg="connecting to shim ca61c7e20f69842a436ebc21513451f00d195884200f6038345a7da4731a600d" address="unix:///run/containerd/s/fb54fe0b47a8509532f502877133c03c725c583f6985410bbccfe7fd783276bd" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:19:53.839059 containerd[1571]: time="2025-09-11T00:19:53.838959479Z" level=info msg="connecting to shim d80416612ba7677ddbbdd4d33ae6edc64cd98043755de5e35f484ab050e58991" address="unix:///run/containerd/s/50d2bb62ec449f1633058cf9a7433cd001b540f509e44db8360e8269b3363064" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:19:53.847780 systemd[1]: Started cri-containerd-29594caeae105787dbd058cc3cfe42ece73874f0cfcc01be15ef05e86c4caf61.scope - libcontainer container 29594caeae105787dbd058cc3cfe42ece73874f0cfcc01be15ef05e86c4caf61. Sep 11 00:19:53.854047 systemd[1]: Started cri-containerd-ca61c7e20f69842a436ebc21513451f00d195884200f6038345a7da4731a600d.scope - libcontainer container ca61c7e20f69842a436ebc21513451f00d195884200f6038345a7da4731a600d. Sep 11 00:19:53.935961 systemd[1]: Started cri-containerd-d80416612ba7677ddbbdd4d33ae6edc64cd98043755de5e35f484ab050e58991.scope - libcontainer container d80416612ba7677ddbbdd4d33ae6edc64cd98043755de5e35f484ab050e58991. Sep 11 00:19:54.001235 containerd[1571]: time="2025-09-11T00:19:54.001160267Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b678d5c6713e936e66aa5bb73166297e,Namespace:kube-system,Attempt:0,} returns sandbox id \"ca61c7e20f69842a436ebc21513451f00d195884200f6038345a7da4731a600d\"" Sep 11 00:19:54.001402 containerd[1571]: time="2025-09-11T00:19:54.001368270Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:074b928878c1347f0b60bd8241148072,Namespace:kube-system,Attempt:0,} returns sandbox id \"29594caeae105787dbd058cc3cfe42ece73874f0cfcc01be15ef05e86c4caf61\"" Sep 11 00:19:54.002818 kubelet[2411]: E0911 00:19:54.002783 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:54.002917 kubelet[2411]: E0911 00:19:54.002844 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:54.009587 kubelet[2411]: E0911 00:19:54.009488 2411 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 11 00:19:54.011944 containerd[1571]: time="2025-09-11T00:19:54.011911607Z" level=info msg="CreateContainer within sandbox \"29594caeae105787dbd058cc3cfe42ece73874f0cfcc01be15ef05e86c4caf61\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 11 00:19:54.013066 containerd[1571]: time="2025-09-11T00:19:54.013026575Z" level=info msg="CreateContainer within sandbox \"ca61c7e20f69842a436ebc21513451f00d195884200f6038345a7da4731a600d\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 11 00:19:54.022274 containerd[1571]: time="2025-09-11T00:19:54.022218357Z" level=info msg="Container 1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:19:54.028743 containerd[1571]: time="2025-09-11T00:19:54.028680586Z" level=info msg="Container 8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:19:54.036827 containerd[1571]: time="2025-09-11T00:19:54.036777317Z" level=info msg="CreateContainer within sandbox \"29594caeae105787dbd058cc3cfe42ece73874f0cfcc01be15ef05e86c4caf61\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756\"" Sep 11 00:19:54.038184 containerd[1571]: time="2025-09-11T00:19:54.037949283Z" level=info msg="StartContainer for \"1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756\"" Sep 11 00:19:54.040259 containerd[1571]: time="2025-09-11T00:19:54.039921062Z" level=info msg="connecting to shim 1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756" address="unix:///run/containerd/s/305b94a721ee6ea95a74cdcf97d1c5210149dfcf637731aaf18609531017d8d9" protocol=ttrpc version=3 Sep 11 00:19:54.043897 containerd[1571]: time="2025-09-11T00:19:54.043819674Z" level=info msg="CreateContainer within sandbox \"ca61c7e20f69842a436ebc21513451f00d195884200f6038345a7da4731a600d\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82\"" Sep 11 00:19:54.044300 containerd[1571]: time="2025-09-11T00:19:54.044272700Z" level=info msg="StartContainer for \"8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82\"" Sep 11 00:19:54.045385 containerd[1571]: time="2025-09-11T00:19:54.045352603Z" level=info msg="connecting to shim 8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82" address="unix:///run/containerd/s/fb54fe0b47a8509532f502877133c03c725c583f6985410bbccfe7fd783276bd" protocol=ttrpc version=3 Sep 11 00:19:54.056393 containerd[1571]: time="2025-09-11T00:19:54.056344207Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7b968cf906b2d9d713a362c43868bef2,Namespace:kube-system,Attempt:0,} returns sandbox id \"d80416612ba7677ddbbdd4d33ae6edc64cd98043755de5e35f484ab050e58991\"" Sep 11 00:19:54.058910 kubelet[2411]: E0911 00:19:54.058687 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:54.065862 systemd[1]: Started cri-containerd-1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756.scope - libcontainer container 1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756. Sep 11 00:19:54.066294 containerd[1571]: time="2025-09-11T00:19:54.066233808Z" level=info msg="CreateContainer within sandbox \"d80416612ba7677ddbbdd4d33ae6edc64cd98043755de5e35f484ab050e58991\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 11 00:19:54.080799 containerd[1571]: time="2025-09-11T00:19:54.080742493Z" level=info msg="Container 28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:19:54.085841 systemd[1]: Started cri-containerd-8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82.scope - libcontainer container 8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82. Sep 11 00:19:54.088918 kubelet[2411]: I0911 00:19:54.088747 2411 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 00:19:54.090382 kubelet[2411]: E0911 00:19:54.090354 2411 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" Sep 11 00:19:54.097328 containerd[1571]: time="2025-09-11T00:19:54.097258704Z" level=info msg="CreateContainer within sandbox \"d80416612ba7677ddbbdd4d33ae6edc64cd98043755de5e35f484ab050e58991\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035\"" Sep 11 00:19:54.098740 containerd[1571]: time="2025-09-11T00:19:54.098713585Z" level=info msg="StartContainer for \"28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035\"" Sep 11 00:19:54.100600 containerd[1571]: time="2025-09-11T00:19:54.100577320Z" level=info msg="connecting to shim 28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035" address="unix:///run/containerd/s/50d2bb62ec449f1633058cf9a7433cd001b540f509e44db8360e8269b3363064" protocol=ttrpc version=3 Sep 11 00:19:54.166594 systemd[1]: Started cri-containerd-28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035.scope - libcontainer container 28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035. Sep 11 00:19:54.184495 containerd[1571]: time="2025-09-11T00:19:54.184413139Z" level=info msg="StartContainer for \"1f329b48c4e48ae50a8e210d93656a86b2a49271c4fbc7dcba3d9f9084dbf756\" returns successfully" Sep 11 00:19:54.281598 containerd[1571]: time="2025-09-11T00:19:54.281450873Z" level=info msg="StartContainer for \"8782fcad480f408c4bca33645e0a8460dab4acaf54ac7cf7d798c486c7e50e82\" returns successfully" Sep 11 00:19:54.321372 kubelet[2411]: E0911 00:19:54.321307 2411 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.82:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 11 00:19:54.341583 containerd[1571]: time="2025-09-11T00:19:54.341374496Z" level=info msg="StartContainer for \"28e4a8a607e42af1647786b6d2208d0982bcb230f7d2b8b8aa9f851fd785f035\" returns successfully" Sep 11 00:19:55.001277 kubelet[2411]: I0911 00:19:55.001220 2411 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 00:19:55.225956 kubelet[2411]: E0911 00:19:55.225919 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:55.226111 kubelet[2411]: E0911 00:19:55.226095 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:55.229793 kubelet[2411]: E0911 00:19:55.229767 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:55.230051 kubelet[2411]: E0911 00:19:55.230033 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:55.235347 kubelet[2411]: E0911 00:19:55.235303 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:55.235451 kubelet[2411]: E0911 00:19:55.235429 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:55.882777 kubelet[2411]: E0911 00:19:55.882712 2411 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 11 00:19:56.106281 kubelet[2411]: I0911 00:19:56.106230 2411 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 11 00:19:56.106281 kubelet[2411]: E0911 00:19:56.106270 2411 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 11 00:19:56.131215 kubelet[2411]: E0911 00:19:56.131158 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.231385 kubelet[2411]: E0911 00:19:56.231223 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.233816 kubelet[2411]: E0911 00:19:56.233781 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:56.233875 kubelet[2411]: E0911 00:19:56.233832 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:56.233946 kubelet[2411]: E0911 00:19:56.233924 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:56.233946 kubelet[2411]: E0911 00:19:56.233934 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:56.233991 kubelet[2411]: E0911 00:19:56.233972 2411 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 11 00:19:56.234074 kubelet[2411]: E0911 00:19:56.234053 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:56.331752 kubelet[2411]: E0911 00:19:56.331689 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.432849 kubelet[2411]: E0911 00:19:56.432764 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.533923 kubelet[2411]: E0911 00:19:56.533848 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.634839 kubelet[2411]: E0911 00:19:56.634777 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.735630 kubelet[2411]: E0911 00:19:56.735540 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.836450 kubelet[2411]: E0911 00:19:56.836290 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:56.937199 kubelet[2411]: E0911 00:19:56.937133 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:57.037461 kubelet[2411]: E0911 00:19:57.037389 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:57.138322 kubelet[2411]: E0911 00:19:57.137742 2411 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 11 00:19:57.170766 kubelet[2411]: I0911 00:19:57.170645 2411 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:57.179356 kubelet[2411]: I0911 00:19:57.179319 2411 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:57.183044 kubelet[2411]: I0911 00:19:57.182997 2411 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 00:19:57.234464 kubelet[2411]: I0911 00:19:57.234420 2411 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 00:19:57.234668 kubelet[2411]: I0911 00:19:57.234623 2411 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:57.239844 kubelet[2411]: E0911 00:19:57.239800 2411 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 11 00:19:57.239994 kubelet[2411]: E0911 00:19:57.239959 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:57.240227 kubelet[2411]: E0911 00:19:57.240200 2411 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:57.240331 kubelet[2411]: E0911 00:19:57.240313 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:58.017367 kubelet[2411]: I0911 00:19:58.017292 2411 apiserver.go:52] "Watching apiserver" Sep 11 00:19:58.020015 kubelet[2411]: E0911 00:19:58.019983 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:58.070287 kubelet[2411]: I0911 00:19:58.070237 2411 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 11 00:19:58.236316 kubelet[2411]: E0911 00:19:58.236269 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:58.236599 kubelet[2411]: E0911 00:19:58.236541 2411 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:19:58.915227 systemd[1]: Reload requested from client PID 2695 ('systemctl') (unit session-9.scope)... Sep 11 00:19:58.915245 systemd[1]: Reloading... Sep 11 00:19:58.995682 zram_generator::config[2741]: No configuration found. Sep 11 00:19:59.243279 systemd[1]: Reloading finished in 327 ms. Sep 11 00:19:59.283857 kubelet[2411]: I0911 00:19:59.283758 2411 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 00:19:59.283874 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:59.307171 systemd[1]: kubelet.service: Deactivated successfully. Sep 11 00:19:59.307517 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:59.307593 systemd[1]: kubelet.service: Consumed 1.592s CPU time, 132.9M memory peak. Sep 11 00:19:59.309669 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 11 00:19:59.587378 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 11 00:19:59.601109 (kubelet)[2783]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 11 00:19:59.656622 kubelet[2783]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 00:19:59.656622 kubelet[2783]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 11 00:19:59.656622 kubelet[2783]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 11 00:19:59.657198 kubelet[2783]: I0911 00:19:59.656622 2783 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 11 00:19:59.667519 kubelet[2783]: I0911 00:19:59.667443 2783 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 11 00:19:59.667519 kubelet[2783]: I0911 00:19:59.667486 2783 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 11 00:19:59.667813 kubelet[2783]: I0911 00:19:59.667781 2783 server.go:956] "Client rotation is on, will bootstrap in background" Sep 11 00:19:59.669427 kubelet[2783]: I0911 00:19:59.669392 2783 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 11 00:19:59.676914 kubelet[2783]: I0911 00:19:59.676823 2783 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 11 00:19:59.681422 kubelet[2783]: I0911 00:19:59.681391 2783 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 11 00:19:59.688126 kubelet[2783]: I0911 00:19:59.688068 2783 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 11 00:19:59.688343 kubelet[2783]: I0911 00:19:59.688304 2783 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 11 00:19:59.688508 kubelet[2783]: I0911 00:19:59.688335 2783 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 11 00:19:59.688682 kubelet[2783]: I0911 00:19:59.688516 2783 topology_manager.go:138] "Creating topology manager with none policy" Sep 11 00:19:59.688682 kubelet[2783]: I0911 00:19:59.688528 2783 container_manager_linux.go:303] "Creating device plugin manager" Sep 11 00:19:59.688682 kubelet[2783]: I0911 00:19:59.688611 2783 state_mem.go:36] "Initialized new in-memory state store" Sep 11 00:19:59.688834 kubelet[2783]: I0911 00:19:59.688808 2783 kubelet.go:480] "Attempting to sync node with API server" Sep 11 00:19:59.688834 kubelet[2783]: I0911 00:19:59.688828 2783 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 11 00:19:59.688905 kubelet[2783]: I0911 00:19:59.688855 2783 kubelet.go:386] "Adding apiserver pod source" Sep 11 00:19:59.688905 kubelet[2783]: I0911 00:19:59.688871 2783 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 11 00:19:59.691674 kubelet[2783]: I0911 00:19:59.690772 2783 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 11 00:19:59.691674 kubelet[2783]: I0911 00:19:59.691354 2783 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 11 00:19:59.694935 kubelet[2783]: I0911 00:19:59.694890 2783 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 11 00:19:59.695120 kubelet[2783]: I0911 00:19:59.695108 2783 server.go:1289] "Started kubelet" Sep 11 00:19:59.696735 kubelet[2783]: I0911 00:19:59.696516 2783 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 11 00:19:59.696921 kubelet[2783]: I0911 00:19:59.696861 2783 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 11 00:19:59.697512 kubelet[2783]: I0911 00:19:59.697492 2783 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 11 00:19:59.698434 kubelet[2783]: I0911 00:19:59.698396 2783 server.go:317] "Adding debug handlers to kubelet server" Sep 11 00:19:59.701874 kubelet[2783]: I0911 00:19:59.701788 2783 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 11 00:19:59.705287 kubelet[2783]: I0911 00:19:59.705056 2783 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 11 00:19:59.707587 kubelet[2783]: I0911 00:19:59.707372 2783 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 11 00:19:59.707587 kubelet[2783]: I0911 00:19:59.707486 2783 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 11 00:19:59.707760 kubelet[2783]: I0911 00:19:59.707629 2783 reconciler.go:26] "Reconciler: start to sync state" Sep 11 00:19:59.709047 kubelet[2783]: E0911 00:19:59.708989 2783 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 11 00:19:59.709507 kubelet[2783]: I0911 00:19:59.709479 2783 factory.go:223] Registration of the systemd container factory successfully Sep 11 00:19:59.709651 kubelet[2783]: I0911 00:19:59.709616 2783 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 11 00:19:59.712811 kubelet[2783]: I0911 00:19:59.712777 2783 factory.go:223] Registration of the containerd container factory successfully Sep 11 00:19:59.767198 kubelet[2783]: I0911 00:19:59.766184 2783 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 11 00:19:59.770026 kubelet[2783]: I0911 00:19:59.769748 2783 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 11 00:19:59.770026 kubelet[2783]: I0911 00:19:59.769781 2783 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 11 00:19:59.770026 kubelet[2783]: I0911 00:19:59.769815 2783 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 11 00:19:59.770026 kubelet[2783]: I0911 00:19:59.769825 2783 kubelet.go:2436] "Starting kubelet main sync loop" Sep 11 00:19:59.770026 kubelet[2783]: E0911 00:19:59.769880 2783 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 11 00:19:59.821694 kubelet[2783]: I0911 00:19:59.821631 2783 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 11 00:19:59.821694 kubelet[2783]: I0911 00:19:59.821671 2783 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 11 00:19:59.821694 kubelet[2783]: I0911 00:19:59.821708 2783 state_mem.go:36] "Initialized new in-memory state store" Sep 11 00:19:59.821968 kubelet[2783]: I0911 00:19:59.821947 2783 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 11 00:19:59.821999 kubelet[2783]: I0911 00:19:59.821967 2783 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 11 00:19:59.821999 kubelet[2783]: I0911 00:19:59.821989 2783 policy_none.go:49] "None policy: Start" Sep 11 00:19:59.822045 kubelet[2783]: I0911 00:19:59.822001 2783 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 11 00:19:59.822045 kubelet[2783]: I0911 00:19:59.822019 2783 state_mem.go:35] "Initializing new in-memory state store" Sep 11 00:19:59.822176 kubelet[2783]: I0911 00:19:59.822159 2783 state_mem.go:75] "Updated machine memory state" Sep 11 00:19:59.829524 kubelet[2783]: E0911 00:19:59.829475 2783 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 11 00:19:59.830907 kubelet[2783]: I0911 00:19:59.830878 2783 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 11 00:19:59.831001 kubelet[2783]: I0911 00:19:59.830900 2783 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 11 00:19:59.831736 kubelet[2783]: I0911 00:19:59.831270 2783 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 11 00:19:59.835881 kubelet[2783]: E0911 00:19:59.835825 2783 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 11 00:19:59.871691 kubelet[2783]: I0911 00:19:59.871471 2783 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 00:19:59.871691 kubelet[2783]: I0911 00:19:59.871575 2783 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 00:19:59.871870 kubelet[2783]: I0911 00:19:59.871575 2783 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 11 00:19:59.941914 kubelet[2783]: I0911 00:19:59.941863 2783 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 11 00:20:00.005384 kubelet[2783]: E0911 00:20:00.005266 2783 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 11 00:20:00.009566 kubelet[2783]: I0911 00:20:00.009523 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.009566 kubelet[2783]: I0911 00:20:00.009583 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.009801 kubelet[2783]: I0911 00:20:00.009608 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/074b928878c1347f0b60bd8241148072-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"074b928878c1347f0b60bd8241148072\") " pod="kube-system/kube-apiserver-localhost" Sep 11 00:20:00.009801 kubelet[2783]: I0911 00:20:00.009629 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.009801 kubelet[2783]: I0911 00:20:00.009690 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.009801 kubelet[2783]: I0911 00:20:00.009730 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7b968cf906b2d9d713a362c43868bef2-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7b968cf906b2d9d713a362c43868bef2\") " pod="kube-system/kube-scheduler-localhost" Sep 11 00:20:00.009961 kubelet[2783]: I0911 00:20:00.009816 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/074b928878c1347f0b60bd8241148072-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"074b928878c1347f0b60bd8241148072\") " pod="kube-system/kube-apiserver-localhost" Sep 11 00:20:00.009961 kubelet[2783]: I0911 00:20:00.009878 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/074b928878c1347f0b60bd8241148072-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"074b928878c1347f0b60bd8241148072\") " pod="kube-system/kube-apiserver-localhost" Sep 11 00:20:00.009961 kubelet[2783]: I0911 00:20:00.009906 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b678d5c6713e936e66aa5bb73166297e-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b678d5c6713e936e66aa5bb73166297e\") " pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.063855 kubelet[2783]: E0911 00:20:00.063766 2783 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.064613 kubelet[2783]: E0911 00:20:00.064570 2783 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 11 00:20:00.258677 kubelet[2783]: I0911 00:20:00.258602 2783 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 11 00:20:00.258832 kubelet[2783]: I0911 00:20:00.258734 2783 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 11 00:20:00.307132 kubelet[2783]: E0911 00:20:00.306451 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:00.366127 kubelet[2783]: E0911 00:20:00.364993 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:00.366127 kubelet[2783]: E0911 00:20:00.365421 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:00.691073 kubelet[2783]: I0911 00:20:00.690906 2783 apiserver.go:52] "Watching apiserver" Sep 11 00:20:00.708222 kubelet[2783]: I0911 00:20:00.708151 2783 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 11 00:20:00.794250 kubelet[2783]: I0911 00:20:00.793814 2783 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.794250 kubelet[2783]: E0911 00:20:00.793831 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:00.794567 kubelet[2783]: I0911 00:20:00.794529 2783 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 11 00:20:00.858750 kubelet[2783]: E0911 00:20:00.811370 2783 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 11 00:20:00.858750 kubelet[2783]: E0911 00:20:00.811822 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:00.860982 kubelet[2783]: E0911 00:20:00.860928 2783 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 11 00:20:00.861284 kubelet[2783]: E0911 00:20:00.861176 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:00.875981 kubelet[2783]: I0911 00:20:00.875757 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.875682964 podStartE2EDuration="3.875682964s" podCreationTimestamp="2025-09-11 00:19:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:20:00.875273611 +0000 UTC m=+1.269078060" watchObservedRunningTime="2025-09-11 00:20:00.875682964 +0000 UTC m=+1.269487393" Sep 11 00:20:00.905646 kubelet[2783]: I0911 00:20:00.902541 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.9025198899999998 podStartE2EDuration="3.90251989s" podCreationTimestamp="2025-09-11 00:19:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:20:00.885540997 +0000 UTC m=+1.279345426" watchObservedRunningTime="2025-09-11 00:20:00.90251989 +0000 UTC m=+1.296324319" Sep 11 00:20:00.924750 kubelet[2783]: I0911 00:20:00.924686 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=3.9246511 podStartE2EDuration="3.9246511s" podCreationTimestamp="2025-09-11 00:19:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:20:00.908614264 +0000 UTC m=+1.302418723" watchObservedRunningTime="2025-09-11 00:20:00.9246511 +0000 UTC m=+1.318455529" Sep 11 00:20:01.796114 kubelet[2783]: E0911 00:20:01.796043 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:01.796114 kubelet[2783]: E0911 00:20:01.796090 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:01.796772 kubelet[2783]: E0911 00:20:01.796323 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:03.577494 kubelet[2783]: E0911 00:20:03.577427 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:04.314954 kubelet[2783]: I0911 00:20:04.314912 2783 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 11 00:20:04.315378 containerd[1571]: time="2025-09-11T00:20:04.315318588Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 11 00:20:04.315932 kubelet[2783]: I0911 00:20:04.315538 2783 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 11 00:20:05.004616 systemd[1]: Created slice kubepods-besteffort-podcd906daf_f478_4b07_8bad_1de384352913.slice - libcontainer container kubepods-besteffort-podcd906daf_f478_4b07_8bad_1de384352913.slice. Sep 11 00:20:05.040083 kubelet[2783]: I0911 00:20:05.040028 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/cd906daf-f478-4b07-8bad-1de384352913-kube-proxy\") pod \"kube-proxy-lxm96\" (UID: \"cd906daf-f478-4b07-8bad-1de384352913\") " pod="kube-system/kube-proxy-lxm96" Sep 11 00:20:05.040083 kubelet[2783]: I0911 00:20:05.040069 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z2fb\" (UniqueName: \"kubernetes.io/projected/cd906daf-f478-4b07-8bad-1de384352913-kube-api-access-5z2fb\") pod \"kube-proxy-lxm96\" (UID: \"cd906daf-f478-4b07-8bad-1de384352913\") " pod="kube-system/kube-proxy-lxm96" Sep 11 00:20:05.040083 kubelet[2783]: I0911 00:20:05.040089 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cd906daf-f478-4b07-8bad-1de384352913-xtables-lock\") pod \"kube-proxy-lxm96\" (UID: \"cd906daf-f478-4b07-8bad-1de384352913\") " pod="kube-system/kube-proxy-lxm96" Sep 11 00:20:05.041022 kubelet[2783]: I0911 00:20:05.040103 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cd906daf-f478-4b07-8bad-1de384352913-lib-modules\") pod \"kube-proxy-lxm96\" (UID: \"cd906daf-f478-4b07-8bad-1de384352913\") " pod="kube-system/kube-proxy-lxm96" Sep 11 00:20:05.050209 systemd[1]: Created slice kubepods-besteffort-poddcbc2fce_c62d_4122_943a_d9eea9779c90.slice - libcontainer container kubepods-besteffort-poddcbc2fce_c62d_4122_943a_d9eea9779c90.slice. Sep 11 00:20:05.141158 kubelet[2783]: I0911 00:20:05.141034 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6w6c\" (UniqueName: \"kubernetes.io/projected/dcbc2fce-c62d-4122-943a-d9eea9779c90-kube-api-access-g6w6c\") pod \"tigera-operator-755d956888-klmrt\" (UID: \"dcbc2fce-c62d-4122-943a-d9eea9779c90\") " pod="tigera-operator/tigera-operator-755d956888-klmrt" Sep 11 00:20:05.141158 kubelet[2783]: I0911 00:20:05.141100 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/dcbc2fce-c62d-4122-943a-d9eea9779c90-var-lib-calico\") pod \"tigera-operator-755d956888-klmrt\" (UID: \"dcbc2fce-c62d-4122-943a-d9eea9779c90\") " pod="tigera-operator/tigera-operator-755d956888-klmrt" Sep 11 00:20:05.313683 kubelet[2783]: E0911 00:20:05.313496 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:05.314698 containerd[1571]: time="2025-09-11T00:20:05.314627284Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-lxm96,Uid:cd906daf-f478-4b07-8bad-1de384352913,Namespace:kube-system,Attempt:0,}" Sep 11 00:20:05.354235 containerd[1571]: time="2025-09-11T00:20:05.354176754Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-klmrt,Uid:dcbc2fce-c62d-4122-943a-d9eea9779c90,Namespace:tigera-operator,Attempt:0,}" Sep 11 00:20:05.399335 containerd[1571]: time="2025-09-11T00:20:05.399271948Z" level=info msg="connecting to shim 6daef41e0a78eee42b2ca11005ecaac760dee81110b8a6853ecf2bd4d5de8b78" address="unix:///run/containerd/s/f7cb46e20a593b3aa16047f37e2521f8484d1abb83e839ce58039df4b9f17c74" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:05.418315 containerd[1571]: time="2025-09-11T00:20:05.418259645Z" level=info msg="connecting to shim e391ba6c5cc8e37dc8b70c47d174f0e7da13abcd37e34bc6e806dd63d7347d70" address="unix:///run/containerd/s/5431e4417576f9c946bf2f2ddc3cf80c17e5a380ed3508d6eaa316a9a6d0c115" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:05.438773 systemd[1]: Started cri-containerd-6daef41e0a78eee42b2ca11005ecaac760dee81110b8a6853ecf2bd4d5de8b78.scope - libcontainer container 6daef41e0a78eee42b2ca11005ecaac760dee81110b8a6853ecf2bd4d5de8b78. Sep 11 00:20:05.449767 systemd[1]: Started cri-containerd-e391ba6c5cc8e37dc8b70c47d174f0e7da13abcd37e34bc6e806dd63d7347d70.scope - libcontainer container e391ba6c5cc8e37dc8b70c47d174f0e7da13abcd37e34bc6e806dd63d7347d70. Sep 11 00:20:05.458585 kubelet[2783]: E0911 00:20:05.457931 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:05.480289 containerd[1571]: time="2025-09-11T00:20:05.480249063Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-lxm96,Uid:cd906daf-f478-4b07-8bad-1de384352913,Namespace:kube-system,Attempt:0,} returns sandbox id \"6daef41e0a78eee42b2ca11005ecaac760dee81110b8a6853ecf2bd4d5de8b78\"" Sep 11 00:20:05.481202 kubelet[2783]: E0911 00:20:05.481171 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:05.488349 containerd[1571]: time="2025-09-11T00:20:05.488296359Z" level=info msg="CreateContainer within sandbox \"6daef41e0a78eee42b2ca11005ecaac760dee81110b8a6853ecf2bd4d5de8b78\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 11 00:20:05.503214 containerd[1571]: time="2025-09-11T00:20:05.503113733Z" level=info msg="Container 9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:05.511113 containerd[1571]: time="2025-09-11T00:20:05.511066229Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-klmrt,Uid:dcbc2fce-c62d-4122-943a-d9eea9779c90,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"e391ba6c5cc8e37dc8b70c47d174f0e7da13abcd37e34bc6e806dd63d7347d70\"" Sep 11 00:20:05.512661 containerd[1571]: time="2025-09-11T00:20:05.512633730Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 11 00:20:05.518815 containerd[1571]: time="2025-09-11T00:20:05.518755259Z" level=info msg="CreateContainer within sandbox \"6daef41e0a78eee42b2ca11005ecaac760dee81110b8a6853ecf2bd4d5de8b78\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788\"" Sep 11 00:20:05.519284 containerd[1571]: time="2025-09-11T00:20:05.519255531Z" level=info msg="StartContainer for \"9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788\"" Sep 11 00:20:05.520729 containerd[1571]: time="2025-09-11T00:20:05.520699270Z" level=info msg="connecting to shim 9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788" address="unix:///run/containerd/s/f7cb46e20a593b3aa16047f37e2521f8484d1abb83e839ce58039df4b9f17c74" protocol=ttrpc version=3 Sep 11 00:20:05.541733 systemd[1]: Started cri-containerd-9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788.scope - libcontainer container 9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788. Sep 11 00:20:05.599407 containerd[1571]: time="2025-09-11T00:20:05.599055801Z" level=info msg="StartContainer for \"9cedd04ac45300d43a61d413662b43e307a44e0c02be94e815f99bb672f8f788\" returns successfully" Sep 11 00:20:05.807953 kubelet[2783]: E0911 00:20:05.807909 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:05.811692 kubelet[2783]: E0911 00:20:05.811635 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:08.119305 kubelet[2783]: E0911 00:20:08.119253 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:08.133999 kubelet[2783]: I0911 00:20:08.133922 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-lxm96" podStartSLOduration=4.133902708 podStartE2EDuration="4.133902708s" podCreationTimestamp="2025-09-11 00:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:20:05.831802471 +0000 UTC m=+6.225606900" watchObservedRunningTime="2025-09-11 00:20:08.133902708 +0000 UTC m=+8.527707137" Sep 11 00:20:08.812118 kubelet[2783]: E0911 00:20:08.812044 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:09.823478 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2515442518.mount: Deactivated successfully. Sep 11 00:20:10.236097 containerd[1571]: time="2025-09-11T00:20:10.235953027Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:10.236714 containerd[1571]: time="2025-09-11T00:20:10.236663714Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Sep 11 00:20:10.237933 containerd[1571]: time="2025-09-11T00:20:10.237889359Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:10.239945 containerd[1571]: time="2025-09-11T00:20:10.239903869Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:10.240707 containerd[1571]: time="2025-09-11T00:20:10.240666363Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 4.727934167s" Sep 11 00:20:10.240707 containerd[1571]: time="2025-09-11T00:20:10.240702291Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Sep 11 00:20:10.247019 containerd[1571]: time="2025-09-11T00:20:10.246981984Z" level=info msg="CreateContainer within sandbox \"e391ba6c5cc8e37dc8b70c47d174f0e7da13abcd37e34bc6e806dd63d7347d70\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 11 00:20:10.256051 containerd[1571]: time="2025-09-11T00:20:10.256018533Z" level=info msg="Container 2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:10.451568 containerd[1571]: time="2025-09-11T00:20:10.451427645Z" level=info msg="CreateContainer within sandbox \"e391ba6c5cc8e37dc8b70c47d174f0e7da13abcd37e34bc6e806dd63d7347d70\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1\"" Sep 11 00:20:10.453192 containerd[1571]: time="2025-09-11T00:20:10.453136900Z" level=info msg="StartContainer for \"2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1\"" Sep 11 00:20:10.454672 containerd[1571]: time="2025-09-11T00:20:10.454606144Z" level=info msg="connecting to shim 2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1" address="unix:///run/containerd/s/5431e4417576f9c946bf2f2ddc3cf80c17e5a380ed3508d6eaa316a9a6d0c115" protocol=ttrpc version=3 Sep 11 00:20:10.508689 systemd[1]: Started cri-containerd-2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1.scope - libcontainer container 2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1. Sep 11 00:20:10.594176 containerd[1571]: time="2025-09-11T00:20:10.594115721Z" level=info msg="StartContainer for \"2c0af4240b9e6d5959fd4a08347825ceea59d4303fa9318fca9048df5c9a76e1\" returns successfully" Sep 11 00:20:10.829814 kubelet[2783]: I0911 00:20:10.829618 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-klmrt" podStartSLOduration=2.100423304 podStartE2EDuration="6.829598367s" podCreationTimestamp="2025-09-11 00:20:04 +0000 UTC" firstStartedPulling="2025-09-11 00:20:05.512280155 +0000 UTC m=+5.906084574" lastFinishedPulling="2025-09-11 00:20:10.241455207 +0000 UTC m=+10.635259637" observedRunningTime="2025-09-11 00:20:10.829170612 +0000 UTC m=+11.222975041" watchObservedRunningTime="2025-09-11 00:20:10.829598367 +0000 UTC m=+11.223402796" Sep 11 00:20:13.583163 kubelet[2783]: E0911 00:20:13.583087 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:13.822872 kubelet[2783]: E0911 00:20:13.822819 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:16.441103 sudo[1797]: pam_unix(sudo:session): session closed for user root Sep 11 00:20:16.442724 sshd[1796]: Connection closed by 10.0.0.1 port 33884 Sep 11 00:20:16.443866 sshd-session[1793]: pam_unix(sshd:session): session closed for user core Sep 11 00:20:16.447932 systemd[1]: sshd@8-10.0.0.82:22-10.0.0.1:33884.service: Deactivated successfully. Sep 11 00:20:16.450179 systemd[1]: session-9.scope: Deactivated successfully. Sep 11 00:20:16.450437 systemd[1]: session-9.scope: Consumed 8.358s CPU time, 227.8M memory peak. Sep 11 00:20:16.453428 systemd-logind[1552]: Session 9 logged out. Waiting for processes to exit. Sep 11 00:20:16.455755 systemd-logind[1552]: Removed session 9. Sep 11 00:20:19.295614 systemd[1]: Created slice kubepods-besteffort-pod4786e90c_1b6c_48a7_abd2_3fc29de9399d.slice - libcontainer container kubepods-besteffort-pod4786e90c_1b6c_48a7_abd2_3fc29de9399d.slice. Sep 11 00:20:19.333888 kubelet[2783]: I0911 00:20:19.333827 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/4786e90c-1b6c-48a7-abd2-3fc29de9399d-typha-certs\") pod \"calico-typha-65dd79988f-7kwv5\" (UID: \"4786e90c-1b6c-48a7-abd2-3fc29de9399d\") " pod="calico-system/calico-typha-65dd79988f-7kwv5" Sep 11 00:20:19.333888 kubelet[2783]: I0911 00:20:19.333882 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dchxc\" (UniqueName: \"kubernetes.io/projected/4786e90c-1b6c-48a7-abd2-3fc29de9399d-kube-api-access-dchxc\") pod \"calico-typha-65dd79988f-7kwv5\" (UID: \"4786e90c-1b6c-48a7-abd2-3fc29de9399d\") " pod="calico-system/calico-typha-65dd79988f-7kwv5" Sep 11 00:20:19.334355 kubelet[2783]: I0911 00:20:19.333911 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4786e90c-1b6c-48a7-abd2-3fc29de9399d-tigera-ca-bundle\") pod \"calico-typha-65dd79988f-7kwv5\" (UID: \"4786e90c-1b6c-48a7-abd2-3fc29de9399d\") " pod="calico-system/calico-typha-65dd79988f-7kwv5" Sep 11 00:20:19.601521 kubelet[2783]: E0911 00:20:19.601368 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:19.602723 containerd[1571]: time="2025-09-11T00:20:19.602527618Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-65dd79988f-7kwv5,Uid:4786e90c-1b6c-48a7-abd2-3fc29de9399d,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:19.618842 systemd[1]: Created slice kubepods-besteffort-pod31c180a7_51e1_4d4b_8de0_4257d3b41e99.slice - libcontainer container kubepods-besteffort-pod31c180a7_51e1_4d4b_8de0_4257d3b41e99.slice. Sep 11 00:20:19.636949 kubelet[2783]: I0911 00:20:19.636886 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-lib-modules\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637089 kubelet[2783]: I0911 00:20:19.637008 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-policysync\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637089 kubelet[2783]: I0911 00:20:19.637037 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-var-lib-calico\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637089 kubelet[2783]: I0911 00:20:19.637087 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-cni-net-dir\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637177 kubelet[2783]: I0911 00:20:19.637102 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-flexvol-driver-host\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637177 kubelet[2783]: I0911 00:20:19.637133 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs7kh\" (UniqueName: \"kubernetes.io/projected/31c180a7-51e1-4d4b-8de0-4257d3b41e99-kube-api-access-gs7kh\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637177 kubelet[2783]: I0911 00:20:19.637148 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-xtables-lock\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637177 kubelet[2783]: I0911 00:20:19.637163 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-cni-bin-dir\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637177 kubelet[2783]: I0911 00:20:19.637176 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-cni-log-dir\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637314 kubelet[2783]: I0911 00:20:19.637192 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/31c180a7-51e1-4d4b-8de0-4257d3b41e99-node-certs\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637314 kubelet[2783]: I0911 00:20:19.637208 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31c180a7-51e1-4d4b-8de0-4257d3b41e99-tigera-ca-bundle\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.637314 kubelet[2783]: I0911 00:20:19.637248 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/31c180a7-51e1-4d4b-8de0-4257d3b41e99-var-run-calico\") pod \"calico-node-75tzc\" (UID: \"31c180a7-51e1-4d4b-8de0-4257d3b41e99\") " pod="calico-system/calico-node-75tzc" Sep 11 00:20:19.654654 containerd[1571]: time="2025-09-11T00:20:19.654579249Z" level=info msg="connecting to shim eb9349bf3627e6ff4dbb8a9c16fe035f444142abc5ca679a5a2ee83b3075d8b9" address="unix:///run/containerd/s/df37dc01e0c12a26d76525595d52650901339b0e87f8af79161c5aa7764d30a3" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:19.689761 systemd[1]: Started cri-containerd-eb9349bf3627e6ff4dbb8a9c16fe035f444142abc5ca679a5a2ee83b3075d8b9.scope - libcontainer container eb9349bf3627e6ff4dbb8a9c16fe035f444142abc5ca679a5a2ee83b3075d8b9. Sep 11 00:20:19.746814 kubelet[2783]: E0911 00:20:19.745275 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.746814 kubelet[2783]: W0911 00:20:19.745300 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.746814 kubelet[2783]: E0911 00:20:19.746629 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.749012 kubelet[2783]: E0911 00:20:19.748928 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.749012 kubelet[2783]: W0911 00:20:19.748948 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.749012 kubelet[2783]: E0911 00:20:19.748971 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.750396 kubelet[2783]: E0911 00:20:19.750370 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.750396 kubelet[2783]: W0911 00:20:19.750393 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.750905 kubelet[2783]: E0911 00:20:19.750409 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.778353 containerd[1571]: time="2025-09-11T00:20:19.778303397Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-65dd79988f-7kwv5,Uid:4786e90c-1b6c-48a7-abd2-3fc29de9399d,Namespace:calico-system,Attempt:0,} returns sandbox id \"eb9349bf3627e6ff4dbb8a9c16fe035f444142abc5ca679a5a2ee83b3075d8b9\"" Sep 11 00:20:19.778870 kubelet[2783]: E0911 00:20:19.778844 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:19.780035 containerd[1571]: time="2025-09-11T00:20:19.779971090Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 11 00:20:19.902899 kubelet[2783]: E0911 00:20:19.902083 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:19.920476 kubelet[2783]: E0911 00:20:19.920203 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.920476 kubelet[2783]: W0911 00:20:19.920257 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.920476 kubelet[2783]: E0911 00:20:19.920288 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.920729 kubelet[2783]: E0911 00:20:19.920673 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.920729 kubelet[2783]: W0911 00:20:19.920717 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.920814 kubelet[2783]: E0911 00:20:19.920741 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.920990 kubelet[2783]: E0911 00:20:19.920969 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.921027 kubelet[2783]: W0911 00:20:19.921001 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.921027 kubelet[2783]: E0911 00:20:19.921012 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.921293 kubelet[2783]: E0911 00:20:19.921268 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.921293 kubelet[2783]: W0911 00:20:19.921278 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.921293 kubelet[2783]: E0911 00:20:19.921287 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.921588 kubelet[2783]: E0911 00:20:19.921571 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.921588 kubelet[2783]: W0911 00:20:19.921583 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.921655 kubelet[2783]: E0911 00:20:19.921593 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.921817 kubelet[2783]: E0911 00:20:19.921777 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.921817 kubelet[2783]: W0911 00:20:19.921791 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.921817 kubelet[2783]: E0911 00:20:19.921804 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.922047 kubelet[2783]: E0911 00:20:19.922030 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.922047 kubelet[2783]: W0911 00:20:19.922042 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.922104 kubelet[2783]: E0911 00:20:19.922052 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.922278 kubelet[2783]: E0911 00:20:19.922262 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.922278 kubelet[2783]: W0911 00:20:19.922273 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.922446 kubelet[2783]: E0911 00:20:19.922282 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.922516 kubelet[2783]: E0911 00:20:19.922500 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.922562 kubelet[2783]: W0911 00:20:19.922521 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.922562 kubelet[2783]: E0911 00:20:19.922531 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.922776 kubelet[2783]: E0911 00:20:19.922746 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.922776 kubelet[2783]: W0911 00:20:19.922758 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.922776 kubelet[2783]: E0911 00:20:19.922767 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.923009 kubelet[2783]: E0911 00:20:19.922983 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.923009 kubelet[2783]: W0911 00:20:19.923005 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.923063 kubelet[2783]: E0911 00:20:19.923027 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.923298 kubelet[2783]: E0911 00:20:19.923281 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.923298 kubelet[2783]: W0911 00:20:19.923293 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.923365 kubelet[2783]: E0911 00:20:19.923303 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.923559 kubelet[2783]: E0911 00:20:19.923531 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.923559 kubelet[2783]: W0911 00:20:19.923557 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.923628 kubelet[2783]: E0911 00:20:19.923568 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.923773 kubelet[2783]: E0911 00:20:19.923750 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.923773 kubelet[2783]: W0911 00:20:19.923762 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.923773 kubelet[2783]: E0911 00:20:19.923772 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.924021 kubelet[2783]: E0911 00:20:19.924002 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.924021 kubelet[2783]: W0911 00:20:19.924013 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.924021 kubelet[2783]: E0911 00:20:19.924022 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.924250 kubelet[2783]: E0911 00:20:19.924228 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.924250 kubelet[2783]: W0911 00:20:19.924240 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.924250 kubelet[2783]: E0911 00:20:19.924248 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.924519 kubelet[2783]: E0911 00:20:19.924498 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.924519 kubelet[2783]: W0911 00:20:19.924513 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.924648 kubelet[2783]: E0911 00:20:19.924525 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.924732 kubelet[2783]: E0911 00:20:19.924710 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.924732 kubelet[2783]: W0911 00:20:19.924724 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.924732 kubelet[2783]: E0911 00:20:19.924733 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.924956 kubelet[2783]: E0911 00:20:19.924934 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.924956 kubelet[2783]: W0911 00:20:19.924946 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.925034 kubelet[2783]: E0911 00:20:19.924967 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.925183 kubelet[2783]: E0911 00:20:19.925162 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.925183 kubelet[2783]: W0911 00:20:19.925174 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.925183 kubelet[2783]: E0911 00:20:19.925183 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.926021 containerd[1571]: time="2025-09-11T00:20:19.925986116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-75tzc,Uid:31c180a7-51e1-4d4b-8de0-4257d3b41e99,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:19.938509 kubelet[2783]: E0911 00:20:19.938474 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.938509 kubelet[2783]: W0911 00:20:19.938493 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.938509 kubelet[2783]: E0911 00:20:19.938512 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.938737 kubelet[2783]: I0911 00:20:19.938538 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kczv7\" (UniqueName: \"kubernetes.io/projected/71ecab39-5d3e-4469-82ca-831c45fb7dcd-kube-api-access-kczv7\") pod \"csi-node-driver-f8r22\" (UID: \"71ecab39-5d3e-4469-82ca-831c45fb7dcd\") " pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:19.938874 kubelet[2783]: E0911 00:20:19.938835 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.938918 kubelet[2783]: W0911 00:20:19.938878 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.938918 kubelet[2783]: E0911 00:20:19.938911 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.938970 kubelet[2783]: I0911 00:20:19.938955 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71ecab39-5d3e-4469-82ca-831c45fb7dcd-kubelet-dir\") pod \"csi-node-driver-f8r22\" (UID: \"71ecab39-5d3e-4469-82ca-831c45fb7dcd\") " pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:19.939231 kubelet[2783]: E0911 00:20:19.939203 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.939231 kubelet[2783]: W0911 00:20:19.939227 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.939295 kubelet[2783]: E0911 00:20:19.939238 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.939676 kubelet[2783]: E0911 00:20:19.939646 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.939676 kubelet[2783]: W0911 00:20:19.939663 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.939676 kubelet[2783]: E0911 00:20:19.939672 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.940783 kubelet[2783]: E0911 00:20:19.939977 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.940783 kubelet[2783]: W0911 00:20:19.939993 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.940783 kubelet[2783]: E0911 00:20:19.940002 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.940783 kubelet[2783]: I0911 00:20:19.940025 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/71ecab39-5d3e-4469-82ca-831c45fb7dcd-registration-dir\") pod \"csi-node-driver-f8r22\" (UID: \"71ecab39-5d3e-4469-82ca-831c45fb7dcd\") " pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:19.940783 kubelet[2783]: E0911 00:20:19.940259 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.940783 kubelet[2783]: W0911 00:20:19.940268 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.940783 kubelet[2783]: E0911 00:20:19.940277 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.940783 kubelet[2783]: I0911 00:20:19.940396 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/71ecab39-5d3e-4469-82ca-831c45fb7dcd-socket-dir\") pod \"csi-node-driver-f8r22\" (UID: \"71ecab39-5d3e-4469-82ca-831c45fb7dcd\") " pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:19.940783 kubelet[2783]: E0911 00:20:19.940582 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.941030 kubelet[2783]: W0911 00:20:19.940598 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.941030 kubelet[2783]: E0911 00:20:19.940643 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.941030 kubelet[2783]: E0911 00:20:19.941004 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.941030 kubelet[2783]: W0911 00:20:19.941018 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.941125 kubelet[2783]: E0911 00:20:19.941038 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.941474 kubelet[2783]: E0911 00:20:19.941459 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.941474 kubelet[2783]: W0911 00:20:19.941471 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.941538 kubelet[2783]: E0911 00:20:19.941480 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.941538 kubelet[2783]: I0911 00:20:19.941499 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/71ecab39-5d3e-4469-82ca-831c45fb7dcd-varrun\") pod \"csi-node-driver-f8r22\" (UID: \"71ecab39-5d3e-4469-82ca-831c45fb7dcd\") " pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:19.941748 kubelet[2783]: E0911 00:20:19.941726 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.941748 kubelet[2783]: W0911 00:20:19.941742 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.941820 kubelet[2783]: E0911 00:20:19.941753 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.942091 kubelet[2783]: E0911 00:20:19.941953 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.942091 kubelet[2783]: W0911 00:20:19.941966 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.942091 kubelet[2783]: E0911 00:20:19.941976 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.942242 kubelet[2783]: E0911 00:20:19.942226 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.942242 kubelet[2783]: W0911 00:20:19.942238 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.942318 kubelet[2783]: E0911 00:20:19.942247 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.942655 kubelet[2783]: E0911 00:20:19.942627 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.942655 kubelet[2783]: W0911 00:20:19.942639 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.942655 kubelet[2783]: E0911 00:20:19.942649 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.943308 kubelet[2783]: E0911 00:20:19.943287 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.943308 kubelet[2783]: W0911 00:20:19.943298 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.943308 kubelet[2783]: E0911 00:20:19.943307 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.943524 kubelet[2783]: E0911 00:20:19.943509 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:19.943524 kubelet[2783]: W0911 00:20:19.943520 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:19.943712 kubelet[2783]: E0911 00:20:19.943529 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:19.951118 containerd[1571]: time="2025-09-11T00:20:19.950965015Z" level=info msg="connecting to shim 5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8" address="unix:///run/containerd/s/15910c11511e97773dc5b38149f47630b2db591ecc6a83b5332847c742ca5b5b" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:19.978690 systemd[1]: Started cri-containerd-5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8.scope - libcontainer container 5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8. Sep 11 00:20:20.043728 kubelet[2783]: E0911 00:20:20.043670 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.043728 kubelet[2783]: W0911 00:20:20.043705 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.043728 kubelet[2783]: E0911 00:20:20.043729 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.044081 kubelet[2783]: E0911 00:20:20.044045 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.044112 kubelet[2783]: W0911 00:20:20.044076 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.044112 kubelet[2783]: E0911 00:20:20.044105 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.044656 kubelet[2783]: E0911 00:20:20.044623 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.044717 kubelet[2783]: W0911 00:20:20.044654 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.044717 kubelet[2783]: E0911 00:20:20.044682 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.045010 kubelet[2783]: E0911 00:20:20.044992 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.045010 kubelet[2783]: W0911 00:20:20.045006 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.045081 kubelet[2783]: E0911 00:20:20.045020 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.045377 kubelet[2783]: E0911 00:20:20.045333 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.045377 kubelet[2783]: W0911 00:20:20.045370 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.045592 kubelet[2783]: E0911 00:20:20.045402 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.045825 kubelet[2783]: E0911 00:20:20.045799 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.045825 kubelet[2783]: W0911 00:20:20.045811 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.045825 kubelet[2783]: E0911 00:20:20.045821 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.046087 kubelet[2783]: E0911 00:20:20.046054 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.046087 kubelet[2783]: W0911 00:20:20.046070 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.046087 kubelet[2783]: E0911 00:20:20.046083 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.046350 kubelet[2783]: E0911 00:20:20.046335 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.046350 kubelet[2783]: W0911 00:20:20.046346 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.046404 kubelet[2783]: E0911 00:20:20.046357 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.046592 kubelet[2783]: E0911 00:20:20.046540 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.046592 kubelet[2783]: W0911 00:20:20.046575 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.046592 kubelet[2783]: E0911 00:20:20.046584 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.046777 kubelet[2783]: E0911 00:20:20.046756 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.046777 kubelet[2783]: W0911 00:20:20.046764 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.046777 kubelet[2783]: E0911 00:20:20.046772 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.046988 kubelet[2783]: E0911 00:20:20.046971 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.046988 kubelet[2783]: W0911 00:20:20.046981 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.046988 kubelet[2783]: E0911 00:20:20.046989 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.047572 kubelet[2783]: E0911 00:20:20.047269 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.047572 kubelet[2783]: W0911 00:20:20.047286 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.047572 kubelet[2783]: E0911 00:20:20.047296 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.047660 kubelet[2783]: E0911 00:20:20.047604 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.047660 kubelet[2783]: W0911 00:20:20.047637 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.047660 kubelet[2783]: E0911 00:20:20.047648 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.047927 kubelet[2783]: E0911 00:20:20.047903 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.047927 kubelet[2783]: W0911 00:20:20.047921 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.048041 kubelet[2783]: E0911 00:20:20.047939 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.048236 kubelet[2783]: E0911 00:20:20.048214 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.048236 kubelet[2783]: W0911 00:20:20.048234 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.048320 kubelet[2783]: E0911 00:20:20.048244 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.048612 kubelet[2783]: E0911 00:20:20.048594 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.048612 kubelet[2783]: W0911 00:20:20.048607 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.048681 kubelet[2783]: E0911 00:20:20.048617 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.048893 kubelet[2783]: E0911 00:20:20.048868 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.048893 kubelet[2783]: W0911 00:20:20.048887 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.048956 kubelet[2783]: E0911 00:20:20.048897 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.049202 kubelet[2783]: E0911 00:20:20.049182 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.049202 kubelet[2783]: W0911 00:20:20.049194 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.049202 kubelet[2783]: E0911 00:20:20.049203 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.049422 kubelet[2783]: E0911 00:20:20.049405 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.049422 kubelet[2783]: W0911 00:20:20.049416 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.049502 kubelet[2783]: E0911 00:20:20.049431 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.049664 kubelet[2783]: E0911 00:20:20.049648 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.049664 kubelet[2783]: W0911 00:20:20.049659 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.049724 kubelet[2783]: E0911 00:20:20.049668 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.049882 kubelet[2783]: E0911 00:20:20.049866 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.049882 kubelet[2783]: W0911 00:20:20.049877 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.049942 kubelet[2783]: E0911 00:20:20.049884 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.050110 kubelet[2783]: E0911 00:20:20.050095 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.050110 kubelet[2783]: W0911 00:20:20.050105 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.050165 kubelet[2783]: E0911 00:20:20.050114 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.050311 kubelet[2783]: E0911 00:20:20.050296 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.050311 kubelet[2783]: W0911 00:20:20.050307 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.050367 kubelet[2783]: E0911 00:20:20.050315 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.050495 kubelet[2783]: E0911 00:20:20.050480 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.050495 kubelet[2783]: W0911 00:20:20.050490 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.050556 kubelet[2783]: E0911 00:20:20.050499 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.050941 kubelet[2783]: E0911 00:20:20.050924 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.050941 kubelet[2783]: W0911 00:20:20.050936 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.050999 kubelet[2783]: E0911 00:20:20.050945 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:20.092753 containerd[1571]: time="2025-09-11T00:20:20.092700522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-75tzc,Uid:31c180a7-51e1-4d4b-8de0-4257d3b41e99,Namespace:calico-system,Attempt:0,} returns sandbox id \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\"" Sep 11 00:20:20.114518 kubelet[2783]: E0911 00:20:20.114466 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:20.114518 kubelet[2783]: W0911 00:20:20.114498 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:20.114518 kubelet[2783]: E0911 00:20:20.114522 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:21.354264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount371480536.mount: Deactivated successfully. Sep 11 00:20:21.771408 kubelet[2783]: E0911 00:20:21.771338 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:23.083338 containerd[1571]: time="2025-09-11T00:20:23.083264702Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:23.084618 containerd[1571]: time="2025-09-11T00:20:23.084580703Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Sep 11 00:20:23.085890 containerd[1571]: time="2025-09-11T00:20:23.085841812Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:23.088941 containerd[1571]: time="2025-09-11T00:20:23.088901378Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:23.089693 containerd[1571]: time="2025-09-11T00:20:23.089621200Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 3.309608712s" Sep 11 00:20:23.089693 containerd[1571]: time="2025-09-11T00:20:23.089683397Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Sep 11 00:20:23.090679 containerd[1571]: time="2025-09-11T00:20:23.090645904Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 11 00:20:23.105236 containerd[1571]: time="2025-09-11T00:20:23.105172136Z" level=info msg="CreateContainer within sandbox \"eb9349bf3627e6ff4dbb8a9c16fe035f444142abc5ca679a5a2ee83b3075d8b9\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 11 00:20:23.114083 containerd[1571]: time="2025-09-11T00:20:23.114037435Z" level=info msg="Container 6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:23.123156 containerd[1571]: time="2025-09-11T00:20:23.123092381Z" level=info msg="CreateContainer within sandbox \"eb9349bf3627e6ff4dbb8a9c16fe035f444142abc5ca679a5a2ee83b3075d8b9\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286\"" Sep 11 00:20:23.123678 containerd[1571]: time="2025-09-11T00:20:23.123646351Z" level=info msg="StartContainer for \"6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286\"" Sep 11 00:20:23.124757 containerd[1571]: time="2025-09-11T00:20:23.124723715Z" level=info msg="connecting to shim 6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286" address="unix:///run/containerd/s/df37dc01e0c12a26d76525595d52650901339b0e87f8af79161c5aa7764d30a3" protocol=ttrpc version=3 Sep 11 00:20:23.151723 systemd[1]: Started cri-containerd-6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286.scope - libcontainer container 6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286. Sep 11 00:20:23.204694 containerd[1571]: time="2025-09-11T00:20:23.204624613Z" level=info msg="StartContainer for \"6bd2d8ea182a0aa2dd34535197068426e77eeb1a50e0d5f3661529b39b5c1286\" returns successfully" Sep 11 00:20:23.772003 kubelet[2783]: E0911 00:20:23.770771 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:23.844845 kubelet[2783]: E0911 00:20:23.844791 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:23.847505 kubelet[2783]: E0911 00:20:23.847082 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.847505 kubelet[2783]: W0911 00:20:23.847122 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.847505 kubelet[2783]: E0911 00:20:23.847148 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.847505 kubelet[2783]: E0911 00:20:23.847344 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.847505 kubelet[2783]: W0911 00:20:23.847355 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.847505 kubelet[2783]: E0911 00:20:23.847366 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.847830 kubelet[2783]: E0911 00:20:23.847627 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.847830 kubelet[2783]: W0911 00:20:23.847646 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.847830 kubelet[2783]: E0911 00:20:23.847671 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.848592 kubelet[2783]: E0911 00:20:23.847977 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.848592 kubelet[2783]: W0911 00:20:23.847993 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.848592 kubelet[2783]: E0911 00:20:23.848002 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.848592 kubelet[2783]: E0911 00:20:23.848215 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.848592 kubelet[2783]: W0911 00:20:23.848225 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.848592 kubelet[2783]: E0911 00:20:23.848233 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.848592 kubelet[2783]: E0911 00:20:23.848408 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.848592 kubelet[2783]: W0911 00:20:23.848415 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.848592 kubelet[2783]: E0911 00:20:23.848423 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.848888 kubelet[2783]: E0911 00:20:23.848643 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.848888 kubelet[2783]: W0911 00:20:23.848651 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.848888 kubelet[2783]: E0911 00:20:23.848661 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.848888 kubelet[2783]: E0911 00:20:23.848820 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.848888 kubelet[2783]: W0911 00:20:23.848828 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.848888 kubelet[2783]: E0911 00:20:23.848836 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.849088 kubelet[2783]: E0911 00:20:23.848999 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.849088 kubelet[2783]: W0911 00:20:23.849006 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.849088 kubelet[2783]: E0911 00:20:23.849014 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.849197 kubelet[2783]: E0911 00:20:23.849176 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.849197 kubelet[2783]: W0911 00:20:23.849183 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.849197 kubelet[2783]: E0911 00:20:23.849192 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.849368 kubelet[2783]: E0911 00:20:23.849350 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.849368 kubelet[2783]: W0911 00:20:23.849361 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.849368 kubelet[2783]: E0911 00:20:23.849368 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.849587 kubelet[2783]: E0911 00:20:23.849572 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.849587 kubelet[2783]: W0911 00:20:23.849583 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.849658 kubelet[2783]: E0911 00:20:23.849592 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.849863 kubelet[2783]: E0911 00:20:23.849844 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.849913 kubelet[2783]: W0911 00:20:23.849862 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.849913 kubelet[2783]: E0911 00:20:23.849883 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.850134 kubelet[2783]: E0911 00:20:23.850109 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.850134 kubelet[2783]: W0911 00:20:23.850130 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.850205 kubelet[2783]: E0911 00:20:23.850140 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.850341 kubelet[2783]: E0911 00:20:23.850327 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.850341 kubelet[2783]: W0911 00:20:23.850337 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.850388 kubelet[2783]: E0911 00:20:23.850346 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.876802 kubelet[2783]: E0911 00:20:23.876746 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.876802 kubelet[2783]: W0911 00:20:23.876777 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.876802 kubelet[2783]: E0911 00:20:23.876802 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.877278 kubelet[2783]: E0911 00:20:23.877244 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.877278 kubelet[2783]: W0911 00:20:23.877258 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.877278 kubelet[2783]: E0911 00:20:23.877270 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.877608 kubelet[2783]: E0911 00:20:23.877586 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.877608 kubelet[2783]: W0911 00:20:23.877606 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.877707 kubelet[2783]: E0911 00:20:23.877621 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.877879 kubelet[2783]: E0911 00:20:23.877849 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.877879 kubelet[2783]: W0911 00:20:23.877865 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.877879 kubelet[2783]: E0911 00:20:23.877878 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.878121 kubelet[2783]: E0911 00:20:23.878095 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.878121 kubelet[2783]: W0911 00:20:23.878106 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.878183 kubelet[2783]: E0911 00:20:23.878129 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.878390 kubelet[2783]: E0911 00:20:23.878371 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.878390 kubelet[2783]: W0911 00:20:23.878384 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.878462 kubelet[2783]: E0911 00:20:23.878395 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.879002 kubelet[2783]: E0911 00:20:23.878711 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.879002 kubelet[2783]: W0911 00:20:23.878727 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.879002 kubelet[2783]: E0911 00:20:23.878739 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.879677 kubelet[2783]: E0911 00:20:23.879658 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.879677 kubelet[2783]: W0911 00:20:23.879673 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.879752 kubelet[2783]: E0911 00:20:23.879687 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.879904 kubelet[2783]: E0911 00:20:23.879887 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.879904 kubelet[2783]: W0911 00:20:23.879899 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.880001 kubelet[2783]: E0911 00:20:23.879908 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.880142 kubelet[2783]: E0911 00:20:23.880108 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.880142 kubelet[2783]: W0911 00:20:23.880130 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.880142 kubelet[2783]: E0911 00:20:23.880139 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.880409 kubelet[2783]: E0911 00:20:23.880383 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.880409 kubelet[2783]: W0911 00:20:23.880395 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.880409 kubelet[2783]: E0911 00:20:23.880404 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.880646 kubelet[2783]: E0911 00:20:23.880628 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.880646 kubelet[2783]: W0911 00:20:23.880640 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.880720 kubelet[2783]: E0911 00:20:23.880650 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.880880 kubelet[2783]: E0911 00:20:23.880863 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.880880 kubelet[2783]: W0911 00:20:23.880874 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.880952 kubelet[2783]: E0911 00:20:23.880884 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.881232 kubelet[2783]: E0911 00:20:23.881197 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.881232 kubelet[2783]: W0911 00:20:23.881214 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.881232 kubelet[2783]: E0911 00:20:23.881228 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.881459 kubelet[2783]: E0911 00:20:23.881431 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.881459 kubelet[2783]: W0911 00:20:23.881444 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.881459 kubelet[2783]: E0911 00:20:23.881455 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.881708 kubelet[2783]: E0911 00:20:23.881688 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.881708 kubelet[2783]: W0911 00:20:23.881700 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.881793 kubelet[2783]: E0911 00:20:23.881711 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.882029 kubelet[2783]: E0911 00:20:23.881998 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.882029 kubelet[2783]: W0911 00:20:23.882023 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.882101 kubelet[2783]: E0911 00:20:23.882045 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:23.882275 kubelet[2783]: E0911 00:20:23.882256 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:23.882275 kubelet[2783]: W0911 00:20:23.882267 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:23.882275 kubelet[2783]: E0911 00:20:23.882275 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.845346 kubelet[2783]: I0911 00:20:24.845303 2783 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 11 00:20:24.845791 kubelet[2783]: E0911 00:20:24.845656 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:24.857640 kubelet[2783]: E0911 00:20:24.857609 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.857640 kubelet[2783]: W0911 00:20:24.857627 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.857763 kubelet[2783]: E0911 00:20:24.857644 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.857943 kubelet[2783]: E0911 00:20:24.857922 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.857943 kubelet[2783]: W0911 00:20:24.857933 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.857943 kubelet[2783]: E0911 00:20:24.857941 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.858170 kubelet[2783]: E0911 00:20:24.858157 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.858170 kubelet[2783]: W0911 00:20:24.858167 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.858220 kubelet[2783]: E0911 00:20:24.858176 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.858383 kubelet[2783]: E0911 00:20:24.858370 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.858383 kubelet[2783]: W0911 00:20:24.858379 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.858447 kubelet[2783]: E0911 00:20:24.858387 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.858605 kubelet[2783]: E0911 00:20:24.858591 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.858605 kubelet[2783]: W0911 00:20:24.858600 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.858662 kubelet[2783]: E0911 00:20:24.858609 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.858784 kubelet[2783]: E0911 00:20:24.858773 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.858784 kubelet[2783]: W0911 00:20:24.858782 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.858835 kubelet[2783]: E0911 00:20:24.858789 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.858964 kubelet[2783]: E0911 00:20:24.858953 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.858964 kubelet[2783]: W0911 00:20:24.858961 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.859011 kubelet[2783]: E0911 00:20:24.858969 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.859143 kubelet[2783]: E0911 00:20:24.859131 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.859143 kubelet[2783]: W0911 00:20:24.859140 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.859203 kubelet[2783]: E0911 00:20:24.859147 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.859322 kubelet[2783]: E0911 00:20:24.859310 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.859322 kubelet[2783]: W0911 00:20:24.859319 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.859362 kubelet[2783]: E0911 00:20:24.859327 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.859477 kubelet[2783]: E0911 00:20:24.859466 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.859477 kubelet[2783]: W0911 00:20:24.859474 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.859538 kubelet[2783]: E0911 00:20:24.859482 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.859665 kubelet[2783]: E0911 00:20:24.859653 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.859665 kubelet[2783]: W0911 00:20:24.859662 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.859714 kubelet[2783]: E0911 00:20:24.859670 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.859841 kubelet[2783]: E0911 00:20:24.859829 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.859841 kubelet[2783]: W0911 00:20:24.859838 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.859895 kubelet[2783]: E0911 00:20:24.859845 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.860014 kubelet[2783]: E0911 00:20:24.860002 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.860014 kubelet[2783]: W0911 00:20:24.860011 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.860062 kubelet[2783]: E0911 00:20:24.860019 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.860199 kubelet[2783]: E0911 00:20:24.860187 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.860199 kubelet[2783]: W0911 00:20:24.860196 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.860249 kubelet[2783]: E0911 00:20:24.860205 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.860374 kubelet[2783]: E0911 00:20:24.860363 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.860374 kubelet[2783]: W0911 00:20:24.860371 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.860423 kubelet[2783]: E0911 00:20:24.860378 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.886007 kubelet[2783]: E0911 00:20:24.885968 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.886007 kubelet[2783]: W0911 00:20:24.885987 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.886007 kubelet[2783]: E0911 00:20:24.886007 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.886404 kubelet[2783]: E0911 00:20:24.886348 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.886448 kubelet[2783]: W0911 00:20:24.886397 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.886448 kubelet[2783]: E0911 00:20:24.886433 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.886845 kubelet[2783]: E0911 00:20:24.886823 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.886845 kubelet[2783]: W0911 00:20:24.886840 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.886919 kubelet[2783]: E0911 00:20:24.886853 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.887163 kubelet[2783]: E0911 00:20:24.887138 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.887163 kubelet[2783]: W0911 00:20:24.887153 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.887218 kubelet[2783]: E0911 00:20:24.887165 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.887351 kubelet[2783]: E0911 00:20:24.887338 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.887351 kubelet[2783]: W0911 00:20:24.887347 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.887420 kubelet[2783]: E0911 00:20:24.887356 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.887586 kubelet[2783]: E0911 00:20:24.887573 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.887586 kubelet[2783]: W0911 00:20:24.887582 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.887652 kubelet[2783]: E0911 00:20:24.887590 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.887973 kubelet[2783]: E0911 00:20:24.887951 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.887973 kubelet[2783]: W0911 00:20:24.887968 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.888046 kubelet[2783]: E0911 00:20:24.887981 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.888234 kubelet[2783]: E0911 00:20:24.888219 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.888234 kubelet[2783]: W0911 00:20:24.888230 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.888289 kubelet[2783]: E0911 00:20:24.888240 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.888424 kubelet[2783]: E0911 00:20:24.888411 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.888424 kubelet[2783]: W0911 00:20:24.888420 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.888487 kubelet[2783]: E0911 00:20:24.888429 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.888636 kubelet[2783]: E0911 00:20:24.888624 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.888636 kubelet[2783]: W0911 00:20:24.888633 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.888636 kubelet[2783]: E0911 00:20:24.888643 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.888893 kubelet[2783]: E0911 00:20:24.888874 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.888893 kubelet[2783]: W0911 00:20:24.888889 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.888981 kubelet[2783]: E0911 00:20:24.888903 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.889216 kubelet[2783]: E0911 00:20:24.889196 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.889216 kubelet[2783]: W0911 00:20:24.889211 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.889276 kubelet[2783]: E0911 00:20:24.889222 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.889436 kubelet[2783]: E0911 00:20:24.889419 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.889436 kubelet[2783]: W0911 00:20:24.889431 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.889498 kubelet[2783]: E0911 00:20:24.889439 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.889651 kubelet[2783]: E0911 00:20:24.889634 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.889651 kubelet[2783]: W0911 00:20:24.889646 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.889713 kubelet[2783]: E0911 00:20:24.889654 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.889826 kubelet[2783]: E0911 00:20:24.889810 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.889826 kubelet[2783]: W0911 00:20:24.889820 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.889885 kubelet[2783]: E0911 00:20:24.889828 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.890042 kubelet[2783]: E0911 00:20:24.890026 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.890042 kubelet[2783]: W0911 00:20:24.890037 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.890105 kubelet[2783]: E0911 00:20:24.890044 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.890331 kubelet[2783]: E0911 00:20:24.890303 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.890331 kubelet[2783]: W0911 00:20:24.890318 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.890331 kubelet[2783]: E0911 00:20:24.890328 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:24.890516 kubelet[2783]: E0911 00:20:24.890500 2783 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 11 00:20:24.890516 kubelet[2783]: W0911 00:20:24.890511 2783 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 11 00:20:24.890583 kubelet[2783]: E0911 00:20:24.890519 2783 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 11 00:20:25.221629 containerd[1571]: time="2025-09-11T00:20:25.221435894Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:25.222585 containerd[1571]: time="2025-09-11T00:20:25.222534256Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Sep 11 00:20:25.224336 containerd[1571]: time="2025-09-11T00:20:25.224200104Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:25.226851 containerd[1571]: time="2025-09-11T00:20:25.226807711Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:25.227732 containerd[1571]: time="2025-09-11T00:20:25.227693974Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 2.137013254s" Sep 11 00:20:25.227732 containerd[1571]: time="2025-09-11T00:20:25.227734331Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Sep 11 00:20:25.233370 containerd[1571]: time="2025-09-11T00:20:25.233314068Z" level=info msg="CreateContainer within sandbox \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 11 00:20:25.243607 containerd[1571]: time="2025-09-11T00:20:25.242761669Z" level=info msg="Container b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:25.252684 containerd[1571]: time="2025-09-11T00:20:25.252636742Z" level=info msg="CreateContainer within sandbox \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\"" Sep 11 00:20:25.253258 containerd[1571]: time="2025-09-11T00:20:25.253224325Z" level=info msg="StartContainer for \"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\"" Sep 11 00:20:25.254672 containerd[1571]: time="2025-09-11T00:20:25.254641246Z" level=info msg="connecting to shim b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc" address="unix:///run/containerd/s/15910c11511e97773dc5b38149f47630b2db591ecc6a83b5332847c742ca5b5b" protocol=ttrpc version=3 Sep 11 00:20:25.277716 systemd[1]: Started cri-containerd-b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc.scope - libcontainer container b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc. Sep 11 00:20:25.336450 systemd[1]: cri-containerd-b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc.scope: Deactivated successfully. Sep 11 00:20:25.339083 containerd[1571]: time="2025-09-11T00:20:25.339028275Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\" id:\"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\" pid:3513 exited_at:{seconds:1757550025 nanos:338298826}" Sep 11 00:20:25.612266 kubelet[2783]: I0911 00:20:25.611672 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-65dd79988f-7kwv5" podStartSLOduration=3.300846932 podStartE2EDuration="6.611648354s" podCreationTimestamp="2025-09-11 00:20:19 +0000 UTC" firstStartedPulling="2025-09-11 00:20:19.779633516 +0000 UTC m=+20.173437945" lastFinishedPulling="2025-09-11 00:20:23.090434938 +0000 UTC m=+23.484239367" observedRunningTime="2025-09-11 00:20:24.140404421 +0000 UTC m=+24.534208870" watchObservedRunningTime="2025-09-11 00:20:25.611648354 +0000 UTC m=+26.005452783" Sep 11 00:20:25.752002 containerd[1571]: time="2025-09-11T00:20:25.751914994Z" level=info msg="received exit event container_id:\"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\" id:\"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\" pid:3513 exited_at:{seconds:1757550025 nanos:338298826}" Sep 11 00:20:25.756567 containerd[1571]: time="2025-09-11T00:20:25.756454808Z" level=info msg="StartContainer for \"b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc\" returns successfully" Sep 11 00:20:25.770756 kubelet[2783]: E0911 00:20:25.770685 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:25.786945 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b8982d32995c2622be1537f8ef3bbc9052b7b6ad18637c8e4673767c71103bcc-rootfs.mount: Deactivated successfully. Sep 11 00:20:25.849836 kubelet[2783]: E0911 00:20:25.849783 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:26.853401 kubelet[2783]: E0911 00:20:26.853362 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:26.857506 containerd[1571]: time="2025-09-11T00:20:26.857454337Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 11 00:20:27.770527 kubelet[2783]: E0911 00:20:27.770461 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:29.771081 kubelet[2783]: E0911 00:20:29.770998 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:30.727900 containerd[1571]: time="2025-09-11T00:20:30.727822242Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:30.728724 containerd[1571]: time="2025-09-11T00:20:30.728697326Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Sep 11 00:20:30.730162 containerd[1571]: time="2025-09-11T00:20:30.730110248Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:30.732729 containerd[1571]: time="2025-09-11T00:20:30.732697996Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:30.733414 containerd[1571]: time="2025-09-11T00:20:30.733386609Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 3.875855197s" Sep 11 00:20:30.733414 containerd[1571]: time="2025-09-11T00:20:30.733413599Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Sep 11 00:20:30.740642 containerd[1571]: time="2025-09-11T00:20:30.740505492Z" level=info msg="CreateContainer within sandbox \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 11 00:20:30.751999 containerd[1571]: time="2025-09-11T00:20:30.751913326Z" level=info msg="Container 462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:30.769092 containerd[1571]: time="2025-09-11T00:20:30.769024338Z" level=info msg="CreateContainer within sandbox \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\"" Sep 11 00:20:30.769769 containerd[1571]: time="2025-09-11T00:20:30.769712379Z" level=info msg="StartContainer for \"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\"" Sep 11 00:20:30.771380 containerd[1571]: time="2025-09-11T00:20:30.771345434Z" level=info msg="connecting to shim 462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5" address="unix:///run/containerd/s/15910c11511e97773dc5b38149f47630b2db591ecc6a83b5332847c742ca5b5b" protocol=ttrpc version=3 Sep 11 00:20:30.797878 systemd[1]: Started cri-containerd-462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5.scope - libcontainer container 462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5. Sep 11 00:20:30.848794 containerd[1571]: time="2025-09-11T00:20:30.848736864Z" level=info msg="StartContainer for \"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\" returns successfully" Sep 11 00:20:31.770339 kubelet[2783]: E0911 00:20:31.770250 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:33.064068 containerd[1571]: time="2025-09-11T00:20:33.064009078Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 11 00:20:33.067620 systemd[1]: cri-containerd-462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5.scope: Deactivated successfully. Sep 11 00:20:33.067998 systemd[1]: cri-containerd-462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5.scope: Consumed 608ms CPU time, 177.6M memory peak, 3.8M read from disk, 171.3M written to disk. Sep 11 00:20:33.068643 containerd[1571]: time="2025-09-11T00:20:33.068417903Z" level=info msg="received exit event container_id:\"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\" id:\"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\" pid:3575 exited_at:{seconds:1757550033 nanos:68182822}" Sep 11 00:20:33.068643 containerd[1571]: time="2025-09-11T00:20:33.068592010Z" level=info msg="TaskExit event in podsandbox handler container_id:\"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\" id:\"462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5\" pid:3575 exited_at:{seconds:1757550033 nanos:68182822}" Sep 11 00:20:33.093350 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-462a13d93962aa4daf33997668e5c63c2e3c5579d31b3170d957ca8ca2047be5-rootfs.mount: Deactivated successfully. Sep 11 00:20:33.131738 kubelet[2783]: I0911 00:20:33.131693 2783 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 11 00:20:33.280443 systemd[1]: Created slice kubepods-burstable-podf4246163_a863_4efc_b1a5_f23f0e9c31b9.slice - libcontainer container kubepods-burstable-podf4246163_a863_4efc_b1a5_f23f0e9c31b9.slice. Sep 11 00:20:33.290603 systemd[1]: Created slice kubepods-besteffort-pod78b30ee7_7aa4_4eb9_858b_ae86f51e74f5.slice - libcontainer container kubepods-besteffort-pod78b30ee7_7aa4_4eb9_858b_ae86f51e74f5.slice. Sep 11 00:20:33.300142 systemd[1]: Created slice kubepods-burstable-pod2f2a0328_8013_43e7_aa22_1250dbd184a5.slice - libcontainer container kubepods-burstable-pod2f2a0328_8013_43e7_aa22_1250dbd184a5.slice. Sep 11 00:20:33.311390 systemd[1]: Created slice kubepods-besteffort-podd228d31d_37f9_4d90_bbcc_389dcb1a76f8.slice - libcontainer container kubepods-besteffort-podd228d31d_37f9_4d90_bbcc_389dcb1a76f8.slice. Sep 11 00:20:33.321367 systemd[1]: Created slice kubepods-besteffort-pod25061c9c_4e54_43e9_a333_25e7b1cb5982.slice - libcontainer container kubepods-besteffort-pod25061c9c_4e54_43e9_a333_25e7b1cb5982.slice. Sep 11 00:20:33.331837 systemd[1]: Created slice kubepods-besteffort-pod90243719_01e5_455a_a7a2_8d684145bb0e.slice - libcontainer container kubepods-besteffort-pod90243719_01e5_455a_a7a2_8d684145bb0e.slice. Sep 11 00:20:33.339445 systemd[1]: Created slice kubepods-besteffort-podabb44784_55a4_4f8a_ac18_b17c33a6fc96.slice - libcontainer container kubepods-besteffort-podabb44784_55a4_4f8a_ac18_b17c33a6fc96.slice. Sep 11 00:20:33.375690 kubelet[2783]: I0911 00:20:33.375621 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r6xg\" (UniqueName: \"kubernetes.io/projected/f4246163-a863-4efc-b1a5-f23f0e9c31b9-kube-api-access-6r6xg\") pod \"coredns-674b8bbfcf-bcvvs\" (UID: \"f4246163-a863-4efc-b1a5-f23f0e9c31b9\") " pod="kube-system/coredns-674b8bbfcf-bcvvs" Sep 11 00:20:33.375690 kubelet[2783]: I0911 00:20:33.375672 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvkss\" (UniqueName: \"kubernetes.io/projected/25061c9c-4e54-43e9-a333-25e7b1cb5982-kube-api-access-tvkss\") pod \"calico-apiserver-7bd7c7b5ff-rr4tc\" (UID: \"25061c9c-4e54-43e9-a333-25e7b1cb5982\") " pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" Sep 11 00:20:33.375690 kubelet[2783]: I0911 00:20:33.375693 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/78b30ee7-7aa4-4eb9-858b-ae86f51e74f5-calico-apiserver-certs\") pod \"calico-apiserver-7bd7c7b5ff-pfr6f\" (UID: \"78b30ee7-7aa4-4eb9-858b-ae86f51e74f5\") " pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-pfr6f" Sep 11 00:20:33.375690 kubelet[2783]: I0911 00:20:33.375707 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snr2d\" (UniqueName: \"kubernetes.io/projected/78b30ee7-7aa4-4eb9-858b-ae86f51e74f5-kube-api-access-snr2d\") pod \"calico-apiserver-7bd7c7b5ff-pfr6f\" (UID: \"78b30ee7-7aa4-4eb9-858b-ae86f51e74f5\") " pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-pfr6f" Sep 11 00:20:33.375990 kubelet[2783]: I0911 00:20:33.375723 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkmr7\" (UniqueName: \"kubernetes.io/projected/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-kube-api-access-nkmr7\") pod \"whisker-7d85b6b7df-wkzk6\" (UID: \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\") " pod="calico-system/whisker-7d85b6b7df-wkzk6" Sep 11 00:20:33.375990 kubelet[2783]: I0911 00:20:33.375821 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bfwv\" (UniqueName: \"kubernetes.io/projected/90243719-01e5-455a-a7a2-8d684145bb0e-kube-api-access-5bfwv\") pod \"calico-kube-controllers-6544b88998-5nqjt\" (UID: \"90243719-01e5-455a-a7a2-8d684145bb0e\") " pod="calico-system/calico-kube-controllers-6544b88998-5nqjt" Sep 11 00:20:33.376104 kubelet[2783]: I0911 00:20:33.376023 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-backend-key-pair\") pod \"whisker-7d85b6b7df-wkzk6\" (UID: \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\") " pod="calico-system/whisker-7d85b6b7df-wkzk6" Sep 11 00:20:33.376166 kubelet[2783]: I0911 00:20:33.376141 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90243719-01e5-455a-a7a2-8d684145bb0e-tigera-ca-bundle\") pod \"calico-kube-controllers-6544b88998-5nqjt\" (UID: \"90243719-01e5-455a-a7a2-8d684145bb0e\") " pod="calico-system/calico-kube-controllers-6544b88998-5nqjt" Sep 11 00:20:33.376209 kubelet[2783]: I0911 00:20:33.376171 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-ca-bundle\") pod \"whisker-7d85b6b7df-wkzk6\" (UID: \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\") " pod="calico-system/whisker-7d85b6b7df-wkzk6" Sep 11 00:20:33.376209 kubelet[2783]: I0911 00:20:33.376198 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abb44784-55a4-4f8a-ac18-b17c33a6fc96-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-k7tqf\" (UID: \"abb44784-55a4-4f8a-ac18-b17c33a6fc96\") " pod="calico-system/goldmane-54d579b49d-k7tqf" Sep 11 00:20:33.376273 kubelet[2783]: I0911 00:20:33.376234 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f4246163-a863-4efc-b1a5-f23f0e9c31b9-config-volume\") pod \"coredns-674b8bbfcf-bcvvs\" (UID: \"f4246163-a863-4efc-b1a5-f23f0e9c31b9\") " pod="kube-system/coredns-674b8bbfcf-bcvvs" Sep 11 00:20:33.376312 kubelet[2783]: I0911 00:20:33.376286 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f2a0328-8013-43e7-aa22-1250dbd184a5-config-volume\") pod \"coredns-674b8bbfcf-2h255\" (UID: \"2f2a0328-8013-43e7-aa22-1250dbd184a5\") " pod="kube-system/coredns-674b8bbfcf-2h255" Sep 11 00:20:33.376346 kubelet[2783]: I0911 00:20:33.376315 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb44784-55a4-4f8a-ac18-b17c33a6fc96-config\") pod \"goldmane-54d579b49d-k7tqf\" (UID: \"abb44784-55a4-4f8a-ac18-b17c33a6fc96\") " pod="calico-system/goldmane-54d579b49d-k7tqf" Sep 11 00:20:33.376346 kubelet[2783]: I0911 00:20:33.376339 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/abb44784-55a4-4f8a-ac18-b17c33a6fc96-goldmane-key-pair\") pod \"goldmane-54d579b49d-k7tqf\" (UID: \"abb44784-55a4-4f8a-ac18-b17c33a6fc96\") " pod="calico-system/goldmane-54d579b49d-k7tqf" Sep 11 00:20:33.376406 kubelet[2783]: I0911 00:20:33.376364 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95mhn\" (UniqueName: \"kubernetes.io/projected/abb44784-55a4-4f8a-ac18-b17c33a6fc96-kube-api-access-95mhn\") pod \"goldmane-54d579b49d-k7tqf\" (UID: \"abb44784-55a4-4f8a-ac18-b17c33a6fc96\") " pod="calico-system/goldmane-54d579b49d-k7tqf" Sep 11 00:20:33.376406 kubelet[2783]: I0911 00:20:33.376400 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gq9l\" (UniqueName: \"kubernetes.io/projected/2f2a0328-8013-43e7-aa22-1250dbd184a5-kube-api-access-7gq9l\") pod \"coredns-674b8bbfcf-2h255\" (UID: \"2f2a0328-8013-43e7-aa22-1250dbd184a5\") " pod="kube-system/coredns-674b8bbfcf-2h255" Sep 11 00:20:33.376476 kubelet[2783]: I0911 00:20:33.376431 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/25061c9c-4e54-43e9-a333-25e7b1cb5982-calico-apiserver-certs\") pod \"calico-apiserver-7bd7c7b5ff-rr4tc\" (UID: \"25061c9c-4e54-43e9-a333-25e7b1cb5982\") " pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" Sep 11 00:20:33.588781 kubelet[2783]: E0911 00:20:33.588654 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:33.589224 containerd[1571]: time="2025-09-11T00:20:33.589193079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-bcvvs,Uid:f4246163-a863-4efc-b1a5-f23f0e9c31b9,Namespace:kube-system,Attempt:0,}" Sep 11 00:20:33.595028 containerd[1571]: time="2025-09-11T00:20:33.595003154Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-pfr6f,Uid:78b30ee7-7aa4-4eb9-858b-ae86f51e74f5,Namespace:calico-apiserver,Attempt:0,}" Sep 11 00:20:33.604562 kubelet[2783]: E0911 00:20:33.604524 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:33.605910 containerd[1571]: time="2025-09-11T00:20:33.605848040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2h255,Uid:2f2a0328-8013-43e7-aa22-1250dbd184a5,Namespace:kube-system,Attempt:0,}" Sep 11 00:20:33.618593 containerd[1571]: time="2025-09-11T00:20:33.618497562Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7d85b6b7df-wkzk6,Uid:d228d31d-37f9-4d90-bbcc-389dcb1a76f8,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:33.629604 containerd[1571]: time="2025-09-11T00:20:33.629034760Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-rr4tc,Uid:25061c9c-4e54-43e9-a333-25e7b1cb5982,Namespace:calico-apiserver,Attempt:0,}" Sep 11 00:20:33.639010 containerd[1571]: time="2025-09-11T00:20:33.638954979Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6544b88998-5nqjt,Uid:90243719-01e5-455a-a7a2-8d684145bb0e,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:33.644196 containerd[1571]: time="2025-09-11T00:20:33.644169607Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-k7tqf,Uid:abb44784-55a4-4f8a-ac18-b17c33a6fc96,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:33.732880 containerd[1571]: time="2025-09-11T00:20:33.732806934Z" level=error msg="Failed to destroy network for sandbox \"294026c2af69d8f4ff7bf0ef8bbe86880f53440d084a301e147ca58ba63e7123\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.753646 containerd[1571]: time="2025-09-11T00:20:33.745630494Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-pfr6f,Uid:78b30ee7-7aa4-4eb9-858b-ae86f51e74f5,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"294026c2af69d8f4ff7bf0ef8bbe86880f53440d084a301e147ca58ba63e7123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.763874 kubelet[2783]: E0911 00:20:33.763796 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"294026c2af69d8f4ff7bf0ef8bbe86880f53440d084a301e147ca58ba63e7123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.764067 kubelet[2783]: E0911 00:20:33.763898 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"294026c2af69d8f4ff7bf0ef8bbe86880f53440d084a301e147ca58ba63e7123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-pfr6f" Sep 11 00:20:33.764067 kubelet[2783]: E0911 00:20:33.763933 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"294026c2af69d8f4ff7bf0ef8bbe86880f53440d084a301e147ca58ba63e7123\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-pfr6f" Sep 11 00:20:33.764067 kubelet[2783]: E0911 00:20:33.764002 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7bd7c7b5ff-pfr6f_calico-apiserver(78b30ee7-7aa4-4eb9-858b-ae86f51e74f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7bd7c7b5ff-pfr6f_calico-apiserver(78b30ee7-7aa4-4eb9-858b-ae86f51e74f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"294026c2af69d8f4ff7bf0ef8bbe86880f53440d084a301e147ca58ba63e7123\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-pfr6f" podUID="78b30ee7-7aa4-4eb9-858b-ae86f51e74f5" Sep 11 00:20:33.781339 containerd[1571]: time="2025-09-11T00:20:33.781273824Z" level=error msg="Failed to destroy network for sandbox \"54c644ee715ab0b646d0ccb543db08bd08f86dc9c45a97b36ad337bd4c7205cb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.783178 containerd[1571]: time="2025-09-11T00:20:33.783138865Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7d85b6b7df-wkzk6,Uid:d228d31d-37f9-4d90-bbcc-389dcb1a76f8,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"54c644ee715ab0b646d0ccb543db08bd08f86dc9c45a97b36ad337bd4c7205cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.783429 kubelet[2783]: E0911 00:20:33.783396 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"54c644ee715ab0b646d0ccb543db08bd08f86dc9c45a97b36ad337bd4c7205cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.784018 kubelet[2783]: E0911 00:20:33.783991 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"54c644ee715ab0b646d0ccb543db08bd08f86dc9c45a97b36ad337bd4c7205cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7d85b6b7df-wkzk6" Sep 11 00:20:33.784168 kubelet[2783]: E0911 00:20:33.784114 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"54c644ee715ab0b646d0ccb543db08bd08f86dc9c45a97b36ad337bd4c7205cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7d85b6b7df-wkzk6" Sep 11 00:20:33.785689 kubelet[2783]: E0911 00:20:33.785624 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7d85b6b7df-wkzk6_calico-system(d228d31d-37f9-4d90-bbcc-389dcb1a76f8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7d85b6b7df-wkzk6_calico-system(d228d31d-37f9-4d90-bbcc-389dcb1a76f8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"54c644ee715ab0b646d0ccb543db08bd08f86dc9c45a97b36ad337bd4c7205cb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7d85b6b7df-wkzk6" podUID="d228d31d-37f9-4d90-bbcc-389dcb1a76f8" Sep 11 00:20:33.788137 systemd[1]: Created slice kubepods-besteffort-pod71ecab39_5d3e_4469_82ca_831c45fb7dcd.slice - libcontainer container kubepods-besteffort-pod71ecab39_5d3e_4469_82ca_831c45fb7dcd.slice. Sep 11 00:20:33.791736 containerd[1571]: time="2025-09-11T00:20:33.791688782Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-f8r22,Uid:71ecab39-5d3e-4469-82ca-831c45fb7dcd,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:33.798657 containerd[1571]: time="2025-09-11T00:20:33.798598762Z" level=error msg="Failed to destroy network for sandbox \"1e3da5ba63d2cc8c36c3918026dad3e3df1497ee98da203b688da1caf02f04e3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.800428 containerd[1571]: time="2025-09-11T00:20:33.800393440Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-k7tqf,Uid:abb44784-55a4-4f8a-ac18-b17c33a6fc96,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e3da5ba63d2cc8c36c3918026dad3e3df1497ee98da203b688da1caf02f04e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.801017 kubelet[2783]: E0911 00:20:33.800972 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e3da5ba63d2cc8c36c3918026dad3e3df1497ee98da203b688da1caf02f04e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.801183 kubelet[2783]: E0911 00:20:33.801152 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e3da5ba63d2cc8c36c3918026dad3e3df1497ee98da203b688da1caf02f04e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-k7tqf" Sep 11 00:20:33.801301 kubelet[2783]: E0911 00:20:33.801278 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e3da5ba63d2cc8c36c3918026dad3e3df1497ee98da203b688da1caf02f04e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-k7tqf" Sep 11 00:20:33.802467 kubelet[2783]: E0911 00:20:33.801426 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-k7tqf_calico-system(abb44784-55a4-4f8a-ac18-b17c33a6fc96)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-k7tqf_calico-system(abb44784-55a4-4f8a-ac18-b17c33a6fc96)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1e3da5ba63d2cc8c36c3918026dad3e3df1497ee98da203b688da1caf02f04e3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-k7tqf" podUID="abb44784-55a4-4f8a-ac18-b17c33a6fc96" Sep 11 00:20:33.812649 containerd[1571]: time="2025-09-11T00:20:33.812582208Z" level=error msg="Failed to destroy network for sandbox \"de647377d92aac9b1d69e26fb1f1f6a7404cb7e26bab2d443e7e1632117d7418\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.815046 containerd[1571]: time="2025-09-11T00:20:33.814988955Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-bcvvs,Uid:f4246163-a863-4efc-b1a5-f23f0e9c31b9,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"de647377d92aac9b1d69e26fb1f1f6a7404cb7e26bab2d443e7e1632117d7418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.815300 kubelet[2783]: E0911 00:20:33.815258 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"de647377d92aac9b1d69e26fb1f1f6a7404cb7e26bab2d443e7e1632117d7418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.815384 kubelet[2783]: E0911 00:20:33.815328 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"de647377d92aac9b1d69e26fb1f1f6a7404cb7e26bab2d443e7e1632117d7418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-bcvvs" Sep 11 00:20:33.815384 kubelet[2783]: E0911 00:20:33.815351 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"de647377d92aac9b1d69e26fb1f1f6a7404cb7e26bab2d443e7e1632117d7418\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-bcvvs" Sep 11 00:20:33.815465 kubelet[2783]: E0911 00:20:33.815398 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-bcvvs_kube-system(f4246163-a863-4efc-b1a5-f23f0e9c31b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-bcvvs_kube-system(f4246163-a863-4efc-b1a5-f23f0e9c31b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"de647377d92aac9b1d69e26fb1f1f6a7404cb7e26bab2d443e7e1632117d7418\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-bcvvs" podUID="f4246163-a863-4efc-b1a5-f23f0e9c31b9" Sep 11 00:20:33.825901 containerd[1571]: time="2025-09-11T00:20:33.825747358Z" level=error msg="Failed to destroy network for sandbox \"610364fd03d64c2c519d4e1c874b0bf0183772316f6c14542ddc3dd7ed28e0a3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.826444 containerd[1571]: time="2025-09-11T00:20:33.826410112Z" level=error msg="Failed to destroy network for sandbox \"1e4144365760f64e4f9393f797fd3c95888ba256d2e7c7d4781edab85b714161\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.827726 containerd[1571]: time="2025-09-11T00:20:33.827693331Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-rr4tc,Uid:25061c9c-4e54-43e9-a333-25e7b1cb5982,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"610364fd03d64c2c519d4e1c874b0bf0183772316f6c14542ddc3dd7ed28e0a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.828797 kubelet[2783]: E0911 00:20:33.828198 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"610364fd03d64c2c519d4e1c874b0bf0183772316f6c14542ddc3dd7ed28e0a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.828797 kubelet[2783]: E0911 00:20:33.828268 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"610364fd03d64c2c519d4e1c874b0bf0183772316f6c14542ddc3dd7ed28e0a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" Sep 11 00:20:33.828797 kubelet[2783]: E0911 00:20:33.828290 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"610364fd03d64c2c519d4e1c874b0bf0183772316f6c14542ddc3dd7ed28e0a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" Sep 11 00:20:33.828974 kubelet[2783]: E0911 00:20:33.828366 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7bd7c7b5ff-rr4tc_calico-apiserver(25061c9c-4e54-43e9-a333-25e7b1cb5982)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7bd7c7b5ff-rr4tc_calico-apiserver(25061c9c-4e54-43e9-a333-25e7b1cb5982)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"610364fd03d64c2c519d4e1c874b0bf0183772316f6c14542ddc3dd7ed28e0a3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" podUID="25061c9c-4e54-43e9-a333-25e7b1cb5982" Sep 11 00:20:33.829687 containerd[1571]: time="2025-09-11T00:20:33.829650955Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2h255,Uid:2f2a0328-8013-43e7-aa22-1250dbd184a5,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e4144365760f64e4f9393f797fd3c95888ba256d2e7c7d4781edab85b714161\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.829980 kubelet[2783]: E0911 00:20:33.829953 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e4144365760f64e4f9393f797fd3c95888ba256d2e7c7d4781edab85b714161\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.830110 kubelet[2783]: E0911 00:20:33.830088 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e4144365760f64e4f9393f797fd3c95888ba256d2e7c7d4781edab85b714161\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-2h255" Sep 11 00:20:33.830224 kubelet[2783]: E0911 00:20:33.830201 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e4144365760f64e4f9393f797fd3c95888ba256d2e7c7d4781edab85b714161\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-2h255" Sep 11 00:20:33.830403 kubelet[2783]: E0911 00:20:33.830338 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-2h255_kube-system(2f2a0328-8013-43e7-aa22-1250dbd184a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-2h255_kube-system(2f2a0328-8013-43e7-aa22-1250dbd184a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1e4144365760f64e4f9393f797fd3c95888ba256d2e7c7d4781edab85b714161\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-2h255" podUID="2f2a0328-8013-43e7-aa22-1250dbd184a5" Sep 11 00:20:33.833868 containerd[1571]: time="2025-09-11T00:20:33.833794121Z" level=error msg="Failed to destroy network for sandbox \"c8f5c12e1d2276b6f61fdeb8f3c735a3ea20995b4d115a0ad55d8dd68e114c6d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.835730 containerd[1571]: time="2025-09-11T00:20:33.835652380Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6544b88998-5nqjt,Uid:90243719-01e5-455a-a7a2-8d684145bb0e,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c8f5c12e1d2276b6f61fdeb8f3c735a3ea20995b4d115a0ad55d8dd68e114c6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.836041 kubelet[2783]: E0911 00:20:33.835989 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c8f5c12e1d2276b6f61fdeb8f3c735a3ea20995b4d115a0ad55d8dd68e114c6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.836134 kubelet[2783]: E0911 00:20:33.836067 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c8f5c12e1d2276b6f61fdeb8f3c735a3ea20995b4d115a0ad55d8dd68e114c6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6544b88998-5nqjt" Sep 11 00:20:33.836134 kubelet[2783]: E0911 00:20:33.836087 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c8f5c12e1d2276b6f61fdeb8f3c735a3ea20995b4d115a0ad55d8dd68e114c6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6544b88998-5nqjt" Sep 11 00:20:33.836206 kubelet[2783]: E0911 00:20:33.836135 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6544b88998-5nqjt_calico-system(90243719-01e5-455a-a7a2-8d684145bb0e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6544b88998-5nqjt_calico-system(90243719-01e5-455a-a7a2-8d684145bb0e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c8f5c12e1d2276b6f61fdeb8f3c735a3ea20995b4d115a0ad55d8dd68e114c6d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6544b88998-5nqjt" podUID="90243719-01e5-455a-a7a2-8d684145bb0e" Sep 11 00:20:33.881126 containerd[1571]: time="2025-09-11T00:20:33.880948026Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 11 00:20:33.890623 containerd[1571]: time="2025-09-11T00:20:33.890491619Z" level=error msg="Failed to destroy network for sandbox \"dfed2541ca2c3ae95c1456e23a0c09618590e595d855d5e465f7318e373ffddd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.892576 containerd[1571]: time="2025-09-11T00:20:33.892475103Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-f8r22,Uid:71ecab39-5d3e-4469-82ca-831c45fb7dcd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfed2541ca2c3ae95c1456e23a0c09618590e595d855d5e465f7318e373ffddd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.892967 kubelet[2783]: E0911 00:20:33.892807 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfed2541ca2c3ae95c1456e23a0c09618590e595d855d5e465f7318e373ffddd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:33.892967 kubelet[2783]: E0911 00:20:33.892884 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfed2541ca2c3ae95c1456e23a0c09618590e595d855d5e465f7318e373ffddd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:33.892967 kubelet[2783]: E0911 00:20:33.892930 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfed2541ca2c3ae95c1456e23a0c09618590e595d855d5e465f7318e373ffddd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:33.893202 kubelet[2783]: E0911 00:20:33.892990 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-f8r22_calico-system(71ecab39-5d3e-4469-82ca-831c45fb7dcd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-f8r22_calico-system(71ecab39-5d3e-4469-82ca-831c45fb7dcd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dfed2541ca2c3ae95c1456e23a0c09618590e595d855d5e465f7318e373ffddd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:43.215756 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1462666797.mount: Deactivated successfully. Sep 11 00:20:44.771318 kubelet[2783]: E0911 00:20:44.771241 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:44.772033 containerd[1571]: time="2025-09-11T00:20:44.771715832Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-f8r22,Uid:71ecab39-5d3e-4469-82ca-831c45fb7dcd,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:44.772033 containerd[1571]: time="2025-09-11T00:20:44.771757072Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-bcvvs,Uid:f4246163-a863-4efc-b1a5-f23f0e9c31b9,Namespace:kube-system,Attempt:0,}" Sep 11 00:20:44.772033 containerd[1571]: time="2025-09-11T00:20:44.771721182Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-rr4tc,Uid:25061c9c-4e54-43e9-a333-25e7b1cb5982,Namespace:calico-apiserver,Attempt:0,}" Sep 11 00:20:45.051293 containerd[1571]: time="2025-09-11T00:20:45.051149303Z" level=error msg="Failed to destroy network for sandbox \"c730206e38b3f80c91acbbfcb5f08d0f861d2db2ff9d617cd0bf2f0cbe89346b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.054263 systemd[1]: run-netns-cni\x2d61943d2a\x2d6b04\x2d4b8a\x2d7c5b\x2dfd846bef74b6.mount: Deactivated successfully. Sep 11 00:20:45.070836 containerd[1571]: time="2025-09-11T00:20:45.070774273Z" level=error msg="Failed to destroy network for sandbox \"e3a0bdd38a760b8b7705a32c57baae5192f5bd5df781acedbbc3cf075f2ba140\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.072823 containerd[1571]: time="2025-09-11T00:20:45.072380931Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:45.073691 systemd[1]: run-netns-cni\x2d2743a5ea\x2d7dca\x2d56fd\x2d1367\x2d9b8996859276.mount: Deactivated successfully. Sep 11 00:20:45.090047 containerd[1571]: time="2025-09-11T00:20:45.089974382Z" level=error msg="Failed to destroy network for sandbox \"929dce07d0c06c26df811b6ca736f3db012d15960f37a22415989867aa9715f5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.263043 containerd[1571]: time="2025-09-11T00:20:45.262776158Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Sep 11 00:20:45.296883 containerd[1571]: time="2025-09-11T00:20:45.296771088Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-bcvvs,Uid:f4246163-a863-4efc-b1a5-f23f0e9c31b9,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c730206e38b3f80c91acbbfcb5f08d0f861d2db2ff9d617cd0bf2f0cbe89346b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.297240 kubelet[2783]: E0911 00:20:45.297185 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c730206e38b3f80c91acbbfcb5f08d0f861d2db2ff9d617cd0bf2f0cbe89346b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.297370 kubelet[2783]: E0911 00:20:45.297275 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c730206e38b3f80c91acbbfcb5f08d0f861d2db2ff9d617cd0bf2f0cbe89346b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-bcvvs" Sep 11 00:20:45.297370 kubelet[2783]: E0911 00:20:45.297298 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c730206e38b3f80c91acbbfcb5f08d0f861d2db2ff9d617cd0bf2f0cbe89346b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-bcvvs" Sep 11 00:20:45.297435 kubelet[2783]: E0911 00:20:45.297372 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-bcvvs_kube-system(f4246163-a863-4efc-b1a5-f23f0e9c31b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-bcvvs_kube-system(f4246163-a863-4efc-b1a5-f23f0e9c31b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c730206e38b3f80c91acbbfcb5f08d0f861d2db2ff9d617cd0bf2f0cbe89346b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-bcvvs" podUID="f4246163-a863-4efc-b1a5-f23f0e9c31b9" Sep 11 00:20:45.299270 containerd[1571]: time="2025-09-11T00:20:45.299209033Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-f8r22,Uid:71ecab39-5d3e-4469-82ca-831c45fb7dcd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3a0bdd38a760b8b7705a32c57baae5192f5bd5df781acedbbc3cf075f2ba140\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.299530 kubelet[2783]: E0911 00:20:45.299478 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3a0bdd38a760b8b7705a32c57baae5192f5bd5df781acedbbc3cf075f2ba140\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.300280 kubelet[2783]: E0911 00:20:45.300173 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3a0bdd38a760b8b7705a32c57baae5192f5bd5df781acedbbc3cf075f2ba140\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:45.300280 kubelet[2783]: E0911 00:20:45.300207 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3a0bdd38a760b8b7705a32c57baae5192f5bd5df781acedbbc3cf075f2ba140\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-f8r22" Sep 11 00:20:45.301834 kubelet[2783]: E0911 00:20:45.300289 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-f8r22_calico-system(71ecab39-5d3e-4469-82ca-831c45fb7dcd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-f8r22_calico-system(71ecab39-5d3e-4469-82ca-831c45fb7dcd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e3a0bdd38a760b8b7705a32c57baae5192f5bd5df781acedbbc3cf075f2ba140\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-f8r22" podUID="71ecab39-5d3e-4469-82ca-831c45fb7dcd" Sep 11 00:20:45.301834 kubelet[2783]: E0911 00:20:45.300938 2783 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"929dce07d0c06c26df811b6ca736f3db012d15960f37a22415989867aa9715f5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.301834 kubelet[2783]: E0911 00:20:45.300972 2783 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"929dce07d0c06c26df811b6ca736f3db012d15960f37a22415989867aa9715f5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" Sep 11 00:20:45.302083 containerd[1571]: time="2025-09-11T00:20:45.300752086Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-rr4tc,Uid:25061c9c-4e54-43e9-a333-25e7b1cb5982,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"929dce07d0c06c26df811b6ca736f3db012d15960f37a22415989867aa9715f5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 11 00:20:45.302185 kubelet[2783]: E0911 00:20:45.300993 2783 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"929dce07d0c06c26df811b6ca736f3db012d15960f37a22415989867aa9715f5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" Sep 11 00:20:45.302185 kubelet[2783]: E0911 00:20:45.301027 2783 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7bd7c7b5ff-rr4tc_calico-apiserver(25061c9c-4e54-43e9-a333-25e7b1cb5982)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7bd7c7b5ff-rr4tc_calico-apiserver(25061c9c-4e54-43e9-a333-25e7b1cb5982)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"929dce07d0c06c26df811b6ca736f3db012d15960f37a22415989867aa9715f5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" podUID="25061c9c-4e54-43e9-a333-25e7b1cb5982" Sep 11 00:20:45.304738 containerd[1571]: time="2025-09-11T00:20:45.304666295Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:45.307822 containerd[1571]: time="2025-09-11T00:20:45.307767904Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:45.309080 containerd[1571]: time="2025-09-11T00:20:45.308808355Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 11.427805275s" Sep 11 00:20:45.309080 containerd[1571]: time="2025-09-11T00:20:45.308847300Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Sep 11 00:20:45.323390 containerd[1571]: time="2025-09-11T00:20:45.323335328Z" level=info msg="CreateContainer within sandbox \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 11 00:20:45.343451 containerd[1571]: time="2025-09-11T00:20:45.343388777Z" level=info msg="Container 0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:45.358135 containerd[1571]: time="2025-09-11T00:20:45.358059348Z" level=info msg="CreateContainer within sandbox \"5d685316fbebace9d922625b348c9ac32511230378bf1f7fb14703b22470c5a8\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829\"" Sep 11 00:20:45.358776 containerd[1571]: time="2025-09-11T00:20:45.358720586Z" level=info msg="StartContainer for \"0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829\"" Sep 11 00:20:45.363373 containerd[1571]: time="2025-09-11T00:20:45.363324319Z" level=info msg="connecting to shim 0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829" address="unix:///run/containerd/s/15910c11511e97773dc5b38149f47630b2db591ecc6a83b5332847c742ca5b5b" protocol=ttrpc version=3 Sep 11 00:20:45.386807 systemd[1]: Started cri-containerd-0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829.scope - libcontainer container 0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829. Sep 11 00:20:45.438831 containerd[1571]: time="2025-09-11T00:20:45.438778162Z" level=info msg="StartContainer for \"0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829\" returns successfully" Sep 11 00:20:45.519415 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 11 00:20:45.520282 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 11 00:20:45.730917 systemd[1]: Started sshd@9-10.0.0.82:22-10.0.0.1:42424.service - OpenSSH per-connection server daemon (10.0.0.1:42424). Sep 11 00:20:45.765279 kubelet[2783]: I0911 00:20:45.765231 2783 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkmr7\" (UniqueName: \"kubernetes.io/projected/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-kube-api-access-nkmr7\") pod \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\" (UID: \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\") " Sep 11 00:20:45.765279 kubelet[2783]: I0911 00:20:45.765274 2783 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-ca-bundle\") pod \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\" (UID: \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\") " Sep 11 00:20:45.765279 kubelet[2783]: I0911 00:20:45.765297 2783 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-backend-key-pair\") pod \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\" (UID: \"d228d31d-37f9-4d90-bbcc-389dcb1a76f8\") " Sep 11 00:20:45.766166 kubelet[2783]: I0911 00:20:45.766117 2783 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "d228d31d-37f9-4d90-bbcc-389dcb1a76f8" (UID: "d228d31d-37f9-4d90-bbcc-389dcb1a76f8"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 11 00:20:45.769733 kubelet[2783]: I0911 00:20:45.769589 2783 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "d228d31d-37f9-4d90-bbcc-389dcb1a76f8" (UID: "d228d31d-37f9-4d90-bbcc-389dcb1a76f8"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 11 00:20:45.773992 kubelet[2783]: I0911 00:20:45.773895 2783 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-kube-api-access-nkmr7" (OuterVolumeSpecName: "kube-api-access-nkmr7") pod "d228d31d-37f9-4d90-bbcc-389dcb1a76f8" (UID: "d228d31d-37f9-4d90-bbcc-389dcb1a76f8"). InnerVolumeSpecName "kube-api-access-nkmr7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 11 00:20:45.775082 containerd[1571]: time="2025-09-11T00:20:45.774317631Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6544b88998-5nqjt,Uid:90243719-01e5-455a-a7a2-8d684145bb0e,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:45.784464 systemd[1]: Removed slice kubepods-besteffort-podd228d31d_37f9_4d90_bbcc_389dcb1a76f8.slice - libcontainer container kubepods-besteffort-podd228d31d_37f9_4d90_bbcc_389dcb1a76f8.slice. Sep 11 00:20:45.820055 sshd[4035]: Accepted publickey for core from 10.0.0.1 port 42424 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:20:45.822525 sshd-session[4035]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:20:45.830612 systemd-logind[1552]: New session 10 of user core. Sep 11 00:20:45.835775 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 11 00:20:45.866668 kubelet[2783]: I0911 00:20:45.866616 2783 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nkmr7\" (UniqueName: \"kubernetes.io/projected/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-kube-api-access-nkmr7\") on node \"localhost\" DevicePath \"\"" Sep 11 00:20:45.866668 kubelet[2783]: I0911 00:20:45.866656 2783 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Sep 11 00:20:45.866668 kubelet[2783]: I0911 00:20:45.866666 2783 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d228d31d-37f9-4d90-bbcc-389dcb1a76f8-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Sep 11 00:20:45.937389 kubelet[2783]: I0911 00:20:45.937298 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-75tzc" podStartSLOduration=1.723118897 podStartE2EDuration="26.937274058s" podCreationTimestamp="2025-09-11 00:20:19 +0000 UTC" firstStartedPulling="2025-09-11 00:20:20.096034035 +0000 UTC m=+20.489838464" lastFinishedPulling="2025-09-11 00:20:45.310189196 +0000 UTC m=+45.703993625" observedRunningTime="2025-09-11 00:20:45.934070742 +0000 UTC m=+46.327875172" watchObservedRunningTime="2025-09-11 00:20:45.937274058 +0000 UTC m=+46.331078487" Sep 11 00:20:45.965246 systemd[1]: run-netns-cni\x2d025a550e\x2d83b3\x2d990f\x2d124e\x2dec0b3f722906.mount: Deactivated successfully. Sep 11 00:20:45.966925 systemd[1]: var-lib-kubelet-pods-d228d31d\x2d37f9\x2d4d90\x2dbbcc\x2d389dcb1a76f8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dnkmr7.mount: Deactivated successfully. Sep 11 00:20:45.967035 systemd[1]: var-lib-kubelet-pods-d228d31d\x2d37f9\x2d4d90\x2dbbcc\x2d389dcb1a76f8-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 11 00:20:45.977394 systemd-networkd[1474]: cali3c1b10ce3f0: Link UP Sep 11 00:20:45.979077 systemd-networkd[1474]: cali3c1b10ce3f0: Gained carrier Sep 11 00:20:46.014410 containerd[1571]: 2025-09-11 00:20:45.811 [INFO][4040] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 00:20:46.014410 containerd[1571]: 2025-09-11 00:20:45.835 [INFO][4040] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0 calico-kube-controllers-6544b88998- calico-system 90243719-01e5-455a-a7a2-8d684145bb0e 852 0 2025-09-11 00:20:20 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6544b88998 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-6544b88998-5nqjt eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali3c1b10ce3f0 [] [] }} ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-" Sep 11 00:20:46.014410 containerd[1571]: 2025-09-11 00:20:45.835 [INFO][4040] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.014410 containerd[1571]: 2025-09-11 00:20:45.906 [INFO][4065] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" HandleID="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Workload="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.907 [INFO][4065] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" HandleID="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Workload="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000396d60), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-6544b88998-5nqjt", "timestamp":"2025-09-11 00:20:45.906813803 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.907 [INFO][4065] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.908 [INFO][4065] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.908 [INFO][4065] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.921 [INFO][4065] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" host="localhost" Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.932 [INFO][4065] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.940 [INFO][4065] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.942 [INFO][4065] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.944 [INFO][4065] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:46.016348 containerd[1571]: 2025-09-11 00:20:45.944 [INFO][4065] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" host="localhost" Sep 11 00:20:46.016626 containerd[1571]: 2025-09-11 00:20:45.946 [INFO][4065] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47 Sep 11 00:20:46.016626 containerd[1571]: 2025-09-11 00:20:45.951 [INFO][4065] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" host="localhost" Sep 11 00:20:46.016626 containerd[1571]: 2025-09-11 00:20:45.958 [INFO][4065] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" host="localhost" Sep 11 00:20:46.016626 containerd[1571]: 2025-09-11 00:20:45.959 [INFO][4065] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" host="localhost" Sep 11 00:20:46.016626 containerd[1571]: 2025-09-11 00:20:45.959 [INFO][4065] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:20:46.016626 containerd[1571]: 2025-09-11 00:20:45.959 [INFO][4065] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" HandleID="k8s-pod-network.e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Workload="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.016765 containerd[1571]: 2025-09-11 00:20:45.963 [INFO][4040] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0", GenerateName:"calico-kube-controllers-6544b88998-", Namespace:"calico-system", SelfLink:"", UID:"90243719-01e5-455a-a7a2-8d684145bb0e", ResourceVersion:"852", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6544b88998", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-6544b88998-5nqjt", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali3c1b10ce3f0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:46.016824 containerd[1571]: 2025-09-11 00:20:45.963 [INFO][4040] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.016824 containerd[1571]: 2025-09-11 00:20:45.963 [INFO][4040] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3c1b10ce3f0 ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.016824 containerd[1571]: 2025-09-11 00:20:45.977 [INFO][4040] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.016888 containerd[1571]: 2025-09-11 00:20:45.980 [INFO][4040] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0", GenerateName:"calico-kube-controllers-6544b88998-", Namespace:"calico-system", SelfLink:"", UID:"90243719-01e5-455a-a7a2-8d684145bb0e", ResourceVersion:"852", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6544b88998", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47", Pod:"calico-kube-controllers-6544b88998-5nqjt", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali3c1b10ce3f0", MAC:"ee:94:23:64:7d:5e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:46.016941 containerd[1571]: 2025-09-11 00:20:46.007 [INFO][4040] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" Namespace="calico-system" Pod="calico-kube-controllers-6544b88998-5nqjt" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6544b88998--5nqjt-eth0" Sep 11 00:20:46.032357 sshd[4062]: Connection closed by 10.0.0.1 port 42424 Sep 11 00:20:46.033224 systemd[1]: Created slice kubepods-besteffort-pod8b354a0d_10f9_428e_9fe2_17739608aaef.slice - libcontainer container kubepods-besteffort-pod8b354a0d_10f9_428e_9fe2_17739608aaef.slice. Sep 11 00:20:46.034561 sshd-session[4035]: pam_unix(sshd:session): session closed for user core Sep 11 00:20:46.042257 systemd[1]: sshd@9-10.0.0.82:22-10.0.0.1:42424.service: Deactivated successfully. Sep 11 00:20:46.045609 systemd[1]: session-10.scope: Deactivated successfully. Sep 11 00:20:46.047464 systemd-logind[1552]: Session 10 logged out. Waiting for processes to exit. Sep 11 00:20:46.050209 systemd-logind[1552]: Removed session 10. Sep 11 00:20:46.119311 containerd[1571]: time="2025-09-11T00:20:46.119229596Z" level=info msg="connecting to shim e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47" address="unix:///run/containerd/s/df7ad671cf1334eb289af1e30b33b00c0f14e239428b0814ceb3c6709ad017cf" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:46.156923 systemd[1]: Started cri-containerd-e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47.scope - libcontainer container e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47. Sep 11 00:20:46.170735 kubelet[2783]: I0911 00:20:46.170675 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbnm\" (UniqueName: \"kubernetes.io/projected/8b354a0d-10f9-428e-9fe2-17739608aaef-kube-api-access-4fbnm\") pod \"whisker-78dbcbf6cb-5w25q\" (UID: \"8b354a0d-10f9-428e-9fe2-17739608aaef\") " pod="calico-system/whisker-78dbcbf6cb-5w25q" Sep 11 00:20:46.171063 kubelet[2783]: I0911 00:20:46.170976 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b354a0d-10f9-428e-9fe2-17739608aaef-whisker-ca-bundle\") pod \"whisker-78dbcbf6cb-5w25q\" (UID: \"8b354a0d-10f9-428e-9fe2-17739608aaef\") " pod="calico-system/whisker-78dbcbf6cb-5w25q" Sep 11 00:20:46.171063 kubelet[2783]: I0911 00:20:46.171016 2783 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/8b354a0d-10f9-428e-9fe2-17739608aaef-whisker-backend-key-pair\") pod \"whisker-78dbcbf6cb-5w25q\" (UID: \"8b354a0d-10f9-428e-9fe2-17739608aaef\") " pod="calico-system/whisker-78dbcbf6cb-5w25q" Sep 11 00:20:46.172523 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:20:46.215658 containerd[1571]: time="2025-09-11T00:20:46.215610744Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6544b88998-5nqjt,Uid:90243719-01e5-455a-a7a2-8d684145bb0e,Namespace:calico-system,Attempt:0,} returns sandbox id \"e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47\"" Sep 11 00:20:46.221903 containerd[1571]: time="2025-09-11T00:20:46.221863596Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 11 00:20:46.339579 containerd[1571]: time="2025-09-11T00:20:46.339449366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-78dbcbf6cb-5w25q,Uid:8b354a0d-10f9-428e-9fe2-17739608aaef,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:46.442714 systemd-networkd[1474]: cali94d3f2344c3: Link UP Sep 11 00:20:46.443357 systemd-networkd[1474]: cali94d3f2344c3: Gained carrier Sep 11 00:20:46.457662 containerd[1571]: 2025-09-11 00:20:46.365 [INFO][4143] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 00:20:46.457662 containerd[1571]: 2025-09-11 00:20:46.376 [INFO][4143] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0 whisker-78dbcbf6cb- calico-system 8b354a0d-10f9-428e-9fe2-17739608aaef 967 0 2025-09-11 00:20:45 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:78dbcbf6cb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-78dbcbf6cb-5w25q eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali94d3f2344c3 [] [] }} ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-" Sep 11 00:20:46.457662 containerd[1571]: 2025-09-11 00:20:46.377 [INFO][4143] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.457662 containerd[1571]: 2025-09-11 00:20:46.405 [INFO][4157] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" HandleID="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Workload="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.405 [INFO][4157] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" HandleID="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Workload="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f7b0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-78dbcbf6cb-5w25q", "timestamp":"2025-09-11 00:20:46.405162717 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.405 [INFO][4157] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.405 [INFO][4157] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.405 [INFO][4157] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.412 [INFO][4157] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" host="localhost" Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.418 [INFO][4157] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.422 [INFO][4157] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.423 [INFO][4157] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.426 [INFO][4157] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:46.457969 containerd[1571]: 2025-09-11 00:20:46.426 [INFO][4157] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" host="localhost" Sep 11 00:20:46.458281 containerd[1571]: 2025-09-11 00:20:46.427 [INFO][4157] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d Sep 11 00:20:46.458281 containerd[1571]: 2025-09-11 00:20:46.431 [INFO][4157] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" host="localhost" Sep 11 00:20:46.458281 containerd[1571]: 2025-09-11 00:20:46.436 [INFO][4157] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" host="localhost" Sep 11 00:20:46.458281 containerd[1571]: 2025-09-11 00:20:46.437 [INFO][4157] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" host="localhost" Sep 11 00:20:46.458281 containerd[1571]: 2025-09-11 00:20:46.437 [INFO][4157] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:20:46.458281 containerd[1571]: 2025-09-11 00:20:46.437 [INFO][4157] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" HandleID="k8s-pod-network.c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Workload="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.458461 containerd[1571]: 2025-09-11 00:20:46.440 [INFO][4143] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0", GenerateName:"whisker-78dbcbf6cb-", Namespace:"calico-system", SelfLink:"", UID:"8b354a0d-10f9-428e-9fe2-17739608aaef", ResourceVersion:"967", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"78dbcbf6cb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-78dbcbf6cb-5w25q", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali94d3f2344c3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:46.458461 containerd[1571]: 2025-09-11 00:20:46.440 [INFO][4143] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.458563 containerd[1571]: 2025-09-11 00:20:46.440 [INFO][4143] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali94d3f2344c3 ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.458563 containerd[1571]: 2025-09-11 00:20:46.443 [INFO][4143] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.458619 containerd[1571]: 2025-09-11 00:20:46.443 [INFO][4143] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0", GenerateName:"whisker-78dbcbf6cb-", Namespace:"calico-system", SelfLink:"", UID:"8b354a0d-10f9-428e-9fe2-17739608aaef", ResourceVersion:"967", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"78dbcbf6cb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d", Pod:"whisker-78dbcbf6cb-5w25q", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali94d3f2344c3", MAC:"82:2f:2a:b0:b4:aa", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:46.458669 containerd[1571]: 2025-09-11 00:20:46.453 [INFO][4143] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" Namespace="calico-system" Pod="whisker-78dbcbf6cb-5w25q" WorkloadEndpoint="localhost-k8s-whisker--78dbcbf6cb--5w25q-eth0" Sep 11 00:20:46.497348 containerd[1571]: time="2025-09-11T00:20:46.497287772Z" level=info msg="connecting to shim c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d" address="unix:///run/containerd/s/b5acd7ab6b4a25f27ad2ae2c1700d3565f70cd9730f8d581f01526cb3281548a" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:46.534862 systemd[1]: Started cri-containerd-c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d.scope - libcontainer container c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d. Sep 11 00:20:46.548903 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:20:46.581499 containerd[1571]: time="2025-09-11T00:20:46.581441865Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-78dbcbf6cb-5w25q,Uid:8b354a0d-10f9-428e-9fe2-17739608aaef,Namespace:calico-system,Attempt:0,} returns sandbox id \"c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d\"" Sep 11 00:20:46.771320 containerd[1571]: time="2025-09-11T00:20:46.771267406Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-pfr6f,Uid:78b30ee7-7aa4-4eb9-858b-ae86f51e74f5,Namespace:calico-apiserver,Attempt:0,}" Sep 11 00:20:47.148656 systemd-networkd[1474]: cali6aa75c30f8a: Link UP Sep 11 00:20:47.150325 systemd-networkd[1474]: cali6aa75c30f8a: Gained carrier Sep 11 00:20:47.317061 containerd[1571]: 2025-09-11 00:20:46.955 [INFO][4296] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 11 00:20:47.317061 containerd[1571]: 2025-09-11 00:20:46.982 [INFO][4296] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0 calico-apiserver-7bd7c7b5ff- calico-apiserver 78b30ee7-7aa4-4eb9-858b-ae86f51e74f5 849 0 2025-09-11 00:20:16 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7bd7c7b5ff projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7bd7c7b5ff-pfr6f eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali6aa75c30f8a [] [] }} ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-" Sep 11 00:20:47.317061 containerd[1571]: 2025-09-11 00:20:46.983 [INFO][4296] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.317061 containerd[1571]: 2025-09-11 00:20:47.055 [INFO][4326] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" HandleID="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Workload="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.055 [INFO][4326] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" HandleID="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Workload="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a4e70), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7bd7c7b5ff-pfr6f", "timestamp":"2025-09-11 00:20:47.0555267 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.055 [INFO][4326] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.055 [INFO][4326] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.055 [INFO][4326] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.064 [INFO][4326] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" host="localhost" Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.070 [INFO][4326] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.109 [INFO][4326] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.115 [INFO][4326] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.119 [INFO][4326] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:47.317665 containerd[1571]: 2025-09-11 00:20:47.120 [INFO][4326] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" host="localhost" Sep 11 00:20:47.318002 containerd[1571]: 2025-09-11 00:20:47.122 [INFO][4326] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f Sep 11 00:20:47.318002 containerd[1571]: 2025-09-11 00:20:47.128 [INFO][4326] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" host="localhost" Sep 11 00:20:47.318002 containerd[1571]: 2025-09-11 00:20:47.140 [INFO][4326] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" host="localhost" Sep 11 00:20:47.318002 containerd[1571]: 2025-09-11 00:20:47.140 [INFO][4326] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" host="localhost" Sep 11 00:20:47.318002 containerd[1571]: 2025-09-11 00:20:47.140 [INFO][4326] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:20:47.318002 containerd[1571]: 2025-09-11 00:20:47.140 [INFO][4326] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" HandleID="k8s-pod-network.e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Workload="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.318137 containerd[1571]: 2025-09-11 00:20:47.145 [INFO][4296] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0", GenerateName:"calico-apiserver-7bd7c7b5ff-", Namespace:"calico-apiserver", SelfLink:"", UID:"78b30ee7-7aa4-4eb9-858b-ae86f51e74f5", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7bd7c7b5ff", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7bd7c7b5ff-pfr6f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6aa75c30f8a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:47.318195 containerd[1571]: 2025-09-11 00:20:47.146 [INFO][4296] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.318195 containerd[1571]: 2025-09-11 00:20:47.146 [INFO][4296] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6aa75c30f8a ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.318195 containerd[1571]: 2025-09-11 00:20:47.148 [INFO][4296] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.318276 containerd[1571]: 2025-09-11 00:20:47.148 [INFO][4296] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0", GenerateName:"calico-apiserver-7bd7c7b5ff-", Namespace:"calico-apiserver", SelfLink:"", UID:"78b30ee7-7aa4-4eb9-858b-ae86f51e74f5", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7bd7c7b5ff", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f", Pod:"calico-apiserver-7bd7c7b5ff-pfr6f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6aa75c30f8a", MAC:"2e:6c:95:56:df:b2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:47.318347 containerd[1571]: 2025-09-11 00:20:47.311 [INFO][4296] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-pfr6f" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--pfr6f-eth0" Sep 11 00:20:47.364633 containerd[1571]: time="2025-09-11T00:20:47.364530340Z" level=info msg="connecting to shim e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f" address="unix:///run/containerd/s/d5ff460140417960c7f292935d2cc4f1f38e5e9675279cf118966487ff1cb164" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:47.402881 systemd[1]: Started cri-containerd-e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f.scope - libcontainer container e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f. Sep 11 00:20:47.423381 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:20:47.486966 containerd[1571]: time="2025-09-11T00:20:47.486909321Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-pfr6f,Uid:78b30ee7-7aa4-4eb9-858b-ae86f51e74f5,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f\"" Sep 11 00:20:47.573883 systemd-networkd[1474]: cali3c1b10ce3f0: Gained IPv6LL Sep 11 00:20:47.677268 systemd-networkd[1474]: vxlan.calico: Link UP Sep 11 00:20:47.677283 systemd-networkd[1474]: vxlan.calico: Gained carrier Sep 11 00:20:47.774459 kubelet[2783]: I0911 00:20:47.774405 2783 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d228d31d-37f9-4d90-bbcc-389dcb1a76f8" path="/var/lib/kubelet/pods/d228d31d-37f9-4d90-bbcc-389dcb1a76f8/volumes" Sep 11 00:20:47.779899 kubelet[2783]: E0911 00:20:47.779584 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:47.780148 containerd[1571]: time="2025-09-11T00:20:47.780108587Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2h255,Uid:2f2a0328-8013-43e7-aa22-1250dbd184a5,Namespace:kube-system,Attempt:0,}" Sep 11 00:20:47.829983 systemd-networkd[1474]: cali94d3f2344c3: Gained IPv6LL Sep 11 00:20:48.112518 systemd-networkd[1474]: cali62752163036: Link UP Sep 11 00:20:48.114166 systemd-networkd[1474]: cali62752163036: Gained carrier Sep 11 00:20:48.168017 containerd[1571]: 2025-09-11 00:20:47.972 [INFO][4458] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--2h255-eth0 coredns-674b8bbfcf- kube-system 2f2a0328-8013-43e7-aa22-1250dbd184a5 848 0 2025-09-11 00:20:05 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-2h255 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali62752163036 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-" Sep 11 00:20:48.168017 containerd[1571]: 2025-09-11 00:20:47.972 [INFO][4458] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.168017 containerd[1571]: 2025-09-11 00:20:48.026 [INFO][4472] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" HandleID="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Workload="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.026 [INFO][4472] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" HandleID="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Workload="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000504980), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-2h255", "timestamp":"2025-09-11 00:20:48.026209846 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.027 [INFO][4472] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.027 [INFO][4472] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.027 [INFO][4472] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.039 [INFO][4472] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" host="localhost" Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.049 [INFO][4472] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.057 [INFO][4472] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.063 [INFO][4472] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.067 [INFO][4472] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:48.168457 containerd[1571]: 2025-09-11 00:20:48.067 [INFO][4472] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" host="localhost" Sep 11 00:20:48.168897 containerd[1571]: 2025-09-11 00:20:48.072 [INFO][4472] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2 Sep 11 00:20:48.168897 containerd[1571]: 2025-09-11 00:20:48.088 [INFO][4472] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" host="localhost" Sep 11 00:20:48.168897 containerd[1571]: 2025-09-11 00:20:48.100 [INFO][4472] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" host="localhost" Sep 11 00:20:48.168897 containerd[1571]: 2025-09-11 00:20:48.101 [INFO][4472] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" host="localhost" Sep 11 00:20:48.168897 containerd[1571]: 2025-09-11 00:20:48.101 [INFO][4472] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:20:48.168897 containerd[1571]: 2025-09-11 00:20:48.101 [INFO][4472] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" HandleID="k8s-pod-network.dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Workload="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.169102 containerd[1571]: 2025-09-11 00:20:48.107 [INFO][4458] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--2h255-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"2f2a0328-8013-43e7-aa22-1250dbd184a5", ResourceVersion:"848", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-2h255", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali62752163036", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:48.169214 containerd[1571]: 2025-09-11 00:20:48.107 [INFO][4458] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.169214 containerd[1571]: 2025-09-11 00:20:48.107 [INFO][4458] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali62752163036 ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.169214 containerd[1571]: 2025-09-11 00:20:48.115 [INFO][4458] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.169312 containerd[1571]: 2025-09-11 00:20:48.115 [INFO][4458] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--2h255-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"2f2a0328-8013-43e7-aa22-1250dbd184a5", ResourceVersion:"848", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2", Pod:"coredns-674b8bbfcf-2h255", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali62752163036", MAC:"c6:fd:9f:48:03:9c", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:48.169312 containerd[1571]: 2025-09-11 00:20:48.159 [INFO][4458] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" Namespace="kube-system" Pod="coredns-674b8bbfcf-2h255" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--2h255-eth0" Sep 11 00:20:48.216387 containerd[1571]: time="2025-09-11T00:20:48.216315329Z" level=info msg="connecting to shim dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2" address="unix:///run/containerd/s/4ea45168cc2f99f1ef31f3c292e255b42ee20cb01eeebf46a33204f62f744747" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:48.262167 systemd[1]: Started cri-containerd-dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2.scope - libcontainer container dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2. Sep 11 00:20:48.279939 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:20:48.314566 containerd[1571]: time="2025-09-11T00:20:48.314503799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-2h255,Uid:2f2a0328-8013-43e7-aa22-1250dbd184a5,Namespace:kube-system,Attempt:0,} returns sandbox id \"dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2\"" Sep 11 00:20:48.315479 kubelet[2783]: E0911 00:20:48.315438 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:48.322830 containerd[1571]: time="2025-09-11T00:20:48.322695642Z" level=info msg="CreateContainer within sandbox \"dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 11 00:20:48.345222 containerd[1571]: time="2025-09-11T00:20:48.345115674Z" level=info msg="Container bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:48.350540 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2884786640.mount: Deactivated successfully. Sep 11 00:20:48.358057 containerd[1571]: time="2025-09-11T00:20:48.357977697Z" level=info msg="CreateContainer within sandbox \"dfcf659501a673c97e60374d64ab5af8f326b2044cc52b96f704b8665336aff2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4\"" Sep 11 00:20:48.359173 containerd[1571]: time="2025-09-11T00:20:48.359080734Z" level=info msg="StartContainer for \"bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4\"" Sep 11 00:20:48.360439 containerd[1571]: time="2025-09-11T00:20:48.360388948Z" level=info msg="connecting to shim bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4" address="unix:///run/containerd/s/4ea45168cc2f99f1ef31f3c292e255b42ee20cb01eeebf46a33204f62f744747" protocol=ttrpc version=3 Sep 11 00:20:48.421924 systemd[1]: Started cri-containerd-bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4.scope - libcontainer container bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4. Sep 11 00:20:48.474242 containerd[1571]: time="2025-09-11T00:20:48.474197472Z" level=info msg="StartContainer for \"bea70acc3d491978b4fbe8f19aba7ac2b81587612e662500b6bf25fd5a7b59a4\" returns successfully" Sep 11 00:20:48.773626 containerd[1571]: time="2025-09-11T00:20:48.772299426Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-k7tqf,Uid:abb44784-55a4-4f8a-ac18-b17c33a6fc96,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:48.790691 systemd-networkd[1474]: vxlan.calico: Gained IPv6LL Sep 11 00:20:48.934058 kubelet[2783]: E0911 00:20:48.932802 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:48.949850 systemd-networkd[1474]: cali54aaf2998cc: Link UP Sep 11 00:20:48.951607 systemd-networkd[1474]: cali54aaf2998cc: Gained carrier Sep 11 00:20:48.979455 kubelet[2783]: I0911 00:20:48.978781 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-2h255" podStartSLOduration=43.978751502 podStartE2EDuration="43.978751502s" podCreationTimestamp="2025-09-11 00:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:20:48.957895789 +0000 UTC m=+49.351700228" watchObservedRunningTime="2025-09-11 00:20:48.978751502 +0000 UTC m=+49.372555931" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.825 [INFO][4605] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--54d579b49d--k7tqf-eth0 goldmane-54d579b49d- calico-system abb44784-55a4-4f8a-ac18-b17c33a6fc96 853 0 2025-09-11 00:20:19 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-54d579b49d-k7tqf eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali54aaf2998cc [] [] }} ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.826 [INFO][4605] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.879 [INFO][4621] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" HandleID="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Workload="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.880 [INFO][4621] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" HandleID="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Workload="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000356cb0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-54d579b49d-k7tqf", "timestamp":"2025-09-11 00:20:48.879023422 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.880 [INFO][4621] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.880 [INFO][4621] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.880 [INFO][4621] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.897 [INFO][4621] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.903 [INFO][4621] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.908 [INFO][4621] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.910 [INFO][4621] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.912 [INFO][4621] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.913 [INFO][4621] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.914 [INFO][4621] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3 Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.919 [INFO][4621] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.929 [INFO][4621] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.929 [INFO][4621] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" host="localhost" Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.929 [INFO][4621] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:20:48.982229 containerd[1571]: 2025-09-11 00:20:48.929 [INFO][4621] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" HandleID="k8s-pod-network.8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Workload="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:48.984049 containerd[1571]: 2025-09-11 00:20:48.935 [INFO][4605] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--k7tqf-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"abb44784-55a4-4f8a-ac18-b17c33a6fc96", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-54d579b49d-k7tqf", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali54aaf2998cc", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:48.984049 containerd[1571]: 2025-09-11 00:20:48.935 [INFO][4605] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:48.984049 containerd[1571]: 2025-09-11 00:20:48.935 [INFO][4605] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali54aaf2998cc ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:48.984049 containerd[1571]: 2025-09-11 00:20:48.960 [INFO][4605] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:48.984049 containerd[1571]: 2025-09-11 00:20:48.961 [INFO][4605] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--k7tqf-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"abb44784-55a4-4f8a-ac18-b17c33a6fc96", ResourceVersion:"853", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3", Pod:"goldmane-54d579b49d-k7tqf", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali54aaf2998cc", MAC:"aa:15:64:53:80:c3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:48.984049 containerd[1571]: 2025-09-11 00:20:48.972 [INFO][4605] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" Namespace="calico-system" Pod="goldmane-54d579b49d-k7tqf" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--k7tqf-eth0" Sep 11 00:20:49.061256 containerd[1571]: time="2025-09-11T00:20:49.060880496Z" level=info msg="connecting to shim 8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3" address="unix:///run/containerd/s/dacb55ee252281b9eaff421d5ff23fd66fa6253e0a9c00219b3156ee99eecfea" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:49.098728 systemd[1]: Started cri-containerd-8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3.scope - libcontainer container 8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3. Sep 11 00:20:49.109935 systemd-networkd[1474]: cali6aa75c30f8a: Gained IPv6LL Sep 11 00:20:49.119443 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:20:49.255534 containerd[1571]: time="2025-09-11T00:20:49.255441359Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-k7tqf,Uid:abb44784-55a4-4f8a-ac18-b17c33a6fc96,Namespace:calico-system,Attempt:0,} returns sandbox id \"8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3\"" Sep 11 00:20:49.591132 containerd[1571]: time="2025-09-11T00:20:49.591052315Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:49.592013 containerd[1571]: time="2025-09-11T00:20:49.591973911Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Sep 11 00:20:49.593175 containerd[1571]: time="2025-09-11T00:20:49.593135610Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:49.595731 containerd[1571]: time="2025-09-11T00:20:49.595681118Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:49.596402 containerd[1571]: time="2025-09-11T00:20:49.596373402Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 3.374473265s" Sep 11 00:20:49.596464 containerd[1571]: time="2025-09-11T00:20:49.596406876Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Sep 11 00:20:49.597312 containerd[1571]: time="2025-09-11T00:20:49.597285089Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 11 00:20:49.611607 containerd[1571]: time="2025-09-11T00:20:49.611539819Z" level=info msg="CreateContainer within sandbox \"e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 11 00:20:49.623092 containerd[1571]: time="2025-09-11T00:20:49.623033785Z" level=info msg="Container aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:49.632938 containerd[1571]: time="2025-09-11T00:20:49.632878603Z" level=info msg="CreateContainer within sandbox \"e758935ac234ea9b58610f447148e52a1aff955f5e10da182a611c8c6de63d47\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\"" Sep 11 00:20:49.633926 containerd[1571]: time="2025-09-11T00:20:49.633886726Z" level=info msg="StartContainer for \"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\"" Sep 11 00:20:49.635527 containerd[1571]: time="2025-09-11T00:20:49.635475909Z" level=info msg="connecting to shim aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246" address="unix:///run/containerd/s/df7ad671cf1334eb289af1e30b33b00c0f14e239428b0814ceb3c6709ad017cf" protocol=ttrpc version=3 Sep 11 00:20:49.662870 systemd[1]: Started cri-containerd-aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246.scope - libcontainer container aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246. Sep 11 00:20:49.941813 systemd-networkd[1474]: cali62752163036: Gained IPv6LL Sep 11 00:20:50.047879 containerd[1571]: time="2025-09-11T00:20:50.047744982Z" level=info msg="StartContainer for \"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\" returns successfully" Sep 11 00:20:50.051146 kubelet[2783]: E0911 00:20:50.051112 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:50.966839 systemd-networkd[1474]: cali54aaf2998cc: Gained IPv6LL Sep 11 00:20:51.050446 systemd[1]: Started sshd@10-10.0.0.82:22-10.0.0.1:58468.service - OpenSSH per-connection server daemon (10.0.0.1:58468). Sep 11 00:20:51.053590 kubelet[2783]: E0911 00:20:51.053520 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:20:51.155864 containerd[1571]: time="2025-09-11T00:20:51.155794048Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\" id:\"ff79fdeb4b9731267e0f84058a82fcc9c5110757fb3cc13d406516fe83502067\" pid:4755 exited_at:{seconds:1757550051 nanos:155142684}" Sep 11 00:20:51.414439 sshd[4740]: Accepted publickey for core from 10.0.0.1 port 58468 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:20:51.416448 sshd-session[4740]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:20:51.422238 systemd-logind[1552]: New session 11 of user core. Sep 11 00:20:51.435746 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 11 00:20:51.548985 kubelet[2783]: I0911 00:20:51.548887 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6544b88998-5nqjt" podStartSLOduration=28.173325555 podStartE2EDuration="31.548833336s" podCreationTimestamp="2025-09-11 00:20:20 +0000 UTC" firstStartedPulling="2025-09-11 00:20:46.221626267 +0000 UTC m=+46.615430696" lastFinishedPulling="2025-09-11 00:20:49.597134038 +0000 UTC m=+49.990938477" observedRunningTime="2025-09-11 00:20:51.483723004 +0000 UTC m=+51.877527433" watchObservedRunningTime="2025-09-11 00:20:51.548833336 +0000 UTC m=+51.942637765" Sep 11 00:20:51.760648 sshd[4767]: Connection closed by 10.0.0.1 port 58468 Sep 11 00:20:51.761036 sshd-session[4740]: pam_unix(sshd:session): session closed for user core Sep 11 00:20:51.765010 systemd[1]: sshd@10-10.0.0.82:22-10.0.0.1:58468.service: Deactivated successfully. Sep 11 00:20:51.767257 systemd[1]: session-11.scope: Deactivated successfully. Sep 11 00:20:51.769041 systemd-logind[1552]: Session 11 logged out. Waiting for processes to exit. Sep 11 00:20:51.770520 systemd-logind[1552]: Removed session 11. Sep 11 00:20:53.931894 containerd[1571]: time="2025-09-11T00:20:53.931804986Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:53.933101 containerd[1571]: time="2025-09-11T00:20:53.933054758Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Sep 11 00:20:53.936495 containerd[1571]: time="2025-09-11T00:20:53.936436609Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:53.945229 containerd[1571]: time="2025-09-11T00:20:53.945144097Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:53.945955 containerd[1571]: time="2025-09-11T00:20:53.945901384Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 4.348587198s" Sep 11 00:20:53.945955 containerd[1571]: time="2025-09-11T00:20:53.945942492Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Sep 11 00:20:53.948036 containerd[1571]: time="2025-09-11T00:20:53.947092433Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 11 00:20:53.953746 containerd[1571]: time="2025-09-11T00:20:53.953692249Z" level=info msg="CreateContainer within sandbox \"c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 11 00:20:53.967326 containerd[1571]: time="2025-09-11T00:20:53.967246786Z" level=info msg="Container 2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:53.978619 containerd[1571]: time="2025-09-11T00:20:53.978538872Z" level=info msg="CreateContainer within sandbox \"c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde\"" Sep 11 00:20:53.979285 containerd[1571]: time="2025-09-11T00:20:53.979235913Z" level=info msg="StartContainer for \"2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde\"" Sep 11 00:20:53.980449 containerd[1571]: time="2025-09-11T00:20:53.980410441Z" level=info msg="connecting to shim 2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde" address="unix:///run/containerd/s/b5acd7ab6b4a25f27ad2ae2c1700d3565f70cd9730f8d581f01526cb3281548a" protocol=ttrpc version=3 Sep 11 00:20:54.009779 systemd[1]: Started cri-containerd-2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde.scope - libcontainer container 2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde. Sep 11 00:20:54.099503 containerd[1571]: time="2025-09-11T00:20:54.099449057Z" level=info msg="StartContainer for \"2cbb6ba6969e0a5675a7751579b2474c57bc39778a3933649229e1c0bfecfcde\" returns successfully" Sep 11 00:20:56.776830 systemd[1]: Started sshd@11-10.0.0.82:22-10.0.0.1:58484.service - OpenSSH per-connection server daemon (10.0.0.1:58484). Sep 11 00:20:56.870696 sshd[4832]: Accepted publickey for core from 10.0.0.1 port 58484 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:20:56.872498 sshd-session[4832]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:20:56.878866 systemd-logind[1552]: New session 12 of user core. Sep 11 00:20:56.887774 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 11 00:20:57.060443 sshd[4835]: Connection closed by 10.0.0.1 port 58484 Sep 11 00:20:57.061084 sshd-session[4832]: pam_unix(sshd:session): session closed for user core Sep 11 00:20:57.067073 systemd[1]: sshd@11-10.0.0.82:22-10.0.0.1:58484.service: Deactivated successfully. Sep 11 00:20:57.070580 systemd[1]: session-12.scope: Deactivated successfully. Sep 11 00:20:57.072312 systemd-logind[1552]: Session 12 logged out. Waiting for processes to exit. Sep 11 00:20:57.075236 systemd-logind[1552]: Removed session 12. Sep 11 00:20:57.771875 containerd[1571]: time="2025-09-11T00:20:57.771802790Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-f8r22,Uid:71ecab39-5d3e-4469-82ca-831c45fb7dcd,Namespace:calico-system,Attempt:0,}" Sep 11 00:20:58.250720 systemd-networkd[1474]: calif4152bd9621: Link UP Sep 11 00:20:58.250951 systemd-networkd[1474]: calif4152bd9621: Gained carrier Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.951 [INFO][4849] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--f8r22-eth0 csi-node-driver- calico-system 71ecab39-5d3e-4469-82ca-831c45fb7dcd 726 0 2025-09-11 00:20:19 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-f8r22 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calif4152bd9621 [] [] }} ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.951 [INFO][4849] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.982 [INFO][4863] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" HandleID="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Workload="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.983 [INFO][4863] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" HandleID="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Workload="localhost-k8s-csi--node--driver--f8r22-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000515500), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-f8r22", "timestamp":"2025-09-11 00:20:57.982836983 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.983 [INFO][4863] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.983 [INFO][4863] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.983 [INFO][4863] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.990 [INFO][4863] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:57.996 [INFO][4863] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.001 [INFO][4863] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.003 [INFO][4863] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.005 [INFO][4863] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.005 [INFO][4863] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.007 [INFO][4863] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455 Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.043 [INFO][4863] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.244 [INFO][4863] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.244 [INFO][4863] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" host="localhost" Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.245 [INFO][4863] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:20:58.400202 containerd[1571]: 2025-09-11 00:20:58.245 [INFO][4863] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" HandleID="k8s-pod-network.db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Workload="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.401026 containerd[1571]: 2025-09-11 00:20:58.248 [INFO][4849] cni-plugin/k8s.go 418: Populated endpoint ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--f8r22-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"71ecab39-5d3e-4469-82ca-831c45fb7dcd", ResourceVersion:"726", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-f8r22", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif4152bd9621", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:58.401026 containerd[1571]: 2025-09-11 00:20:58.248 [INFO][4849] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.401026 containerd[1571]: 2025-09-11 00:20:58.248 [INFO][4849] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif4152bd9621 ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.401026 containerd[1571]: 2025-09-11 00:20:58.251 [INFO][4849] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.401026 containerd[1571]: 2025-09-11 00:20:58.251 [INFO][4849] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--f8r22-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"71ecab39-5d3e-4469-82ca-831c45fb7dcd", ResourceVersion:"726", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455", Pod:"csi-node-driver-f8r22", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif4152bd9621", MAC:"d2:6f:7a:e4:ea:33", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:20:58.401026 containerd[1571]: 2025-09-11 00:20:58.396 [INFO][4849] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" Namespace="calico-system" Pod="csi-node-driver-f8r22" WorkloadEndpoint="localhost-k8s-csi--node--driver--f8r22-eth0" Sep 11 00:20:58.572097 containerd[1571]: time="2025-09-11T00:20:58.571883353Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:58.629693 containerd[1571]: time="2025-09-11T00:20:58.629610576Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Sep 11 00:20:58.655787 containerd[1571]: time="2025-09-11T00:20:58.655682112Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:58.798947 containerd[1571]: time="2025-09-11T00:20:58.798852974Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:20:58.799834 containerd[1571]: time="2025-09-11T00:20:58.799774320Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 4.852648674s" Sep 11 00:20:58.799884 containerd[1571]: time="2025-09-11T00:20:58.799832742Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 11 00:20:58.810422 containerd[1571]: time="2025-09-11T00:20:58.810367382Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 11 00:20:58.997625 containerd[1571]: time="2025-09-11T00:20:58.997445171Z" level=info msg="connecting to shim db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455" address="unix:///run/containerd/s/f926348bf826c3b83b2327da4f80fa584801ec988cec151c5d771797ebe48bfa" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:20:59.037071 containerd[1571]: time="2025-09-11T00:20:59.037007483Z" level=info msg="CreateContainer within sandbox \"e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 11 00:20:59.063808 systemd[1]: Started cri-containerd-db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455.scope - libcontainer container db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455. Sep 11 00:20:59.077122 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:20:59.259007 containerd[1571]: time="2025-09-11T00:20:59.258944283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-f8r22,Uid:71ecab39-5d3e-4469-82ca-831c45fb7dcd,Namespace:calico-system,Attempt:0,} returns sandbox id \"db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455\"" Sep 11 00:20:59.349733 systemd-networkd[1474]: calif4152bd9621: Gained IPv6LL Sep 11 00:20:59.372376 containerd[1571]: time="2025-09-11T00:20:59.372306880Z" level=info msg="Container 270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:20:59.575279 containerd[1571]: time="2025-09-11T00:20:59.575123227Z" level=info msg="CreateContainer within sandbox \"e5b963f1a9192a540a2a9ba23931a4666983863211a7207b956646477884819f\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad\"" Sep 11 00:20:59.575851 containerd[1571]: time="2025-09-11T00:20:59.575826434Z" level=info msg="StartContainer for \"270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad\"" Sep 11 00:20:59.577043 containerd[1571]: time="2025-09-11T00:20:59.576953833Z" level=info msg="connecting to shim 270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad" address="unix:///run/containerd/s/d5ff460140417960c7f292935d2cc4f1f38e5e9675279cf118966487ff1cb164" protocol=ttrpc version=3 Sep 11 00:20:59.605875 systemd[1]: Started cri-containerd-270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad.scope - libcontainer container 270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad. Sep 11 00:20:59.773944 containerd[1571]: time="2025-09-11T00:20:59.773878425Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-rr4tc,Uid:25061c9c-4e54-43e9-a333-25e7b1cb5982,Namespace:calico-apiserver,Attempt:0,}" Sep 11 00:21:00.337429 containerd[1571]: time="2025-09-11T00:21:00.337374145Z" level=info msg="StartContainer for \"270a41bf8a3923dd4f57ff2b3cf760b63e64db469f99a44cb40b8548180d79ad\" returns successfully" Sep 11 00:21:00.771473 kubelet[2783]: E0911 00:21:00.770832 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:00.772050 containerd[1571]: time="2025-09-11T00:21:00.771321447Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-bcvvs,Uid:f4246163-a863-4efc-b1a5-f23f0e9c31b9,Namespace:kube-system,Attempt:0,}" Sep 11 00:21:01.836944 systemd-networkd[1474]: cali16c6143322c: Link UP Sep 11 00:21:01.837568 systemd-networkd[1474]: cali16c6143322c: Gained carrier Sep 11 00:21:01.916599 kubelet[2783]: I0911 00:21:01.916448 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-pfr6f" podStartSLOduration=34.595804026 podStartE2EDuration="45.916421016s" podCreationTimestamp="2025-09-11 00:20:16 +0000 UTC" firstStartedPulling="2025-09-11 00:20:47.489432763 +0000 UTC m=+47.883237192" lastFinishedPulling="2025-09-11 00:20:58.810049743 +0000 UTC m=+59.203854182" observedRunningTime="2025-09-11 00:21:01.916064294 +0000 UTC m=+62.309868743" watchObservedRunningTime="2025-09-11 00:21:01.916421016 +0000 UTC m=+62.310225445" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.826 [INFO][4982] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0 calico-apiserver-7bd7c7b5ff- calico-apiserver 25061c9c-4e54-43e9-a333-25e7b1cb5982 850 0 2025-09-11 00:20:16 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7bd7c7b5ff projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7bd7c7b5ff-rr4tc eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali16c6143322c [] [] }} ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.826 [INFO][4982] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.887 [INFO][4998] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" HandleID="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Workload="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.888 [INFO][4998] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" HandleID="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Workload="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00026d630), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7bd7c7b5ff-rr4tc", "timestamp":"2025-09-11 00:21:00.887899453 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.888 [INFO][4998] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.888 [INFO][4998] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.888 [INFO][4998] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:00.902 [INFO][4998] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.167 [INFO][4998] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.172 [INFO][4998] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.194 [INFO][4998] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.197 [INFO][4998] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.197 [INFO][4998] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.257 [INFO][4998] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.582 [INFO][4998] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.818 [INFO][4998] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.818 [INFO][4998] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" host="localhost" Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.818 [INFO][4998] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:21:01.920751 containerd[1571]: 2025-09-11 00:21:01.818 [INFO][4998] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" HandleID="k8s-pod-network.d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Workload="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:01.922052 containerd[1571]: 2025-09-11 00:21:01.822 [INFO][4982] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0", GenerateName:"calico-apiserver-7bd7c7b5ff-", Namespace:"calico-apiserver", SelfLink:"", UID:"25061c9c-4e54-43e9-a333-25e7b1cb5982", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7bd7c7b5ff", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7bd7c7b5ff-rr4tc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali16c6143322c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:21:01.922052 containerd[1571]: 2025-09-11 00:21:01.834 [INFO][4982] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:01.922052 containerd[1571]: 2025-09-11 00:21:01.835 [INFO][4982] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali16c6143322c ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:01.922052 containerd[1571]: 2025-09-11 00:21:01.837 [INFO][4982] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:01.922052 containerd[1571]: 2025-09-11 00:21:01.838 [INFO][4982] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0", GenerateName:"calico-apiserver-7bd7c7b5ff-", Namespace:"calico-apiserver", SelfLink:"", UID:"25061c9c-4e54-43e9-a333-25e7b1cb5982", ResourceVersion:"850", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7bd7c7b5ff", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a", Pod:"calico-apiserver-7bd7c7b5ff-rr4tc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali16c6143322c", MAC:"4e:28:f8:9e:e5:7c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:21:01.922052 containerd[1571]: 2025-09-11 00:21:01.915 [INFO][4982] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" Namespace="calico-apiserver" Pod="calico-apiserver-7bd7c7b5ff-rr4tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bd7c7b5ff--rr4tc-eth0" Sep 11 00:21:02.080746 systemd[1]: Started sshd@12-10.0.0.82:22-10.0.0.1:52454.service - OpenSSH per-connection server daemon (10.0.0.1:52454). Sep 11 00:21:02.132632 containerd[1571]: time="2025-09-11T00:21:02.132528908Z" level=info msg="connecting to shim d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a" address="unix:///run/containerd/s/86dc31d5941886600a722720fe7bbc5d25e49d079f3b2cf612439505f582c45d" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:21:02.170722 systemd[1]: Started cri-containerd-d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a.scope - libcontainer container d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a. Sep 11 00:21:02.184015 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:21:02.286657 sshd[5043]: Accepted publickey for core from 10.0.0.1 port 52454 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:02.289076 sshd-session[5043]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:02.294784 systemd-logind[1552]: New session 13 of user core. Sep 11 00:21:02.302879 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 11 00:21:02.429144 containerd[1571]: time="2025-09-11T00:21:02.428981803Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bd7c7b5ff-rr4tc,Uid:25061c9c-4e54-43e9-a333-25e7b1cb5982,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a\"" Sep 11 00:21:02.493965 containerd[1571]: time="2025-09-11T00:21:02.493903490Z" level=info msg="CreateContainer within sandbox \"d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 11 00:21:02.701848 containerd[1571]: time="2025-09-11T00:21:02.700633086Z" level=info msg="Container 3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:21:02.712962 systemd-networkd[1474]: calicdd99c743a7: Link UP Sep 11 00:21:02.716205 systemd-networkd[1474]: calicdd99c743a7: Gained carrier Sep 11 00:21:02.719494 sshd[5091]: Connection closed by 10.0.0.1 port 52454 Sep 11 00:21:02.720608 containerd[1571]: time="2025-09-11T00:21:02.720572990Z" level=info msg="CreateContainer within sandbox \"d11d83eb856a3aa320918b4d9fff078e7decb4536f6844df513a80e0bdff8c1a\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371\"" Sep 11 00:21:02.721303 sshd-session[5043]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:02.721756 containerd[1571]: time="2025-09-11T00:21:02.721455707Z" level=info msg="StartContainer for \"3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371\"" Sep 11 00:21:02.724087 containerd[1571]: time="2025-09-11T00:21:02.723763080Z" level=info msg="connecting to shim 3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371" address="unix:///run/containerd/s/86dc31d5941886600a722720fe7bbc5d25e49d079f3b2cf612439505f582c45d" protocol=ttrpc version=3 Sep 11 00:21:02.732936 systemd[1]: sshd@12-10.0.0.82:22-10.0.0.1:52454.service: Deactivated successfully. Sep 11 00:21:02.738852 systemd[1]: session-13.scope: Deactivated successfully. Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:01.972 [INFO][5016] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0 coredns-674b8bbfcf- kube-system f4246163-a863-4efc-b1a5-f23f0e9c31b9 845 0 2025-09-11 00:20:04 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-bcvvs eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calicdd99c743a7 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:01.972 [INFO][5016] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.005 [INFO][5034] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" HandleID="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Workload="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.005 [INFO][5034] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" HandleID="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Workload="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c63e0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-bcvvs", "timestamp":"2025-09-11 00:21:02.005276568 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.005 [INFO][5034] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.005 [INFO][5034] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.005 [INFO][5034] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.013 [INFO][5034] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.258 [INFO][5034] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.322 [INFO][5034] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.482 [INFO][5034] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.487 [INFO][5034] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.487 [INFO][5034] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.496 [INFO][5034] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.502 [INFO][5034] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.698 [INFO][5034] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.699 [INFO][5034] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" host="localhost" Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.699 [INFO][5034] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 11 00:21:02.741213 containerd[1571]: 2025-09-11 00:21:02.699 [INFO][5034] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" HandleID="k8s-pod-network.50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Workload="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.742123 containerd[1571]: 2025-09-11 00:21:02.704 [INFO][5016] cni-plugin/k8s.go 418: Populated endpoint ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"f4246163-a863-4efc-b1a5-f23f0e9c31b9", ResourceVersion:"845", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-bcvvs", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calicdd99c743a7", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:21:02.742123 containerd[1571]: 2025-09-11 00:21:02.705 [INFO][5016] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.742123 containerd[1571]: 2025-09-11 00:21:02.705 [INFO][5016] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicdd99c743a7 ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.742123 containerd[1571]: 2025-09-11 00:21:02.717 [INFO][5016] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.742123 containerd[1571]: 2025-09-11 00:21:02.718 [INFO][5016] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"f4246163-a863-4efc-b1a5-f23f0e9c31b9", ResourceVersion:"845", Generation:0, CreationTimestamp:time.Date(2025, time.September, 11, 0, 20, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a", Pod:"coredns-674b8bbfcf-bcvvs", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calicdd99c743a7", MAC:"2a:26:6c:23:9d:8a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 11 00:21:02.742123 containerd[1571]: 2025-09-11 00:21:02.735 [INFO][5016] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" Namespace="kube-system" Pod="coredns-674b8bbfcf-bcvvs" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--bcvvs-eth0" Sep 11 00:21:02.744915 systemd-logind[1552]: Session 13 logged out. Waiting for processes to exit. Sep 11 00:21:02.746675 systemd-logind[1552]: Removed session 13. Sep 11 00:21:02.762847 systemd[1]: Started cri-containerd-3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371.scope - libcontainer container 3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371. Sep 11 00:21:02.773509 containerd[1571]: time="2025-09-11T00:21:02.772872939Z" level=info msg="connecting to shim 50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a" address="unix:///run/containerd/s/8d8dadc3ddb5a43cbc4aea9e4060007926369f85767d6e0338ae57b490b5da57" namespace=k8s.io protocol=ttrpc version=3 Sep 11 00:21:02.821314 systemd[1]: Started cri-containerd-50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a.scope - libcontainer container 50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a. Sep 11 00:21:02.850697 systemd-resolved[1418]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 11 00:21:03.059753 containerd[1571]: time="2025-09-11T00:21:03.059692325Z" level=info msg="StartContainer for \"3bd6d56d595b1c999c573c28484dcfaa7036a1c08ab6776f42e6e33ace927371\" returns successfully" Sep 11 00:21:03.153285 containerd[1571]: time="2025-09-11T00:21:03.153195566Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-bcvvs,Uid:f4246163-a863-4efc-b1a5-f23f0e9c31b9,Namespace:kube-system,Attempt:0,} returns sandbox id \"50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a\"" Sep 11 00:21:03.154896 kubelet[2783]: E0911 00:21:03.154857 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:03.189850 containerd[1571]: time="2025-09-11T00:21:03.189784385Z" level=info msg="CreateContainer within sandbox \"50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 11 00:21:03.253797 systemd-networkd[1474]: cali16c6143322c: Gained IPv6LL Sep 11 00:21:03.432773 kubelet[2783]: I0911 00:21:03.431192 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7bd7c7b5ff-rr4tc" podStartSLOduration=47.430821185 podStartE2EDuration="47.430821185s" podCreationTimestamp="2025-09-11 00:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:21:03.430641722 +0000 UTC m=+63.824446151" watchObservedRunningTime="2025-09-11 00:21:03.430821185 +0000 UTC m=+63.824625614" Sep 11 00:21:03.452631 containerd[1571]: time="2025-09-11T00:21:03.451954731Z" level=info msg="Container 2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:21:03.455535 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount450320368.mount: Deactivated successfully. Sep 11 00:21:03.463125 containerd[1571]: time="2025-09-11T00:21:03.463007809Z" level=info msg="CreateContainer within sandbox \"50612a8c9f65a9055392f095104c41ca05719930174b38fa2aac50b08550117a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e\"" Sep 11 00:21:03.464364 containerd[1571]: time="2025-09-11T00:21:03.464223162Z" level=info msg="StartContainer for \"2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e\"" Sep 11 00:21:03.466301 containerd[1571]: time="2025-09-11T00:21:03.466200903Z" level=info msg="connecting to shim 2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e" address="unix:///run/containerd/s/8d8dadc3ddb5a43cbc4aea9e4060007926369f85767d6e0338ae57b490b5da57" protocol=ttrpc version=3 Sep 11 00:21:03.494907 systemd[1]: Started cri-containerd-2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e.scope - libcontainer container 2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e. Sep 11 00:21:03.559533 containerd[1571]: time="2025-09-11T00:21:03.559419349Z" level=info msg="StartContainer for \"2e18ac710a6c8f9546689d80fc6362800c854dcd252bbb8f25163aa9b84de02e\" returns successfully" Sep 11 00:21:03.767469 systemd-networkd[1474]: calicdd99c743a7: Gained IPv6LL Sep 11 00:21:04.370463 kubelet[2783]: E0911 00:21:04.370410 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:04.459051 kubelet[2783]: I0911 00:21:04.458321 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-bcvvs" podStartSLOduration=60.458298055 podStartE2EDuration="1m0.458298055s" podCreationTimestamp="2025-09-11 00:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-11 00:21:04.389477507 +0000 UTC m=+64.783281936" watchObservedRunningTime="2025-09-11 00:21:04.458298055 +0000 UTC m=+64.852102484" Sep 11 00:21:05.372384 kubelet[2783]: E0911 00:21:05.372299 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:05.913810 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2689466478.mount: Deactivated successfully. Sep 11 00:21:06.373633 kubelet[2783]: E0911 00:21:06.373591 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:07.376035 kubelet[2783]: E0911 00:21:07.375986 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:07.736198 systemd[1]: Started sshd@13-10.0.0.82:22-10.0.0.1:52470.service - OpenSSH per-connection server daemon (10.0.0.1:52470). Sep 11 00:21:08.124973 sshd[5259]: Accepted publickey for core from 10.0.0.1 port 52470 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:08.127108 sshd-session[5259]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:08.131945 systemd-logind[1552]: New session 14 of user core. Sep 11 00:21:08.141787 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 11 00:21:08.323710 sshd[5265]: Connection closed by 10.0.0.1 port 52470 Sep 11 00:21:08.324058 sshd-session[5259]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:08.337779 systemd[1]: sshd@13-10.0.0.82:22-10.0.0.1:52470.service: Deactivated successfully. Sep 11 00:21:08.339919 systemd[1]: session-14.scope: Deactivated successfully. Sep 11 00:21:08.340762 systemd-logind[1552]: Session 14 logged out. Waiting for processes to exit. Sep 11 00:21:08.344080 systemd[1]: Started sshd@14-10.0.0.82:22-10.0.0.1:52472.service - OpenSSH per-connection server daemon (10.0.0.1:52472). Sep 11 00:21:08.344866 systemd-logind[1552]: Removed session 14. Sep 11 00:21:08.407466 sshd[5285]: Accepted publickey for core from 10.0.0.1 port 52472 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:08.409315 sshd-session[5285]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:08.414779 systemd-logind[1552]: New session 15 of user core. Sep 11 00:21:08.423689 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 11 00:21:08.563351 containerd[1571]: time="2025-09-11T00:21:08.563259975Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:08.625385 containerd[1571]: time="2025-09-11T00:21:08.625300862Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Sep 11 00:21:08.724937 containerd[1571]: time="2025-09-11T00:21:08.724772258Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:08.764146 containerd[1571]: time="2025-09-11T00:21:08.764047095Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:08.764953 containerd[1571]: time="2025-09-11T00:21:08.764767920Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 9.954345001s" Sep 11 00:21:08.764953 containerd[1571]: time="2025-09-11T00:21:08.764800692Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Sep 11 00:21:08.770320 sshd[5288]: Connection closed by 10.0.0.1 port 52472 Sep 11 00:21:08.772729 sshd-session[5285]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:08.777849 containerd[1571]: time="2025-09-11T00:21:08.777822228Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 11 00:21:08.780211 systemd[1]: sshd@14-10.0.0.82:22-10.0.0.1:52472.service: Deactivated successfully. Sep 11 00:21:08.783164 systemd[1]: session-15.scope: Deactivated successfully. Sep 11 00:21:08.784630 systemd-logind[1552]: Session 15 logged out. Waiting for processes to exit. Sep 11 00:21:08.788659 systemd[1]: Started sshd@15-10.0.0.82:22-10.0.0.1:52488.service - OpenSSH per-connection server daemon (10.0.0.1:52488). Sep 11 00:21:08.789314 systemd-logind[1552]: Removed session 15. Sep 11 00:21:08.843319 sshd[5299]: Accepted publickey for core from 10.0.0.1 port 52488 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:08.844846 sshd-session[5299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:08.850187 systemd-logind[1552]: New session 16 of user core. Sep 11 00:21:08.858671 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 11 00:21:08.967406 containerd[1571]: time="2025-09-11T00:21:08.967341266Z" level=info msg="CreateContainer within sandbox \"8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 11 00:21:09.280250 sshd[5302]: Connection closed by 10.0.0.1 port 52488 Sep 11 00:21:09.286770 systemd[1]: sshd@15-10.0.0.82:22-10.0.0.1:52488.service: Deactivated successfully. Sep 11 00:21:09.280684 sshd-session[5299]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:09.289200 systemd[1]: session-16.scope: Deactivated successfully. Sep 11 00:21:09.290079 systemd-logind[1552]: Session 16 logged out. Waiting for processes to exit. Sep 11 00:21:09.292201 systemd-logind[1552]: Removed session 16. Sep 11 00:21:09.376741 containerd[1571]: time="2025-09-11T00:21:09.376670497Z" level=info msg="Container f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:21:09.796101 containerd[1571]: time="2025-09-11T00:21:09.796030025Z" level=info msg="CreateContainer within sandbox \"8f31a2506672d3e1835609b4843875b715c72a062f673208c7080c4bad4b08e3\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\"" Sep 11 00:21:09.796571 containerd[1571]: time="2025-09-11T00:21:09.796510100Z" level=info msg="StartContainer for \"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\"" Sep 11 00:21:09.798052 containerd[1571]: time="2025-09-11T00:21:09.798010951Z" level=info msg="connecting to shim f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab" address="unix:///run/containerd/s/dacb55ee252281b9eaff421d5ff23fd66fa6253e0a9c00219b3156ee99eecfea" protocol=ttrpc version=3 Sep 11 00:21:09.827918 systemd[1]: Started cri-containerd-f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab.scope - libcontainer container f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab. Sep 11 00:21:10.129984 containerd[1571]: time="2025-09-11T00:21:10.129828215Z" level=info msg="StartContainer for \"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\" returns successfully" Sep 11 00:21:10.444667 kubelet[2783]: I0911 00:21:10.444314 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-k7tqf" podStartSLOduration=31.924001905 podStartE2EDuration="51.444186515s" podCreationTimestamp="2025-09-11 00:20:19 +0000 UTC" firstStartedPulling="2025-09-11 00:20:49.257389314 +0000 UTC m=+49.651193733" lastFinishedPulling="2025-09-11 00:21:08.777573914 +0000 UTC m=+69.171378343" observedRunningTime="2025-09-11 00:21:10.440804641 +0000 UTC m=+70.834609080" watchObservedRunningTime="2025-09-11 00:21:10.444186515 +0000 UTC m=+70.837990944" Sep 11 00:21:10.828374 containerd[1571]: time="2025-09-11T00:21:10.828225247Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\" id:\"2aca30fd3c65456fd96e4394ea64ad251be000a4c8fa23f2e644319369c888f1\" pid:5364 exited_at:{seconds:1757550070 nanos:827593934}" Sep 11 00:21:11.539047 containerd[1571]: time="2025-09-11T00:21:11.538974790Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\" id:\"0eb1869d63906c2d9e1715ebf23286a5330045c9816538c8dce23a5b84cdbd5b\" pid:5388 exit_status:1 exited_at:{seconds:1757550071 nanos:531782048}" Sep 11 00:21:12.530284 containerd[1571]: time="2025-09-11T00:21:12.530232726Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\" id:\"35c8bb67ff5a09a9cf374f6bb261255b80cfb9102238d9e6f79ce14c3fe4b9c6\" pid:5412 exit_status:1 exited_at:{seconds:1757550072 nanos:529766769}" Sep 11 00:21:12.827649 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4079296324.mount: Deactivated successfully. Sep 11 00:21:14.301498 systemd[1]: Started sshd@16-10.0.0.82:22-10.0.0.1:39948.service - OpenSSH per-connection server daemon (10.0.0.1:39948). Sep 11 00:21:14.476805 sshd[5431]: Accepted publickey for core from 10.0.0.1 port 39948 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:14.479099 sshd-session[5431]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:14.483861 systemd-logind[1552]: New session 17 of user core. Sep 11 00:21:14.491689 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 11 00:21:14.994408 sshd[5435]: Connection closed by 10.0.0.1 port 39948 Sep 11 00:21:14.994834 sshd-session[5431]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:15.000087 systemd[1]: sshd@16-10.0.0.82:22-10.0.0.1:39948.service: Deactivated successfully. Sep 11 00:21:15.003261 systemd[1]: session-17.scope: Deactivated successfully. Sep 11 00:21:15.004329 systemd-logind[1552]: Session 17 logged out. Waiting for processes to exit. Sep 11 00:21:15.006012 systemd-logind[1552]: Removed session 17. Sep 11 00:21:15.163360 containerd[1571]: time="2025-09-11T00:21:15.163274933Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:15.191003 containerd[1571]: time="2025-09-11T00:21:15.190610284Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Sep 11 00:21:15.241266 containerd[1571]: time="2025-09-11T00:21:15.241206785Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:15.350244 containerd[1571]: time="2025-09-11T00:21:15.350166973Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:15.350833 containerd[1571]: time="2025-09-11T00:21:15.350803563Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 6.572839265s" Sep 11 00:21:15.350833 containerd[1571]: time="2025-09-11T00:21:15.350833680Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Sep 11 00:21:15.351965 containerd[1571]: time="2025-09-11T00:21:15.351924596Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 11 00:21:15.599961 containerd[1571]: time="2025-09-11T00:21:15.599912308Z" level=info msg="CreateContainer within sandbox \"c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 11 00:21:15.958047 containerd[1571]: time="2025-09-11T00:21:15.957997532Z" level=info msg="Container 92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:21:16.206317 containerd[1571]: time="2025-09-11T00:21:16.206244421Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829\" id:\"50056dd29c44365c5113614caac871860c781c94b960fb79cf2f924509322576\" pid:5469 exit_status:1 exited_at:{seconds:1757550076 nanos:205882412}" Sep 11 00:21:16.294070 containerd[1571]: time="2025-09-11T00:21:16.293951895Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829\" id:\"3f5038b58e62784c2039beef4b2457a006ac0611a90f7b7013c6fe9273eed8cb\" pid:5493 exit_status:1 exited_at:{seconds:1757550076 nanos:293632557}" Sep 11 00:21:17.544312 containerd[1571]: time="2025-09-11T00:21:17.544233595Z" level=info msg="CreateContainer within sandbox \"c899c08c007fb7229f122a067fca375bdc1d46735cd2c885a2f419ffe2e18b0d\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc\"" Sep 11 00:21:17.545253 containerd[1571]: time="2025-09-11T00:21:17.545200904Z" level=info msg="StartContainer for \"92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc\"" Sep 11 00:21:17.547443 containerd[1571]: time="2025-09-11T00:21:17.547379286Z" level=info msg="connecting to shim 92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc" address="unix:///run/containerd/s/b5acd7ab6b4a25f27ad2ae2c1700d3565f70cd9730f8d581f01526cb3281548a" protocol=ttrpc version=3 Sep 11 00:21:17.573720 systemd[1]: Started cri-containerd-92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc.scope - libcontainer container 92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc. Sep 11 00:21:17.637947 containerd[1571]: time="2025-09-11T00:21:17.637899453Z" level=info msg="StartContainer for \"92e61cbeefd5362821fe3f846b10f8a5cf9dd8ffba8c431a09968227efe7d3fc\" returns successfully" Sep 11 00:21:18.567187 kubelet[2783]: I0911 00:21:18.567107 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-78dbcbf6cb-5w25q" podStartSLOduration=4.798100994 podStartE2EDuration="33.567091466s" podCreationTimestamp="2025-09-11 00:20:45 +0000 UTC" firstStartedPulling="2025-09-11 00:20:46.582807523 +0000 UTC m=+46.976611952" lastFinishedPulling="2025-09-11 00:21:15.351797995 +0000 UTC m=+75.745602424" observedRunningTime="2025-09-11 00:21:18.566656199 +0000 UTC m=+78.960460628" watchObservedRunningTime="2025-09-11 00:21:18.567091466 +0000 UTC m=+78.960895895" Sep 11 00:21:19.471462 containerd[1571]: time="2025-09-11T00:21:19.471400674Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:19.472953 containerd[1571]: time="2025-09-11T00:21:19.472896886Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Sep 11 00:21:19.474253 containerd[1571]: time="2025-09-11T00:21:19.474192248Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:19.476170 containerd[1571]: time="2025-09-11T00:21:19.476116154Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:19.476703 containerd[1571]: time="2025-09-11T00:21:19.476663584Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 4.124704573s" Sep 11 00:21:19.476703 containerd[1571]: time="2025-09-11T00:21:19.476701335Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Sep 11 00:21:19.482761 containerd[1571]: time="2025-09-11T00:21:19.482706154Z" level=info msg="CreateContainer within sandbox \"db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 11 00:21:19.493177 containerd[1571]: time="2025-09-11T00:21:19.493103699Z" level=info msg="Container 01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:21:19.518226 containerd[1571]: time="2025-09-11T00:21:19.518138715Z" level=info msg="CreateContainer within sandbox \"db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e\"" Sep 11 00:21:19.519024 containerd[1571]: time="2025-09-11T00:21:19.518982498Z" level=info msg="StartContainer for \"01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e\"" Sep 11 00:21:19.520922 containerd[1571]: time="2025-09-11T00:21:19.520889040Z" level=info msg="connecting to shim 01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e" address="unix:///run/containerd/s/f926348bf826c3b83b2327da4f80fa584801ec988cec151c5d771797ebe48bfa" protocol=ttrpc version=3 Sep 11 00:21:19.547728 systemd[1]: Started cri-containerd-01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e.scope - libcontainer container 01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e. Sep 11 00:21:19.628374 containerd[1571]: time="2025-09-11T00:21:19.628310477Z" level=info msg="StartContainer for \"01320e5d563ccc05a62736b81183397161ed1395d6d564297659f15e0ae05c8e\" returns successfully" Sep 11 00:21:19.629498 containerd[1571]: time="2025-09-11T00:21:19.629470761Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 11 00:21:20.013283 systemd[1]: Started sshd@17-10.0.0.82:22-10.0.0.1:46892.service - OpenSSH per-connection server daemon (10.0.0.1:46892). Sep 11 00:21:20.105986 sshd[5577]: Accepted publickey for core from 10.0.0.1 port 46892 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:20.108310 sshd-session[5577]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:20.114872 systemd-logind[1552]: New session 18 of user core. Sep 11 00:21:20.132916 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 11 00:21:20.373603 sshd[5581]: Connection closed by 10.0.0.1 port 46892 Sep 11 00:21:20.372436 sshd-session[5577]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:20.378523 systemd[1]: sshd@17-10.0.0.82:22-10.0.0.1:46892.service: Deactivated successfully. Sep 11 00:21:20.382786 systemd[1]: session-18.scope: Deactivated successfully. Sep 11 00:21:20.385796 systemd-logind[1552]: Session 18 logged out. Waiting for processes to exit. Sep 11 00:21:20.387813 systemd-logind[1552]: Removed session 18. Sep 11 00:21:21.134378 containerd[1571]: time="2025-09-11T00:21:21.134327385Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\" id:\"44998756e681730016092b84f922f58eebe791c9cc085894c61d7040b4042a2d\" pid:5606 exited_at:{seconds:1757550081 nanos:133788522}" Sep 11 00:21:22.413745 containerd[1571]: time="2025-09-11T00:21:22.413669293Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:22.414701 containerd[1571]: time="2025-09-11T00:21:22.414655947Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Sep 11 00:21:22.416324 containerd[1571]: time="2025-09-11T00:21:22.416291091Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:22.418482 containerd[1571]: time="2025-09-11T00:21:22.418443547Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 11 00:21:22.419065 containerd[1571]: time="2025-09-11T00:21:22.419025381Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 2.789524472s" Sep 11 00:21:22.419121 containerd[1571]: time="2025-09-11T00:21:22.419065147Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Sep 11 00:21:22.424651 containerd[1571]: time="2025-09-11T00:21:22.424600425Z" level=info msg="CreateContainer within sandbox \"db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 11 00:21:22.435870 containerd[1571]: time="2025-09-11T00:21:22.435809915Z" level=info msg="Container 5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b: CDI devices from CRI Config.CDIDevices: []" Sep 11 00:21:22.447662 containerd[1571]: time="2025-09-11T00:21:22.447612190Z" level=info msg="CreateContainer within sandbox \"db2435133b3dfecf818ea1e93192ccbac2088e0c799104d9973364df3543a455\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b\"" Sep 11 00:21:22.448281 containerd[1571]: time="2025-09-11T00:21:22.448238749Z" level=info msg="StartContainer for \"5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b\"" Sep 11 00:21:22.449930 containerd[1571]: time="2025-09-11T00:21:22.449904011Z" level=info msg="connecting to shim 5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b" address="unix:///run/containerd/s/f926348bf826c3b83b2327da4f80fa584801ec988cec151c5d771797ebe48bfa" protocol=ttrpc version=3 Sep 11 00:21:22.485901 systemd[1]: Started cri-containerd-5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b.scope - libcontainer container 5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b. Sep 11 00:21:22.904921 kubelet[2783]: I0911 00:21:22.904875 2783 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 11 00:21:23.138070 kubelet[2783]: I0911 00:21:23.138001 2783 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 11 00:21:23.233947 containerd[1571]: time="2025-09-11T00:21:23.233827079Z" level=info msg="StartContainer for \"5b37d1c2d3294af6d26e1c3001fe22958dbe7db88a3448d059452d56a6418e4b\" returns successfully" Sep 11 00:21:23.614393 kubelet[2783]: I0911 00:21:23.614149 2783 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-f8r22" podStartSLOduration=41.454562078 podStartE2EDuration="1m4.614129198s" podCreationTimestamp="2025-09-11 00:20:19 +0000 UTC" firstStartedPulling="2025-09-11 00:20:59.26025305 +0000 UTC m=+59.654057479" lastFinishedPulling="2025-09-11 00:21:22.41982017 +0000 UTC m=+82.813624599" observedRunningTime="2025-09-11 00:21:23.612783445 +0000 UTC m=+84.006587874" watchObservedRunningTime="2025-09-11 00:21:23.614129198 +0000 UTC m=+84.007933627" Sep 11 00:21:25.393151 systemd[1]: Started sshd@18-10.0.0.82:22-10.0.0.1:46908.service - OpenSSH per-connection server daemon (10.0.0.1:46908). Sep 11 00:21:25.489956 sshd[5656]: Accepted publickey for core from 10.0.0.1 port 46908 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:25.492467 sshd-session[5656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:25.499171 systemd-logind[1552]: New session 19 of user core. Sep 11 00:21:25.503703 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 11 00:21:25.867853 sshd[5659]: Connection closed by 10.0.0.1 port 46908 Sep 11 00:21:25.868284 sshd-session[5656]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:25.873933 systemd[1]: sshd@18-10.0.0.82:22-10.0.0.1:46908.service: Deactivated successfully. Sep 11 00:21:25.876228 systemd[1]: session-19.scope: Deactivated successfully. Sep 11 00:21:25.877231 systemd-logind[1552]: Session 19 logged out. Waiting for processes to exit. Sep 11 00:21:25.878741 systemd-logind[1552]: Removed session 19. Sep 11 00:21:28.771638 kubelet[2783]: E0911 00:21:28.771504 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:30.882190 systemd[1]: Started sshd@19-10.0.0.82:22-10.0.0.1:55394.service - OpenSSH per-connection server daemon (10.0.0.1:55394). Sep 11 00:21:30.946635 sshd[5681]: Accepted publickey for core from 10.0.0.1 port 55394 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:30.948688 sshd-session[5681]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:30.954484 systemd-logind[1552]: New session 20 of user core. Sep 11 00:21:30.966808 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 11 00:21:31.111312 sshd[5684]: Connection closed by 10.0.0.1 port 55394 Sep 11 00:21:31.111748 sshd-session[5681]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:31.117099 systemd[1]: sshd@19-10.0.0.82:22-10.0.0.1:55394.service: Deactivated successfully. Sep 11 00:21:31.119784 systemd[1]: session-20.scope: Deactivated successfully. Sep 11 00:21:31.120866 systemd-logind[1552]: Session 20 logged out. Waiting for processes to exit. Sep 11 00:21:31.122437 systemd-logind[1552]: Removed session 20. Sep 11 00:21:31.771265 kubelet[2783]: E0911 00:21:31.771217 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:32.202480 containerd[1571]: time="2025-09-11T00:21:32.202359511Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\" id:\"156f0fd7067e16138325a1b135dccbf790065f503c07aa309ff021d966d6ffd9\" pid:5709 exited_at:{seconds:1757550092 nanos:201973510}" Sep 11 00:21:34.770813 kubelet[2783]: E0911 00:21:34.770755 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:35.771593 kubelet[2783]: E0911 00:21:35.771510 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:36.130605 systemd[1]: Started sshd@20-10.0.0.82:22-10.0.0.1:55400.service - OpenSSH per-connection server daemon (10.0.0.1:55400). Sep 11 00:21:36.195154 sshd[5722]: Accepted publickey for core from 10.0.0.1 port 55400 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:36.197094 sshd-session[5722]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:36.202714 systemd-logind[1552]: New session 21 of user core. Sep 11 00:21:36.213730 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 11 00:21:36.356514 sshd[5725]: Connection closed by 10.0.0.1 port 55400 Sep 11 00:21:36.357260 sshd-session[5722]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:36.367106 systemd[1]: sshd@20-10.0.0.82:22-10.0.0.1:55400.service: Deactivated successfully. Sep 11 00:21:36.369384 systemd[1]: session-21.scope: Deactivated successfully. Sep 11 00:21:36.370480 systemd-logind[1552]: Session 21 logged out. Waiting for processes to exit. Sep 11 00:21:36.373967 systemd[1]: Started sshd@21-10.0.0.82:22-10.0.0.1:55410.service - OpenSSH per-connection server daemon (10.0.0.1:55410). Sep 11 00:21:36.375525 systemd-logind[1552]: Removed session 21. Sep 11 00:21:36.446407 sshd[5738]: Accepted publickey for core from 10.0.0.1 port 55410 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:36.448513 sshd-session[5738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:36.454523 systemd-logind[1552]: New session 22 of user core. Sep 11 00:21:36.461773 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 11 00:21:36.823871 sshd[5741]: Connection closed by 10.0.0.1 port 55410 Sep 11 00:21:36.825333 sshd-session[5738]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:36.840321 systemd[1]: sshd@21-10.0.0.82:22-10.0.0.1:55410.service: Deactivated successfully. Sep 11 00:21:36.843236 systemd[1]: session-22.scope: Deactivated successfully. Sep 11 00:21:36.847161 systemd-logind[1552]: Session 22 logged out. Waiting for processes to exit. Sep 11 00:21:36.850675 systemd[1]: Started sshd@22-10.0.0.82:22-10.0.0.1:55426.service - OpenSSH per-connection server daemon (10.0.0.1:55426). Sep 11 00:21:36.852011 systemd-logind[1552]: Removed session 22. Sep 11 00:21:36.916908 sshd[5753]: Accepted publickey for core from 10.0.0.1 port 55426 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:36.918663 sshd-session[5753]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:36.923237 systemd-logind[1552]: New session 23 of user core. Sep 11 00:21:36.933769 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 11 00:21:37.794751 sshd[5756]: Connection closed by 10.0.0.1 port 55426 Sep 11 00:21:37.795501 sshd-session[5753]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:37.808203 systemd[1]: sshd@22-10.0.0.82:22-10.0.0.1:55426.service: Deactivated successfully. Sep 11 00:21:37.811333 systemd[1]: session-23.scope: Deactivated successfully. Sep 11 00:21:37.812971 systemd-logind[1552]: Session 23 logged out. Waiting for processes to exit. Sep 11 00:21:37.815668 systemd-logind[1552]: Removed session 23. Sep 11 00:21:37.817302 systemd[1]: Started sshd@23-10.0.0.82:22-10.0.0.1:55442.service - OpenSSH per-connection server daemon (10.0.0.1:55442). Sep 11 00:21:37.931981 sshd[5781]: Accepted publickey for core from 10.0.0.1 port 55442 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:37.934065 sshd-session[5781]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:37.939372 systemd-logind[1552]: New session 24 of user core. Sep 11 00:21:37.948774 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 11 00:21:38.255604 sshd[5785]: Connection closed by 10.0.0.1 port 55442 Sep 11 00:21:38.255202 sshd-session[5781]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:38.270037 systemd[1]: sshd@23-10.0.0.82:22-10.0.0.1:55442.service: Deactivated successfully. Sep 11 00:21:38.272676 systemd[1]: session-24.scope: Deactivated successfully. Sep 11 00:21:38.273681 systemd-logind[1552]: Session 24 logged out. Waiting for processes to exit. Sep 11 00:21:38.281172 systemd[1]: Started sshd@24-10.0.0.82:22-10.0.0.1:55450.service - OpenSSH per-connection server daemon (10.0.0.1:55450). Sep 11 00:21:38.283449 systemd-logind[1552]: Removed session 24. Sep 11 00:21:38.342887 sshd[5796]: Accepted publickey for core from 10.0.0.1 port 55450 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:38.345238 sshd-session[5796]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:38.350818 systemd-logind[1552]: New session 25 of user core. Sep 11 00:21:38.359688 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 11 00:21:38.482180 sshd[5799]: Connection closed by 10.0.0.1 port 55450 Sep 11 00:21:38.482578 sshd-session[5796]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:38.486247 systemd[1]: sshd@24-10.0.0.82:22-10.0.0.1:55450.service: Deactivated successfully. Sep 11 00:21:38.489055 systemd[1]: session-25.scope: Deactivated successfully. Sep 11 00:21:38.492309 systemd-logind[1552]: Session 25 logged out. Waiting for processes to exit. Sep 11 00:21:38.493406 systemd-logind[1552]: Removed session 25. Sep 11 00:21:42.526147 containerd[1571]: time="2025-09-11T00:21:42.526096500Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f86e5014dbc8d10579cb62c9662218f0c916a77f459c8c73546b659c0f5459ab\" id:\"85937417606f332141bbc5b13f9dca7180aacb0d3d0beb07e25e0c291fcdbafe\" pid:5822 exited_at:{seconds:1757550102 nanos:525752680}" Sep 11 00:21:43.500841 systemd[1]: Started sshd@25-10.0.0.82:22-10.0.0.1:58796.service - OpenSSH per-connection server daemon (10.0.0.1:58796). Sep 11 00:21:43.585836 sshd[5841]: Accepted publickey for core from 10.0.0.1 port 58796 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:43.587963 sshd-session[5841]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:43.593309 systemd-logind[1552]: New session 26 of user core. Sep 11 00:21:43.605770 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 11 00:21:43.727573 sshd[5844]: Connection closed by 10.0.0.1 port 58796 Sep 11 00:21:43.727902 sshd-session[5841]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:43.732337 systemd[1]: sshd@25-10.0.0.82:22-10.0.0.1:58796.service: Deactivated successfully. Sep 11 00:21:43.734671 systemd[1]: session-26.scope: Deactivated successfully. Sep 11 00:21:43.735686 systemd-logind[1552]: Session 26 logged out. Waiting for processes to exit. Sep 11 00:21:43.737485 systemd-logind[1552]: Removed session 26. Sep 11 00:21:46.393185 containerd[1571]: time="2025-09-11T00:21:46.393113726Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0d5fbe653197ca218226ed31580be1898a8366e50c3a0f9d5585e1571f7ab829\" id:\"b3bc7229272c05515d25eb7e47a53dcaa80188e297f8bcf5cf8166540bd84122\" pid:5869 exited_at:{seconds:1757550106 nanos:392653656}" Sep 11 00:21:48.747892 systemd[1]: Started sshd@26-10.0.0.82:22-10.0.0.1:58810.service - OpenSSH per-connection server daemon (10.0.0.1:58810). Sep 11 00:21:48.771130 kubelet[2783]: E0911 00:21:48.771070 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 11 00:21:48.879396 sshd[5882]: Accepted publickey for core from 10.0.0.1 port 58810 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:48.881651 sshd-session[5882]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:48.886479 systemd-logind[1552]: New session 27 of user core. Sep 11 00:21:48.905869 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 11 00:21:49.202181 sshd[5885]: Connection closed by 10.0.0.1 port 58810 Sep 11 00:21:49.204369 sshd-session[5882]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:49.210687 systemd-logind[1552]: Session 27 logged out. Waiting for processes to exit. Sep 11 00:21:49.211635 systemd[1]: sshd@26-10.0.0.82:22-10.0.0.1:58810.service: Deactivated successfully. Sep 11 00:21:49.214945 systemd[1]: session-27.scope: Deactivated successfully. Sep 11 00:21:49.217783 systemd-logind[1552]: Removed session 27. Sep 11 00:21:51.107703 containerd[1571]: time="2025-09-11T00:21:51.107602175Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa95de66fdf91444230fb352c92397b2e53dbf22f25791111d89ec576f151246\" id:\"fa247854e811ee4e8ac69cdb8200235d871122620d855a2ae467aa4af5a2bb78\" pid:5911 exited_at:{seconds:1757550111 nanos:107386277}" Sep 11 00:21:54.220458 systemd[1]: Started sshd@27-10.0.0.82:22-10.0.0.1:56732.service - OpenSSH per-connection server daemon (10.0.0.1:56732). Sep 11 00:21:54.284986 sshd[5922]: Accepted publickey for core from 10.0.0.1 port 56732 ssh2: RSA SHA256:iG/lPcoyZucxTWaZiRVFFdQ+jOuDk1s0lgCqGD+sReM Sep 11 00:21:54.287084 sshd-session[5922]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 11 00:21:54.292564 systemd-logind[1552]: New session 28 of user core. Sep 11 00:21:54.306780 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 11 00:21:54.435864 sshd[5925]: Connection closed by 10.0.0.1 port 56732 Sep 11 00:21:54.436225 sshd-session[5922]: pam_unix(sshd:session): session closed for user core Sep 11 00:21:54.441756 systemd[1]: sshd@27-10.0.0.82:22-10.0.0.1:56732.service: Deactivated successfully. Sep 11 00:21:54.444598 systemd[1]: session-28.scope: Deactivated successfully. Sep 11 00:21:54.445965 systemd-logind[1552]: Session 28 logged out. Waiting for processes to exit. Sep 11 00:21:54.447528 systemd-logind[1552]: Removed session 28. Sep 11 00:21:54.771034 kubelet[2783]: E0911 00:21:54.770978 2783 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"