Sep 16 04:48:42.930952 kernel: Linux version 6.12.47-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Tue Sep 16 03:05:42 -00 2025 Sep 16 04:48:42.930981 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=0b876f86a632750e9937176808a48c2452d5168964273bcfc3c72f2a26140c06 Sep 16 04:48:42.930997 kernel: BIOS-provided physical RAM map: Sep 16 04:48:42.931006 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 16 04:48:42.931014 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 16 04:48:42.931022 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 16 04:48:42.931033 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Sep 16 04:48:42.931042 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 16 04:48:42.931054 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Sep 16 04:48:42.931066 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Sep 16 04:48:42.931075 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Sep 16 04:48:42.931085 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Sep 16 04:48:42.931093 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Sep 16 04:48:42.931102 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Sep 16 04:48:42.931114 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Sep 16 04:48:42.931127 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 16 04:48:42.931140 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Sep 16 04:48:42.931150 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Sep 16 04:48:42.931160 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Sep 16 04:48:42.931170 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Sep 16 04:48:42.931179 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Sep 16 04:48:42.931189 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 16 04:48:42.931199 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 16 04:48:42.931209 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 16 04:48:42.931219 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Sep 16 04:48:42.931267 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 16 04:48:42.931278 kernel: NX (Execute Disable) protection: active Sep 16 04:48:42.931287 kernel: APIC: Static calls initialized Sep 16 04:48:42.931297 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Sep 16 04:48:42.931308 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Sep 16 04:48:42.931317 kernel: extended physical RAM map: Sep 16 04:48:42.931327 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 16 04:48:42.931337 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 16 04:48:42.931347 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 16 04:48:42.931357 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Sep 16 04:48:42.931367 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 16 04:48:42.931380 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Sep 16 04:48:42.931390 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Sep 16 04:48:42.931400 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Sep 16 04:48:42.931410 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Sep 16 04:48:42.931425 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Sep 16 04:48:42.931435 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Sep 16 04:48:42.931448 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Sep 16 04:48:42.931459 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Sep 16 04:48:42.931483 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Sep 16 04:48:42.931493 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Sep 16 04:48:42.931503 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Sep 16 04:48:42.931514 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 16 04:48:42.931525 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Sep 16 04:48:42.931535 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Sep 16 04:48:42.931546 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Sep 16 04:48:42.931556 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Sep 16 04:48:42.931570 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Sep 16 04:48:42.931580 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 16 04:48:42.931591 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 16 04:48:42.931601 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 16 04:48:42.931612 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Sep 16 04:48:42.931622 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 16 04:48:42.931636 kernel: efi: EFI v2.7 by EDK II Sep 16 04:48:42.931647 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Sep 16 04:48:42.931658 kernel: random: crng init done Sep 16 04:48:42.931671 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Sep 16 04:48:42.931682 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Sep 16 04:48:42.931698 kernel: secureboot: Secure boot disabled Sep 16 04:48:42.931708 kernel: SMBIOS 2.8 present. Sep 16 04:48:42.931719 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Sep 16 04:48:42.931729 kernel: DMI: Memory slots populated: 1/1 Sep 16 04:48:42.931740 kernel: Hypervisor detected: KVM Sep 16 04:48:42.931750 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 16 04:48:42.931761 kernel: kvm-clock: using sched offset of 5238590505 cycles Sep 16 04:48:42.931772 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 16 04:48:42.931783 kernel: tsc: Detected 2794.750 MHz processor Sep 16 04:48:42.931794 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 16 04:48:42.931804 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 16 04:48:42.931818 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Sep 16 04:48:42.931828 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 16 04:48:42.931839 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 16 04:48:42.931849 kernel: Using GB pages for direct mapping Sep 16 04:48:42.931859 kernel: ACPI: Early table checksum verification disabled Sep 16 04:48:42.931869 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Sep 16 04:48:42.931879 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Sep 16 04:48:42.931889 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 16 04:48:42.931899 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 16 04:48:42.931912 kernel: ACPI: FACS 0x000000009CBDD000 000040 Sep 16 04:48:42.931922 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 16 04:48:42.931933 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 16 04:48:42.931942 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 16 04:48:42.931952 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 16 04:48:42.931962 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Sep 16 04:48:42.931972 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Sep 16 04:48:42.931982 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Sep 16 04:48:42.931993 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Sep 16 04:48:42.932006 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Sep 16 04:48:42.932016 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Sep 16 04:48:42.932026 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Sep 16 04:48:42.932037 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Sep 16 04:48:42.932047 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Sep 16 04:48:42.932057 kernel: No NUMA configuration found Sep 16 04:48:42.932067 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Sep 16 04:48:42.932077 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Sep 16 04:48:42.932088 kernel: Zone ranges: Sep 16 04:48:42.932101 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 16 04:48:42.932111 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Sep 16 04:48:42.932121 kernel: Normal empty Sep 16 04:48:42.932130 kernel: Device empty Sep 16 04:48:42.932140 kernel: Movable zone start for each node Sep 16 04:48:42.932150 kernel: Early memory node ranges Sep 16 04:48:42.932160 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 16 04:48:42.932170 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Sep 16 04:48:42.932184 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Sep 16 04:48:42.932199 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Sep 16 04:48:42.932210 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Sep 16 04:48:42.932221 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Sep 16 04:48:42.932248 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Sep 16 04:48:42.932259 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Sep 16 04:48:42.932270 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Sep 16 04:48:42.932281 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 16 04:48:42.932295 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 16 04:48:42.932322 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Sep 16 04:48:42.932334 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 16 04:48:42.932345 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Sep 16 04:48:42.932357 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Sep 16 04:48:42.932371 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Sep 16 04:48:42.932383 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Sep 16 04:48:42.932395 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Sep 16 04:48:42.932406 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 16 04:48:42.932417 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 16 04:48:42.932432 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 16 04:48:42.932444 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 16 04:48:42.932455 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 16 04:48:42.932477 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 16 04:48:42.932489 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 16 04:48:42.932500 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 16 04:48:42.932512 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 16 04:48:42.932523 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 16 04:48:42.932534 kernel: TSC deadline timer available Sep 16 04:48:42.932549 kernel: CPU topo: Max. logical packages: 1 Sep 16 04:48:42.932560 kernel: CPU topo: Max. logical dies: 1 Sep 16 04:48:42.932571 kernel: CPU topo: Max. dies per package: 1 Sep 16 04:48:42.932583 kernel: CPU topo: Max. threads per core: 1 Sep 16 04:48:42.932594 kernel: CPU topo: Num. cores per package: 4 Sep 16 04:48:42.932606 kernel: CPU topo: Num. threads per package: 4 Sep 16 04:48:42.932617 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Sep 16 04:48:42.932629 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 16 04:48:42.932640 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 16 04:48:42.932651 kernel: kvm-guest: setup PV sched yield Sep 16 04:48:42.932666 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Sep 16 04:48:42.932676 kernel: Booting paravirtualized kernel on KVM Sep 16 04:48:42.932687 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 16 04:48:42.932698 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 16 04:48:42.932709 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Sep 16 04:48:42.932721 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Sep 16 04:48:42.932733 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 16 04:48:42.932744 kernel: kvm-guest: PV spinlocks enabled Sep 16 04:48:42.932759 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 16 04:48:42.932771 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=0b876f86a632750e9937176808a48c2452d5168964273bcfc3c72f2a26140c06 Sep 16 04:48:42.932786 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 16 04:48:42.932797 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 16 04:48:42.932808 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 16 04:48:42.932819 kernel: Fallback order for Node 0: 0 Sep 16 04:48:42.932829 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Sep 16 04:48:42.932840 kernel: Policy zone: DMA32 Sep 16 04:48:42.932851 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 16 04:48:42.932865 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 16 04:48:42.932875 kernel: ftrace: allocating 40125 entries in 157 pages Sep 16 04:48:42.932886 kernel: ftrace: allocated 157 pages with 5 groups Sep 16 04:48:42.932896 kernel: Dynamic Preempt: voluntary Sep 16 04:48:42.932906 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 16 04:48:42.932918 kernel: rcu: RCU event tracing is enabled. Sep 16 04:48:42.932929 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 16 04:48:42.932940 kernel: Trampoline variant of Tasks RCU enabled. Sep 16 04:48:42.932951 kernel: Rude variant of Tasks RCU enabled. Sep 16 04:48:42.932965 kernel: Tracing variant of Tasks RCU enabled. Sep 16 04:48:42.932976 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 16 04:48:42.932991 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 16 04:48:42.933003 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 16 04:48:42.933014 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 16 04:48:42.933025 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 16 04:48:42.933035 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 16 04:48:42.933046 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 16 04:48:42.933056 kernel: Console: colour dummy device 80x25 Sep 16 04:48:42.933071 kernel: printk: legacy console [ttyS0] enabled Sep 16 04:48:42.933081 kernel: ACPI: Core revision 20240827 Sep 16 04:48:42.933092 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 16 04:48:42.933103 kernel: APIC: Switch to symmetric I/O mode setup Sep 16 04:48:42.933113 kernel: x2apic enabled Sep 16 04:48:42.933124 kernel: APIC: Switched APIC routing to: physical x2apic Sep 16 04:48:42.933134 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 16 04:48:42.933145 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 16 04:48:42.933156 kernel: kvm-guest: setup PV IPIs Sep 16 04:48:42.933173 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 16 04:48:42.933187 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848e100549, max_idle_ns: 440795215505 ns Sep 16 04:48:42.933198 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Sep 16 04:48:42.933209 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 16 04:48:42.933219 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 16 04:48:42.933253 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 16 04:48:42.933264 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 16 04:48:42.933275 kernel: Spectre V2 : Mitigation: Retpolines Sep 16 04:48:42.933286 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 16 04:48:42.933301 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 16 04:48:42.933312 kernel: active return thunk: retbleed_return_thunk Sep 16 04:48:42.933323 kernel: RETBleed: Mitigation: untrained return thunk Sep 16 04:48:42.933338 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 16 04:48:42.933348 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 16 04:48:42.933359 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 16 04:48:42.933371 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 16 04:48:42.933383 kernel: active return thunk: srso_return_thunk Sep 16 04:48:42.933398 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 16 04:48:42.933409 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 16 04:48:42.933421 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 16 04:48:42.933432 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 16 04:48:42.933443 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 16 04:48:42.933455 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 16 04:48:42.933477 kernel: Freeing SMP alternatives memory: 32K Sep 16 04:48:42.933489 kernel: pid_max: default: 32768 minimum: 301 Sep 16 04:48:42.933500 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 16 04:48:42.933515 kernel: landlock: Up and running. Sep 16 04:48:42.933526 kernel: SELinux: Initializing. Sep 16 04:48:42.933537 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 16 04:48:42.933548 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 16 04:48:42.933560 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 16 04:48:42.933571 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 16 04:48:42.933582 kernel: ... version: 0 Sep 16 04:48:42.933593 kernel: ... bit width: 48 Sep 16 04:48:42.933604 kernel: ... generic registers: 6 Sep 16 04:48:42.933619 kernel: ... value mask: 0000ffffffffffff Sep 16 04:48:42.933630 kernel: ... max period: 00007fffffffffff Sep 16 04:48:42.933641 kernel: ... fixed-purpose events: 0 Sep 16 04:48:42.933653 kernel: ... event mask: 000000000000003f Sep 16 04:48:42.933664 kernel: signal: max sigframe size: 1776 Sep 16 04:48:42.933674 kernel: rcu: Hierarchical SRCU implementation. Sep 16 04:48:42.933685 kernel: rcu: Max phase no-delay instances is 400. Sep 16 04:48:42.933701 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 16 04:48:42.933712 kernel: smp: Bringing up secondary CPUs ... Sep 16 04:48:42.933727 kernel: smpboot: x86: Booting SMP configuration: Sep 16 04:48:42.933737 kernel: .... node #0, CPUs: #1 #2 #3 Sep 16 04:48:42.933747 kernel: smp: Brought up 1 node, 4 CPUs Sep 16 04:48:42.933757 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Sep 16 04:48:42.933767 kernel: Memory: 2422676K/2565800K available (14336K kernel code, 2432K rwdata, 9992K rodata, 54096K init, 2868K bss, 137196K reserved, 0K cma-reserved) Sep 16 04:48:42.933778 kernel: devtmpfs: initialized Sep 16 04:48:42.933789 kernel: x86/mm: Memory block size: 128MB Sep 16 04:48:42.933799 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Sep 16 04:48:42.933810 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Sep 16 04:48:42.933825 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Sep 16 04:48:42.933836 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Sep 16 04:48:42.933847 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Sep 16 04:48:42.933858 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Sep 16 04:48:42.933869 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 16 04:48:42.933880 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 16 04:48:42.933894 kernel: pinctrl core: initialized pinctrl subsystem Sep 16 04:48:42.933905 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 16 04:48:42.933914 kernel: audit: initializing netlink subsys (disabled) Sep 16 04:48:42.933925 kernel: audit: type=2000 audit(1757998120.082:1): state=initialized audit_enabled=0 res=1 Sep 16 04:48:42.933933 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 16 04:48:42.933941 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 16 04:48:42.933949 kernel: cpuidle: using governor menu Sep 16 04:48:42.933958 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 16 04:48:42.933966 kernel: dca service started, version 1.12.1 Sep 16 04:48:42.933974 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Sep 16 04:48:42.933982 kernel: PCI: Using configuration type 1 for base access Sep 16 04:48:42.933990 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 16 04:48:42.934001 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 16 04:48:42.934009 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 16 04:48:42.934017 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 16 04:48:42.934025 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 16 04:48:42.934032 kernel: ACPI: Added _OSI(Module Device) Sep 16 04:48:42.934040 kernel: ACPI: Added _OSI(Processor Device) Sep 16 04:48:42.934048 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 16 04:48:42.934056 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 16 04:48:42.934064 kernel: ACPI: Interpreter enabled Sep 16 04:48:42.934074 kernel: ACPI: PM: (supports S0 S3 S5) Sep 16 04:48:42.934082 kernel: ACPI: Using IOAPIC for interrupt routing Sep 16 04:48:42.934090 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 16 04:48:42.934098 kernel: PCI: Using E820 reservations for host bridge windows Sep 16 04:48:42.934106 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 16 04:48:42.934114 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 16 04:48:42.934382 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 16 04:48:42.934527 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 16 04:48:42.934656 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 16 04:48:42.934667 kernel: PCI host bridge to bus 0000:00 Sep 16 04:48:42.934804 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 16 04:48:42.935001 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 16 04:48:42.935182 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 16 04:48:42.935332 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Sep 16 04:48:42.935453 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Sep 16 04:48:42.935589 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Sep 16 04:48:42.935704 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 16 04:48:42.935864 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Sep 16 04:48:42.936092 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Sep 16 04:48:42.936276 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Sep 16 04:48:42.936411 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Sep 16 04:48:42.936613 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Sep 16 04:48:42.936789 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 16 04:48:42.936933 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 16 04:48:42.937058 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Sep 16 04:48:42.937181 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Sep 16 04:48:42.937342 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Sep 16 04:48:42.937496 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Sep 16 04:48:42.937627 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Sep 16 04:48:42.937751 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Sep 16 04:48:42.937923 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Sep 16 04:48:42.938140 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Sep 16 04:48:42.938302 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Sep 16 04:48:42.938432 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Sep 16 04:48:42.938572 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Sep 16 04:48:42.938719 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Sep 16 04:48:42.938989 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Sep 16 04:48:42.939165 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 16 04:48:42.939367 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Sep 16 04:48:42.939509 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Sep 16 04:48:42.939642 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Sep 16 04:48:42.939801 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Sep 16 04:48:42.939928 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Sep 16 04:48:42.939940 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 16 04:48:42.939948 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 16 04:48:42.939957 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 16 04:48:42.939965 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 16 04:48:42.939973 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 16 04:48:42.939981 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 16 04:48:42.939993 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 16 04:48:42.940001 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 16 04:48:42.940009 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 16 04:48:42.940018 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 16 04:48:42.940026 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 16 04:48:42.940034 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 16 04:48:42.940042 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 16 04:48:42.940050 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 16 04:48:42.940058 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 16 04:48:42.940069 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 16 04:48:42.940077 kernel: iommu: Default domain type: Translated Sep 16 04:48:42.940085 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 16 04:48:42.940093 kernel: efivars: Registered efivars operations Sep 16 04:48:42.940101 kernel: PCI: Using ACPI for IRQ routing Sep 16 04:48:42.940110 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 16 04:48:42.940118 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Sep 16 04:48:42.940126 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Sep 16 04:48:42.940134 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Sep 16 04:48:42.940144 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Sep 16 04:48:42.940153 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Sep 16 04:48:42.940161 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Sep 16 04:48:42.940169 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Sep 16 04:48:42.940177 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Sep 16 04:48:42.940340 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 16 04:48:42.940538 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 16 04:48:42.940692 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 16 04:48:42.940712 kernel: vgaarb: loaded Sep 16 04:48:42.940723 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 16 04:48:42.940734 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 16 04:48:42.940744 kernel: clocksource: Switched to clocksource kvm-clock Sep 16 04:48:42.940755 kernel: VFS: Disk quotas dquot_6.6.0 Sep 16 04:48:42.940765 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 16 04:48:42.940776 kernel: pnp: PnP ACPI init Sep 16 04:48:42.941123 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Sep 16 04:48:42.941148 kernel: pnp: PnP ACPI: found 6 devices Sep 16 04:48:42.941159 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 16 04:48:42.941170 kernel: NET: Registered PF_INET protocol family Sep 16 04:48:42.941181 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 16 04:48:42.941195 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 16 04:48:42.941206 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 16 04:48:42.941218 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 16 04:48:42.941247 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 16 04:48:42.941260 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 16 04:48:42.941275 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 16 04:48:42.941286 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 16 04:48:42.941297 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 16 04:48:42.941307 kernel: NET: Registered PF_XDP protocol family Sep 16 04:48:42.941497 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Sep 16 04:48:42.941670 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Sep 16 04:48:42.941842 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 16 04:48:42.942013 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 16 04:48:42.942159 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 16 04:48:42.942345 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Sep 16 04:48:42.942530 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Sep 16 04:48:42.942680 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Sep 16 04:48:42.942698 kernel: PCI: CLS 0 bytes, default 64 Sep 16 04:48:42.942710 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848e100549, max_idle_ns: 440795215505 ns Sep 16 04:48:42.942721 kernel: Initialise system trusted keyrings Sep 16 04:48:42.942738 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 16 04:48:42.942756 kernel: Key type asymmetric registered Sep 16 04:48:42.942767 kernel: Asymmetric key parser 'x509' registered Sep 16 04:48:42.942778 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 16 04:48:42.942790 kernel: io scheduler mq-deadline registered Sep 16 04:48:42.942801 kernel: io scheduler kyber registered Sep 16 04:48:42.942812 kernel: io scheduler bfq registered Sep 16 04:48:42.942828 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 16 04:48:42.942840 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 16 04:48:42.942852 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 16 04:48:42.942863 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 16 04:48:42.942874 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 16 04:48:42.942886 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 16 04:48:42.942898 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 16 04:48:42.942909 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 16 04:48:42.942921 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 16 04:48:42.943119 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 16 04:48:42.943138 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 16 04:48:42.943315 kernel: rtc_cmos 00:04: registered as rtc0 Sep 16 04:48:42.943479 kernel: rtc_cmos 00:04: setting system clock to 2025-09-16T04:48:42 UTC (1757998122) Sep 16 04:48:42.943633 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Sep 16 04:48:42.943651 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 16 04:48:42.943663 kernel: efifb: probing for efifb Sep 16 04:48:42.943680 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Sep 16 04:48:42.943698 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Sep 16 04:48:42.943709 kernel: efifb: scrolling: redraw Sep 16 04:48:42.943721 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 16 04:48:42.943732 kernel: Console: switching to colour frame buffer device 160x50 Sep 16 04:48:42.943743 kernel: fb0: EFI VGA frame buffer device Sep 16 04:48:42.943755 kernel: pstore: Using crash dump compression: deflate Sep 16 04:48:42.943767 kernel: pstore: Registered efi_pstore as persistent store backend Sep 16 04:48:42.943778 kernel: NET: Registered PF_INET6 protocol family Sep 16 04:48:42.943790 kernel: Segment Routing with IPv6 Sep 16 04:48:42.943806 kernel: In-situ OAM (IOAM) with IPv6 Sep 16 04:48:42.943818 kernel: NET: Registered PF_PACKET protocol family Sep 16 04:48:42.943829 kernel: Key type dns_resolver registered Sep 16 04:48:42.943840 kernel: IPI shorthand broadcast: enabled Sep 16 04:48:42.943851 kernel: sched_clock: Marking stable (3747003415, 182953759)->(3977974701, -48017527) Sep 16 04:48:42.943862 kernel: registered taskstats version 1 Sep 16 04:48:42.943874 kernel: Loading compiled-in X.509 certificates Sep 16 04:48:42.943885 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.47-flatcar: d1d5b0d56b9b23dabf19e645632ff93bf659b3bf' Sep 16 04:48:42.943897 kernel: Demotion targets for Node 0: null Sep 16 04:48:42.943912 kernel: Key type .fscrypt registered Sep 16 04:48:42.943923 kernel: Key type fscrypt-provisioning registered Sep 16 04:48:42.943935 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 16 04:48:42.943947 kernel: ima: Allocated hash algorithm: sha1 Sep 16 04:48:42.943958 kernel: ima: No architecture policies found Sep 16 04:48:42.943969 kernel: clk: Disabling unused clocks Sep 16 04:48:42.943981 kernel: Warning: unable to open an initial console. Sep 16 04:48:42.943992 kernel: Freeing unused kernel image (initmem) memory: 54096K Sep 16 04:48:42.944004 kernel: Write protecting the kernel read-only data: 24576k Sep 16 04:48:42.944020 kernel: Freeing unused kernel image (rodata/data gap) memory: 248K Sep 16 04:48:42.944032 kernel: Run /init as init process Sep 16 04:48:42.944043 kernel: with arguments: Sep 16 04:48:42.944055 kernel: /init Sep 16 04:48:42.944066 kernel: with environment: Sep 16 04:48:42.944077 kernel: HOME=/ Sep 16 04:48:42.944088 kernel: TERM=linux Sep 16 04:48:42.944100 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 16 04:48:42.944113 systemd[1]: Successfully made /usr/ read-only. Sep 16 04:48:42.944133 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 16 04:48:42.944146 systemd[1]: Detected virtualization kvm. Sep 16 04:48:42.944159 systemd[1]: Detected architecture x86-64. Sep 16 04:48:42.944171 systemd[1]: Running in initrd. Sep 16 04:48:42.944183 systemd[1]: No hostname configured, using default hostname. Sep 16 04:48:42.944195 systemd[1]: Hostname set to . Sep 16 04:48:42.944207 systemd[1]: Initializing machine ID from VM UUID. Sep 16 04:48:42.944222 systemd[1]: Queued start job for default target initrd.target. Sep 16 04:48:42.944273 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 16 04:48:42.944298 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 16 04:48:42.944312 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 16 04:48:42.944325 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 16 04:48:42.944337 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 16 04:48:42.944351 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 16 04:48:42.944370 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 16 04:48:42.944382 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 16 04:48:42.944395 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 16 04:48:42.944407 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 16 04:48:42.944448 systemd[1]: Reached target paths.target - Path Units. Sep 16 04:48:42.944471 systemd[1]: Reached target slices.target - Slice Units. Sep 16 04:48:42.944484 systemd[1]: Reached target swap.target - Swaps. Sep 16 04:48:42.944496 systemd[1]: Reached target timers.target - Timer Units. Sep 16 04:48:42.944512 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 16 04:48:42.944524 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 16 04:48:42.944536 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 16 04:48:42.944548 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 16 04:48:42.944561 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 16 04:48:42.944573 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 16 04:48:42.944585 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 16 04:48:42.944598 systemd[1]: Reached target sockets.target - Socket Units. Sep 16 04:48:42.944610 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 16 04:48:42.944625 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 16 04:48:42.944637 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 16 04:48:42.944650 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 16 04:48:42.944662 systemd[1]: Starting systemd-fsck-usr.service... Sep 16 04:48:42.944675 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 16 04:48:42.944687 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 16 04:48:42.944699 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:42.944712 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 16 04:48:42.944731 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 16 04:48:42.944743 systemd[1]: Finished systemd-fsck-usr.service. Sep 16 04:48:42.944794 systemd-journald[219]: Collecting audit messages is disabled. Sep 16 04:48:42.944829 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 16 04:48:42.944843 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:42.944855 systemd-journald[219]: Journal started Sep 16 04:48:42.944887 systemd-journald[219]: Runtime Journal (/run/log/journal/117f8f80422942c6b9fb53f6e9a328ca) is 6M, max 48.4M, 42.4M free. Sep 16 04:48:42.936684 systemd-modules-load[222]: Inserted module 'overlay' Sep 16 04:48:42.950888 systemd[1]: Started systemd-journald.service - Journal Service. Sep 16 04:48:42.952123 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 16 04:48:42.957676 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 16 04:48:42.959330 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 16 04:48:42.966390 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 16 04:48:42.968921 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 16 04:48:42.971269 kernel: Bridge firewalling registered Sep 16 04:48:42.970891 systemd-modules-load[222]: Inserted module 'br_netfilter' Sep 16 04:48:42.976434 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 16 04:48:42.978611 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 16 04:48:42.989454 systemd-tmpfiles[240]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 16 04:48:42.990338 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 16 04:48:42.992549 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 16 04:48:42.994855 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 16 04:48:43.008446 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 16 04:48:43.009372 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 16 04:48:43.014455 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 16 04:48:43.027291 dracut-cmdline[259]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=0b876f86a632750e9937176808a48c2452d5168964273bcfc3c72f2a26140c06 Sep 16 04:48:43.075996 systemd-resolved[262]: Positive Trust Anchors: Sep 16 04:48:43.076015 systemd-resolved[262]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 16 04:48:43.076054 systemd-resolved[262]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 16 04:48:43.079480 systemd-resolved[262]: Defaulting to hostname 'linux'. Sep 16 04:48:43.081081 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 16 04:48:43.086055 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 16 04:48:43.146297 kernel: SCSI subsystem initialized Sep 16 04:48:43.159891 kernel: Loading iSCSI transport class v2.0-870. Sep 16 04:48:43.176294 kernel: iscsi: registered transport (tcp) Sep 16 04:48:43.205375 kernel: iscsi: registered transport (qla4xxx) Sep 16 04:48:43.205487 kernel: QLogic iSCSI HBA Driver Sep 16 04:48:43.236129 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 16 04:48:43.263912 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 16 04:48:43.266263 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 16 04:48:43.329380 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 16 04:48:43.332875 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 16 04:48:43.401294 kernel: raid6: avx2x4 gen() 29361 MB/s Sep 16 04:48:43.418272 kernel: raid6: avx2x2 gen() 29185 MB/s Sep 16 04:48:43.444261 kernel: raid6: avx2x1 gen() 22262 MB/s Sep 16 04:48:43.444301 kernel: raid6: using algorithm avx2x4 gen() 29361 MB/s Sep 16 04:48:43.461431 kernel: raid6: .... xor() 6755 MB/s, rmw enabled Sep 16 04:48:43.461533 kernel: raid6: using avx2x2 recovery algorithm Sep 16 04:48:43.506279 kernel: xor: automatically using best checksumming function avx Sep 16 04:48:43.725383 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 16 04:48:43.736030 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 16 04:48:43.754988 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 16 04:48:43.790351 systemd-udevd[472]: Using default interface naming scheme 'v255'. Sep 16 04:48:43.796766 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 16 04:48:43.801819 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 16 04:48:43.841531 dracut-pre-trigger[480]: rd.md=0: removing MD RAID activation Sep 16 04:48:43.885584 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 16 04:48:43.889068 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 16 04:48:44.018907 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 16 04:48:44.022995 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 16 04:48:44.066263 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 16 04:48:44.070712 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 16 04:48:44.080212 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 16 04:48:44.080363 kernel: GPT:9289727 != 19775487 Sep 16 04:48:44.080390 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 16 04:48:44.080420 kernel: GPT:9289727 != 19775487 Sep 16 04:48:44.080458 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 16 04:48:44.080482 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 16 04:48:44.082260 kernel: cryptd: max_cpu_qlen set to 1000 Sep 16 04:48:44.102262 kernel: AES CTR mode by8 optimization enabled Sep 16 04:48:44.113697 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Sep 16 04:48:44.129272 kernel: libata version 3.00 loaded. Sep 16 04:48:44.139278 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 16 04:48:44.140938 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:44.147797 kernel: ahci 0000:00:1f.2: version 3.0 Sep 16 04:48:44.148078 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 16 04:48:44.145008 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:44.157290 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Sep 16 04:48:44.157588 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Sep 16 04:48:44.157785 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 16 04:48:44.151674 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:44.160007 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 16 04:48:44.172272 kernel: scsi host0: ahci Sep 16 04:48:44.184256 kernel: scsi host1: ahci Sep 16 04:48:44.188791 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 16 04:48:44.198265 kernel: scsi host2: ahci Sep 16 04:48:44.201340 kernel: scsi host3: ahci Sep 16 04:48:44.203667 kernel: scsi host4: ahci Sep 16 04:48:44.205295 kernel: scsi host5: ahci Sep 16 04:48:44.205600 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 16 04:48:44.214807 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 1 Sep 16 04:48:44.214833 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 1 Sep 16 04:48:44.214850 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 1 Sep 16 04:48:44.214865 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 1 Sep 16 04:48:44.214891 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 1 Sep 16 04:48:44.214906 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 1 Sep 16 04:48:44.243174 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 16 04:48:44.257143 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 16 04:48:44.258643 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 16 04:48:44.273887 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 16 04:48:44.274623 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 16 04:48:44.274716 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:44.280406 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:44.285288 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:44.289196 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 16 04:48:44.322896 disk-uuid[634]: Primary Header is updated. Sep 16 04:48:44.322896 disk-uuid[634]: Secondary Entries is updated. Sep 16 04:48:44.322896 disk-uuid[634]: Secondary Header is updated. Sep 16 04:48:44.327288 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 16 04:48:44.332290 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 16 04:48:44.338995 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:44.524330 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 16 04:48:44.524465 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 16 04:48:44.524482 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 16 04:48:44.524496 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 16 04:48:44.526291 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 16 04:48:44.526420 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 16 04:48:44.527292 kernel: ata3.00: LPM support broken, forcing max_power Sep 16 04:48:44.528514 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 16 04:48:44.528539 kernel: ata3.00: applying bridge limits Sep 16 04:48:44.529848 kernel: ata3.00: LPM support broken, forcing max_power Sep 16 04:48:44.529893 kernel: ata3.00: configured for UDMA/100 Sep 16 04:48:44.531373 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 16 04:48:44.594398 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 16 04:48:44.594854 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 16 04:48:44.616273 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 16 04:48:44.974813 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 16 04:48:44.977262 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 16 04:48:44.978631 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 16 04:48:44.981045 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 16 04:48:44.984652 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 16 04:48:45.027838 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 16 04:48:45.336861 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 16 04:48:45.336993 disk-uuid[637]: The operation has completed successfully. Sep 16 04:48:45.386146 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 16 04:48:45.386350 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 16 04:48:45.426536 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 16 04:48:45.460049 sh[668]: Success Sep 16 04:48:45.488287 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 16 04:48:45.488386 kernel: device-mapper: uevent: version 1.0.3 Sep 16 04:48:45.490311 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 16 04:48:45.502329 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Sep 16 04:48:45.549084 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 16 04:48:45.552527 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 16 04:48:45.570505 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 16 04:48:45.578991 kernel: BTRFS: device fsid f1b91845-3914-4d21-a370-6d760ee45b2e devid 1 transid 36 /dev/mapper/usr (253:0) scanned by mount (680) Sep 16 04:48:45.579056 kernel: BTRFS info (device dm-0): first mount of filesystem f1b91845-3914-4d21-a370-6d760ee45b2e Sep 16 04:48:45.579072 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 16 04:48:45.585441 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 16 04:48:45.585505 kernel: BTRFS info (device dm-0): enabling free space tree Sep 16 04:48:45.586869 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 16 04:48:45.588718 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 16 04:48:45.590534 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 16 04:48:45.591529 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 16 04:48:45.593559 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 16 04:48:45.624269 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (712) Sep 16 04:48:45.624337 kernel: BTRFS info (device vda6): first mount of filesystem 8b047ef5-4757-404a-b211-2a505a425364 Sep 16 04:48:45.626264 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 16 04:48:45.630266 kernel: BTRFS info (device vda6): turning on async discard Sep 16 04:48:45.630298 kernel: BTRFS info (device vda6): enabling free space tree Sep 16 04:48:45.637283 kernel: BTRFS info (device vda6): last unmount of filesystem 8b047ef5-4757-404a-b211-2a505a425364 Sep 16 04:48:45.639125 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 16 04:48:45.640603 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 16 04:48:45.800767 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 16 04:48:45.806770 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 16 04:48:45.875371 systemd-networkd[855]: lo: Link UP Sep 16 04:48:45.875386 systemd-networkd[855]: lo: Gained carrier Sep 16 04:48:45.878110 systemd-networkd[855]: Enumeration completed Sep 16 04:48:45.878624 systemd-networkd[855]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 16 04:48:45.878629 systemd-networkd[855]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 16 04:48:45.878840 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 16 04:48:45.882194 systemd-networkd[855]: eth0: Link UP Sep 16 04:48:45.882414 systemd-networkd[855]: eth0: Gained carrier Sep 16 04:48:45.882428 systemd-networkd[855]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 16 04:48:45.882943 systemd[1]: Reached target network.target - Network. Sep 16 04:48:45.903374 systemd-networkd[855]: eth0: DHCPv4 address 10.0.0.34/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 16 04:48:45.918623 ignition[754]: Ignition 2.22.0 Sep 16 04:48:45.918643 ignition[754]: Stage: fetch-offline Sep 16 04:48:45.918703 ignition[754]: no configs at "/usr/lib/ignition/base.d" Sep 16 04:48:45.918744 ignition[754]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 16 04:48:45.919470 ignition[754]: parsed url from cmdline: "" Sep 16 04:48:45.919476 ignition[754]: no config URL provided Sep 16 04:48:45.919489 ignition[754]: reading system config file "/usr/lib/ignition/user.ign" Sep 16 04:48:45.919501 ignition[754]: no config at "/usr/lib/ignition/user.ign" Sep 16 04:48:45.919534 ignition[754]: op(1): [started] loading QEMU firmware config module Sep 16 04:48:45.919541 ignition[754]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 16 04:48:45.932040 ignition[754]: op(1): [finished] loading QEMU firmware config module Sep 16 04:48:45.971949 ignition[754]: parsing config with SHA512: be353244d25b02cd31c9ac62b94cb2301251b51322b09e324a14945838e2745b9ebd224930d475dcdf9b01d9e41851a01f8a3a6fb1ed92620cbdebbd39576556 Sep 16 04:48:45.977182 unknown[754]: fetched base config from "system" Sep 16 04:48:45.977199 unknown[754]: fetched user config from "qemu" Sep 16 04:48:45.977630 ignition[754]: fetch-offline: fetch-offline passed Sep 16 04:48:45.981623 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 16 04:48:45.977694 ignition[754]: Ignition finished successfully Sep 16 04:48:46.002475 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 16 04:48:46.003514 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 16 04:48:46.093536 ignition[866]: Ignition 2.22.0 Sep 16 04:48:46.093550 ignition[866]: Stage: kargs Sep 16 04:48:46.093704 ignition[866]: no configs at "/usr/lib/ignition/base.d" Sep 16 04:48:46.093718 ignition[866]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 16 04:48:46.094690 ignition[866]: kargs: kargs passed Sep 16 04:48:46.094750 ignition[866]: Ignition finished successfully Sep 16 04:48:46.100041 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 16 04:48:46.102843 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 16 04:48:46.139631 ignition[874]: Ignition 2.22.0 Sep 16 04:48:46.139649 ignition[874]: Stage: disks Sep 16 04:48:46.140465 ignition[874]: no configs at "/usr/lib/ignition/base.d" Sep 16 04:48:46.140496 ignition[874]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 16 04:48:46.144868 ignition[874]: disks: disks passed Sep 16 04:48:46.145656 ignition[874]: Ignition finished successfully Sep 16 04:48:46.150420 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 16 04:48:46.151811 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 16 04:48:46.153757 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 16 04:48:46.154190 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 16 04:48:46.154693 systemd[1]: Reached target sysinit.target - System Initialization. Sep 16 04:48:46.155165 systemd[1]: Reached target basic.target - Basic System. Sep 16 04:48:46.163110 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 16 04:48:46.206074 systemd-fsck[884]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 16 04:48:46.215481 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 16 04:48:46.219995 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 16 04:48:46.344272 kernel: EXT4-fs (vda9): mounted filesystem fb1cb44f-955b-4cd0-8849-33ce3640d547 r/w with ordered data mode. Quota mode: none. Sep 16 04:48:46.345672 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 16 04:48:46.347221 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 16 04:48:46.349968 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 16 04:48:46.351507 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 16 04:48:46.352841 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 16 04:48:46.352911 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 16 04:48:46.352947 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 16 04:48:46.371771 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 16 04:48:46.375985 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 16 04:48:46.379610 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (892) Sep 16 04:48:46.382852 kernel: BTRFS info (device vda6): first mount of filesystem 8b047ef5-4757-404a-b211-2a505a425364 Sep 16 04:48:46.382884 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 16 04:48:46.386592 kernel: BTRFS info (device vda6): turning on async discard Sep 16 04:48:46.386622 kernel: BTRFS info (device vda6): enabling free space tree Sep 16 04:48:46.389136 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 16 04:48:46.437643 initrd-setup-root[916]: cut: /sysroot/etc/passwd: No such file or directory Sep 16 04:48:46.442535 initrd-setup-root[923]: cut: /sysroot/etc/group: No such file or directory Sep 16 04:48:46.447308 initrd-setup-root[930]: cut: /sysroot/etc/shadow: No such file or directory Sep 16 04:48:46.452013 initrd-setup-root[937]: cut: /sysroot/etc/gshadow: No such file or directory Sep 16 04:48:46.553411 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 16 04:48:46.555682 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 16 04:48:46.557946 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 16 04:48:46.623431 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 16 04:48:46.625108 kernel: BTRFS info (device vda6): last unmount of filesystem 8b047ef5-4757-404a-b211-2a505a425364 Sep 16 04:48:46.654450 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 16 04:48:46.678994 ignition[1005]: INFO : Ignition 2.22.0 Sep 16 04:48:46.678994 ignition[1005]: INFO : Stage: mount Sep 16 04:48:46.681202 ignition[1005]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 16 04:48:46.681202 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 16 04:48:46.681202 ignition[1005]: INFO : mount: mount passed Sep 16 04:48:46.681202 ignition[1005]: INFO : Ignition finished successfully Sep 16 04:48:46.688936 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 16 04:48:46.691344 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 16 04:48:46.724604 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 16 04:48:46.752215 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1018) Sep 16 04:48:46.752294 kernel: BTRFS info (device vda6): first mount of filesystem 8b047ef5-4757-404a-b211-2a505a425364 Sep 16 04:48:46.752307 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 16 04:48:46.757431 kernel: BTRFS info (device vda6): turning on async discard Sep 16 04:48:46.757487 kernel: BTRFS info (device vda6): enabling free space tree Sep 16 04:48:46.759595 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 16 04:48:46.810942 ignition[1035]: INFO : Ignition 2.22.0 Sep 16 04:48:46.810942 ignition[1035]: INFO : Stage: files Sep 16 04:48:46.813752 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 16 04:48:46.813752 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 16 04:48:46.813752 ignition[1035]: DEBUG : files: compiled without relabeling support, skipping Sep 16 04:48:46.813752 ignition[1035]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 16 04:48:46.813752 ignition[1035]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 16 04:48:46.821555 ignition[1035]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 16 04:48:46.821555 ignition[1035]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 16 04:48:46.821555 ignition[1035]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 16 04:48:46.821555 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 16 04:48:46.821555 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Sep 16 04:48:46.816788 unknown[1035]: wrote ssh authorized keys file for user: core Sep 16 04:48:46.863414 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 16 04:48:46.995792 systemd-networkd[855]: eth0: Gained IPv6LL Sep 16 04:48:47.072689 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 16 04:48:47.072689 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 16 04:48:47.076556 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 16 04:48:47.078213 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 16 04:48:47.080084 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 16 04:48:47.081746 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 16 04:48:47.083530 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 16 04:48:47.085207 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 16 04:48:47.087008 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 16 04:48:47.130261 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 16 04:48:47.132550 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 16 04:48:47.132550 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 16 04:48:47.138073 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 16 04:48:47.138073 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 16 04:48:47.142612 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Sep 16 04:48:47.483499 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 16 04:48:48.188908 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 16 04:48:48.188908 ignition[1035]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 16 04:48:48.193377 ignition[1035]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 16 04:48:48.196221 ignition[1035]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 16 04:48:48.196221 ignition[1035]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 16 04:48:48.196221 ignition[1035]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 16 04:48:48.201602 ignition[1035]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 16 04:48:48.201602 ignition[1035]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 16 04:48:48.201602 ignition[1035]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 16 04:48:48.201602 ignition[1035]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 16 04:48:48.219731 ignition[1035]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 16 04:48:48.225220 ignition[1035]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 16 04:48:48.227244 ignition[1035]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 16 04:48:48.227244 ignition[1035]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 16 04:48:48.230923 ignition[1035]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 16 04:48:48.230923 ignition[1035]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 16 04:48:48.230923 ignition[1035]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 16 04:48:48.230923 ignition[1035]: INFO : files: files passed Sep 16 04:48:48.230923 ignition[1035]: INFO : Ignition finished successfully Sep 16 04:48:48.234414 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 16 04:48:48.239266 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 16 04:48:48.242691 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 16 04:48:48.263933 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 16 04:48:48.264066 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 16 04:48:48.268932 initrd-setup-root-after-ignition[1064]: grep: /sysroot/oem/oem-release: No such file or directory Sep 16 04:48:48.273126 initrd-setup-root-after-ignition[1066]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 16 04:48:48.274940 initrd-setup-root-after-ignition[1066]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 16 04:48:48.276606 initrd-setup-root-after-ignition[1070]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 16 04:48:48.279719 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 16 04:48:48.282457 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 16 04:48:48.283558 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 16 04:48:48.342419 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 16 04:48:48.342572 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 16 04:48:48.344009 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 16 04:48:48.346914 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 16 04:48:48.347548 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 16 04:48:48.350778 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 16 04:48:48.374728 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 16 04:48:48.379115 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 16 04:48:48.407853 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 16 04:48:48.408316 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 16 04:48:48.410733 systemd[1]: Stopped target timers.target - Timer Units. Sep 16 04:48:48.411093 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 16 04:48:48.411283 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 16 04:48:48.416843 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 16 04:48:48.418911 systemd[1]: Stopped target basic.target - Basic System. Sep 16 04:48:48.419627 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 16 04:48:48.419947 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 16 04:48:48.420301 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 16 04:48:48.425962 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 16 04:48:48.426289 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 16 04:48:48.426671 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 16 04:48:48.432715 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 16 04:48:48.434934 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 16 04:48:48.435275 systemd[1]: Stopped target swap.target - Swaps. Sep 16 04:48:48.435741 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 16 04:48:48.435880 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 16 04:48:48.440779 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 16 04:48:48.441115 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 16 04:48:48.441573 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 16 04:48:48.448276 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 16 04:48:48.450986 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 16 04:48:48.451166 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 16 04:48:48.453892 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 16 04:48:48.454030 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 16 04:48:48.454675 systemd[1]: Stopped target paths.target - Path Units. Sep 16 04:48:48.457294 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 16 04:48:48.462334 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 16 04:48:48.465144 systemd[1]: Stopped target slices.target - Slice Units. Sep 16 04:48:48.465749 systemd[1]: Stopped target sockets.target - Socket Units. Sep 16 04:48:48.466082 systemd[1]: iscsid.socket: Deactivated successfully. Sep 16 04:48:48.466204 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 16 04:48:48.469810 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 16 04:48:48.469940 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 16 04:48:48.471754 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 16 04:48:48.471895 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 16 04:48:48.473619 systemd[1]: ignition-files.service: Deactivated successfully. Sep 16 04:48:48.473729 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 16 04:48:48.478441 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 16 04:48:48.479647 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 16 04:48:48.481068 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 16 04:48:48.481190 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 16 04:48:48.482868 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 16 04:48:48.482970 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 16 04:48:48.493562 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 16 04:48:48.495471 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 16 04:48:48.549404 ignition[1090]: INFO : Ignition 2.22.0 Sep 16 04:48:48.549404 ignition[1090]: INFO : Stage: umount Sep 16 04:48:48.553540 ignition[1090]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 16 04:48:48.553540 ignition[1090]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 16 04:48:48.553540 ignition[1090]: INFO : umount: umount passed Sep 16 04:48:48.553540 ignition[1090]: INFO : Ignition finished successfully Sep 16 04:48:48.556347 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 16 04:48:48.556533 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 16 04:48:48.557980 systemd[1]: Stopped target network.target - Network. Sep 16 04:48:48.559496 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 16 04:48:48.559562 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 16 04:48:48.559841 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 16 04:48:48.559892 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 16 04:48:48.560170 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 16 04:48:48.560245 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 16 04:48:48.560671 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 16 04:48:48.560716 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 16 04:48:48.561099 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 16 04:48:48.561743 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 16 04:48:48.563212 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 16 04:48:48.579973 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 16 04:48:48.581119 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 16 04:48:48.586361 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 16 04:48:48.586785 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 16 04:48:48.586846 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 16 04:48:48.592382 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 16 04:48:48.592716 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 16 04:48:48.592858 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 16 04:48:48.597481 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 16 04:48:48.597959 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 16 04:48:48.598565 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 16 04:48:48.598619 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 16 04:48:48.602832 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 16 04:48:48.603627 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 16 04:48:48.603696 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 16 04:48:48.605791 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 16 04:48:48.605858 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 16 04:48:48.612252 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 16 04:48:48.612337 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 16 04:48:48.612803 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 16 04:48:48.614061 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 16 04:48:48.632063 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 16 04:48:48.640424 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 16 04:48:48.641212 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 16 04:48:48.641278 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 16 04:48:48.643612 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 16 04:48:48.643652 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 16 04:48:48.643943 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 16 04:48:48.643993 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 16 04:48:48.649587 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 16 04:48:48.649641 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 16 04:48:48.650548 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 16 04:48:48.650600 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 16 04:48:48.655823 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 16 04:48:48.656468 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 16 04:48:48.656535 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 16 04:48:48.661582 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 16 04:48:48.661636 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 16 04:48:48.664949 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 16 04:48:48.664999 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 16 04:48:48.668334 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 16 04:48:48.668393 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 16 04:48:48.668733 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 16 04:48:48.668788 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:48.678575 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 16 04:48:48.678736 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 16 04:48:48.679360 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 16 04:48:48.679481 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 16 04:48:48.702967 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 16 04:48:48.703180 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 16 04:48:48.704215 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 16 04:48:48.708278 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 16 04:48:48.708395 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 16 04:48:48.709986 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 16 04:48:48.752719 systemd[1]: Switching root. Sep 16 04:48:48.795772 systemd-journald[219]: Journal stopped Sep 16 04:48:49.993749 systemd-journald[219]: Received SIGTERM from PID 1 (systemd). Sep 16 04:48:49.993851 kernel: SELinux: policy capability network_peer_controls=1 Sep 16 04:48:49.993872 kernel: SELinux: policy capability open_perms=1 Sep 16 04:48:49.993892 kernel: SELinux: policy capability extended_socket_class=1 Sep 16 04:48:49.993907 kernel: SELinux: policy capability always_check_network=0 Sep 16 04:48:49.993922 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 16 04:48:49.993947 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 16 04:48:49.993964 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 16 04:48:49.993980 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 16 04:48:49.993995 kernel: SELinux: policy capability userspace_initial_context=0 Sep 16 04:48:49.994018 kernel: audit: type=1403 audit(1757998129.124:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 16 04:48:49.994035 systemd[1]: Successfully loaded SELinux policy in 68.441ms. Sep 16 04:48:49.994074 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 12.153ms. Sep 16 04:48:49.994096 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 16 04:48:49.994120 systemd[1]: Detected virtualization kvm. Sep 16 04:48:49.994139 systemd[1]: Detected architecture x86-64. Sep 16 04:48:49.994158 systemd[1]: Detected first boot. Sep 16 04:48:49.994181 systemd[1]: Initializing machine ID from VM UUID. Sep 16 04:48:49.994196 zram_generator::config[1135]: No configuration found. Sep 16 04:48:49.994213 kernel: Guest personality initialized and is inactive Sep 16 04:48:49.994254 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 16 04:48:49.994275 kernel: Initialized host personality Sep 16 04:48:49.994299 kernel: NET: Registered PF_VSOCK protocol family Sep 16 04:48:49.994324 systemd[1]: Populated /etc with preset unit settings. Sep 16 04:48:49.994345 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 16 04:48:49.994362 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 16 04:48:49.994386 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 16 04:48:49.994402 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 16 04:48:49.994419 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 16 04:48:49.994441 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 16 04:48:49.994459 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 16 04:48:49.994484 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 16 04:48:49.994501 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 16 04:48:49.994518 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 16 04:48:49.994534 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 16 04:48:49.994550 systemd[1]: Created slice user.slice - User and Session Slice. Sep 16 04:48:49.994566 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 16 04:48:49.994583 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 16 04:48:49.994600 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 16 04:48:49.994616 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 16 04:48:49.994641 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 16 04:48:49.994657 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 16 04:48:49.994676 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 16 04:48:49.994694 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 16 04:48:49.994711 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 16 04:48:49.994726 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 16 04:48:49.994742 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 16 04:48:49.994759 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 16 04:48:49.994784 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 16 04:48:49.994802 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 16 04:48:49.994825 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 16 04:48:49.994841 systemd[1]: Reached target slices.target - Slice Units. Sep 16 04:48:49.994857 systemd[1]: Reached target swap.target - Swaps. Sep 16 04:48:49.994872 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 16 04:48:49.994889 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 16 04:48:49.994905 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 16 04:48:49.994922 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 16 04:48:49.994945 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 16 04:48:49.994961 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 16 04:48:49.994977 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 16 04:48:49.994994 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 16 04:48:49.995010 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 16 04:48:49.995026 systemd[1]: Mounting media.mount - External Media Directory... Sep 16 04:48:49.995044 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:49.995060 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 16 04:48:49.995077 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 16 04:48:49.995101 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 16 04:48:49.995117 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 16 04:48:49.995133 systemd[1]: Reached target machines.target - Containers. Sep 16 04:48:49.995149 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 16 04:48:49.995165 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 16 04:48:49.995181 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 16 04:48:49.995197 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 16 04:48:49.995213 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 16 04:48:49.995255 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 16 04:48:49.995274 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 16 04:48:49.995299 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 16 04:48:49.995316 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 16 04:48:49.995333 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 16 04:48:49.995349 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 16 04:48:49.995365 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 16 04:48:49.995385 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 16 04:48:49.995401 systemd[1]: Stopped systemd-fsck-usr.service. Sep 16 04:48:49.995427 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 16 04:48:49.995446 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 16 04:48:49.995463 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 16 04:48:49.995479 kernel: fuse: init (API version 7.41) Sep 16 04:48:49.995494 kernel: loop: module loaded Sep 16 04:48:49.995511 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 16 04:48:49.995529 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 16 04:48:49.995545 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 16 04:48:49.995571 kernel: ACPI: bus type drm_connector registered Sep 16 04:48:49.995588 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 16 04:48:49.995607 systemd[1]: verity-setup.service: Deactivated successfully. Sep 16 04:48:49.995624 systemd[1]: Stopped verity-setup.service. Sep 16 04:48:49.995649 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:49.995666 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 16 04:48:49.995682 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 16 04:48:49.995728 systemd-journald[1203]: Collecting audit messages is disabled. Sep 16 04:48:49.995760 systemd[1]: Mounted media.mount - External Media Directory. Sep 16 04:48:49.995786 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 16 04:48:49.995804 systemd-journald[1203]: Journal started Sep 16 04:48:49.995833 systemd-journald[1203]: Runtime Journal (/run/log/journal/117f8f80422942c6b9fb53f6e9a328ca) is 6M, max 48.4M, 42.4M free. Sep 16 04:48:49.721685 systemd[1]: Queued start job for default target multi-user.target. Sep 16 04:48:49.735755 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 16 04:48:49.736348 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 16 04:48:49.998469 systemd[1]: Started systemd-journald.service - Journal Service. Sep 16 04:48:49.999617 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 16 04:48:50.000903 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 16 04:48:50.002296 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 16 04:48:50.003823 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 16 04:48:50.005450 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 16 04:48:50.005763 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 16 04:48:50.007361 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 16 04:48:50.007632 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 16 04:48:50.009110 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 16 04:48:50.009644 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 16 04:48:50.011062 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 16 04:48:50.011332 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 16 04:48:50.013081 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 16 04:48:50.013346 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 16 04:48:50.014883 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 16 04:48:50.015126 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 16 04:48:50.016827 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 16 04:48:50.018866 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 16 04:48:50.020716 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 16 04:48:50.022456 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 16 04:48:50.040485 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 16 04:48:50.043749 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 16 04:48:50.049337 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 16 04:48:50.050714 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 16 04:48:50.050766 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 16 04:48:50.053125 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 16 04:48:50.056565 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 16 04:48:50.057934 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 16 04:48:50.062678 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 16 04:48:50.066907 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 16 04:48:50.068410 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 16 04:48:50.071402 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 16 04:48:50.072641 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 16 04:48:50.075547 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 16 04:48:50.078989 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 16 04:48:50.083406 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 16 04:48:50.086949 systemd-journald[1203]: Time spent on flushing to /var/log/journal/117f8f80422942c6b9fb53f6e9a328ca is 23.518ms for 1072 entries. Sep 16 04:48:50.086949 systemd-journald[1203]: System Journal (/var/log/journal/117f8f80422942c6b9fb53f6e9a328ca) is 8M, max 195.6M, 187.6M free. Sep 16 04:48:50.122549 systemd-journald[1203]: Received client request to flush runtime journal. Sep 16 04:48:50.087282 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 16 04:48:50.090332 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 16 04:48:50.102467 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 16 04:48:50.104650 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 16 04:48:50.116630 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 16 04:48:50.124409 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 16 04:48:50.125270 kernel: loop0: detected capacity change from 0 to 128016 Sep 16 04:48:50.127855 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 16 04:48:50.159266 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 16 04:48:50.161705 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 16 04:48:50.167173 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. Sep 16 04:48:50.167197 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. Sep 16 04:48:50.169557 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 16 04:48:50.175675 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 16 04:48:50.180006 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 16 04:48:50.186262 kernel: loop1: detected capacity change from 0 to 110984 Sep 16 04:48:50.217351 kernel: loop2: detected capacity change from 0 to 224512 Sep 16 04:48:50.229353 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 16 04:48:50.234884 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 16 04:48:50.272278 kernel: loop3: detected capacity change from 0 to 128016 Sep 16 04:48:50.279412 systemd-tmpfiles[1276]: ACLs are not supported, ignoring. Sep 16 04:48:50.279438 systemd-tmpfiles[1276]: ACLs are not supported, ignoring. Sep 16 04:48:50.285926 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 16 04:48:50.298297 kernel: loop4: detected capacity change from 0 to 110984 Sep 16 04:48:50.311388 kernel: loop5: detected capacity change from 0 to 224512 Sep 16 04:48:50.320341 (sd-merge)[1278]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 16 04:48:50.321156 (sd-merge)[1278]: Merged extensions into '/usr'. Sep 16 04:48:50.327256 systemd[1]: Reload requested from client PID 1254 ('systemd-sysext') (unit systemd-sysext.service)... Sep 16 04:48:50.327276 systemd[1]: Reloading... Sep 16 04:48:50.430613 zram_generator::config[1301]: No configuration found. Sep 16 04:48:50.644461 ldconfig[1249]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 16 04:48:50.695158 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 16 04:48:50.695347 systemd[1]: Reloading finished in 367 ms. Sep 16 04:48:50.725032 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 16 04:48:50.726891 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 16 04:48:50.743983 systemd[1]: Starting ensure-sysext.service... Sep 16 04:48:50.755326 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 16 04:48:50.777917 systemd[1]: Reload requested from client PID 1342 ('systemctl') (unit ensure-sysext.service)... Sep 16 04:48:50.777935 systemd[1]: Reloading... Sep 16 04:48:50.797829 systemd-tmpfiles[1343]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 16 04:48:50.797884 systemd-tmpfiles[1343]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 16 04:48:50.798405 systemd-tmpfiles[1343]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 16 04:48:50.798794 systemd-tmpfiles[1343]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 16 04:48:50.800218 systemd-tmpfiles[1343]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 16 04:48:50.800650 systemd-tmpfiles[1343]: ACLs are not supported, ignoring. Sep 16 04:48:50.800734 systemd-tmpfiles[1343]: ACLs are not supported, ignoring. Sep 16 04:48:50.807067 systemd-tmpfiles[1343]: Detected autofs mount point /boot during canonicalization of boot. Sep 16 04:48:50.807084 systemd-tmpfiles[1343]: Skipping /boot Sep 16 04:48:50.830203 systemd-tmpfiles[1343]: Detected autofs mount point /boot during canonicalization of boot. Sep 16 04:48:50.830632 systemd-tmpfiles[1343]: Skipping /boot Sep 16 04:48:50.926298 zram_generator::config[1370]: No configuration found. Sep 16 04:48:51.135708 systemd[1]: Reloading finished in 357 ms. Sep 16 04:48:51.182392 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 16 04:48:51.192245 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 16 04:48:51.195369 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 16 04:48:51.198442 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 16 04:48:51.203901 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 16 04:48:51.304042 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 16 04:48:51.307110 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 16 04:48:51.314444 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:51.314702 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 16 04:48:51.316385 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 16 04:48:51.320494 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 16 04:48:51.323816 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 16 04:48:51.325381 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 16 04:48:51.325526 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 16 04:48:51.330507 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 16 04:48:51.334507 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 16 04:48:51.337396 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:51.339339 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 16 04:48:51.341924 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 16 04:48:51.342487 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 16 04:48:51.345717 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 16 04:48:51.346164 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 16 04:48:51.351939 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 16 04:48:51.352748 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 16 04:48:51.364737 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:51.365121 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 16 04:48:51.368508 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 16 04:48:51.373704 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 16 04:48:51.383508 systemd-udevd[1424]: Using default interface naming scheme 'v255'. Sep 16 04:48:51.383645 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 16 04:48:51.385672 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 16 04:48:51.386879 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 16 04:48:51.389731 augenrules[1444]: No rules Sep 16 04:48:51.393444 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 16 04:48:51.395332 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:51.397987 systemd[1]: audit-rules.service: Deactivated successfully. Sep 16 04:48:51.398366 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 16 04:48:51.400174 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 16 04:48:51.402535 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 16 04:48:51.402762 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 16 04:48:51.404525 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 16 04:48:51.406844 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 16 04:48:51.407137 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 16 04:48:51.409333 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 16 04:48:51.409619 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 16 04:48:51.422748 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 16 04:48:51.425427 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 16 04:48:51.430215 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 16 04:48:51.439119 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:51.441812 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 16 04:48:51.443514 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 16 04:48:51.445096 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 16 04:48:51.449755 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 16 04:48:51.452622 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 16 04:48:51.462004 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 16 04:48:51.463562 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 16 04:48:51.463728 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 16 04:48:51.470998 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 16 04:48:51.472369 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 16 04:48:51.472500 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 16 04:48:51.481151 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 16 04:48:51.481530 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 16 04:48:51.483771 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 16 04:48:51.484104 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 16 04:48:51.491194 systemd[1]: Finished ensure-sysext.service. Sep 16 04:48:51.504709 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 16 04:48:51.507041 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 16 04:48:51.507389 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 16 04:48:51.509578 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 16 04:48:51.509959 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 16 04:48:51.515848 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 16 04:48:51.516070 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 16 04:48:51.531438 augenrules[1479]: /sbin/augenrules: No change Sep 16 04:48:51.546520 augenrules[1519]: No rules Sep 16 04:48:51.547626 systemd[1]: audit-rules.service: Deactivated successfully. Sep 16 04:48:51.547976 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 16 04:48:51.563951 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 16 04:48:51.582271 systemd-resolved[1411]: Positive Trust Anchors: Sep 16 04:48:51.582301 systemd-resolved[1411]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 16 04:48:51.582340 systemd-resolved[1411]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 16 04:48:51.593296 systemd-resolved[1411]: Defaulting to hostname 'linux'. Sep 16 04:48:51.599518 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 16 04:48:51.604160 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 16 04:48:51.627537 kernel: mousedev: PS/2 mouse device common for all mice Sep 16 04:48:51.668869 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 16 04:48:51.666187 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 16 04:48:51.675265 kernel: ACPI: button: Power Button [PWRF] Sep 16 04:48:51.677071 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 16 04:48:51.709272 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Sep 16 04:48:51.712557 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 16 04:48:51.712771 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 16 04:48:51.728716 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 16 04:48:51.740365 systemd-networkd[1495]: lo: Link UP Sep 16 04:48:51.740376 systemd-networkd[1495]: lo: Gained carrier Sep 16 04:48:51.743132 systemd-networkd[1495]: Enumeration completed Sep 16 04:48:51.743265 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 16 04:48:51.744720 systemd-networkd[1495]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 16 04:48:51.744732 systemd-networkd[1495]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 16 04:48:51.745476 systemd-networkd[1495]: eth0: Link UP Sep 16 04:48:51.745702 systemd-networkd[1495]: eth0: Gained carrier Sep 16 04:48:51.745739 systemd-networkd[1495]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 16 04:48:51.746224 systemd[1]: Reached target network.target - Network. Sep 16 04:48:51.750862 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 16 04:48:51.753588 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 16 04:48:51.760871 systemd-networkd[1495]: eth0: DHCPv4 address 10.0.0.34/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 16 04:48:53.093379 systemd-resolved[1411]: Clock change detected. Flushing caches. Sep 16 04:48:53.093635 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 16 04:48:53.093727 systemd-timesyncd[1505]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 16 04:48:53.094100 systemd-timesyncd[1505]: Initial clock synchronization to Tue 2025-09-16 04:48:53.093310 UTC. Sep 16 04:48:53.096946 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 16 04:48:53.103387 systemd[1]: Reached target time-set.target - System Time Set. Sep 16 04:48:53.108898 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:53.164201 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 16 04:48:53.164556 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:53.170092 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 16 04:48:53.205765 kernel: kvm_amd: TSC scaling supported Sep 16 04:48:53.205868 kernel: kvm_amd: Nested Virtualization enabled Sep 16 04:48:53.205890 kernel: kvm_amd: Nested Paging enabled Sep 16 04:48:53.206904 kernel: kvm_amd: LBR virtualization supported Sep 16 04:48:53.206946 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 16 04:48:53.208550 kernel: kvm_amd: Virtual GIF supported Sep 16 04:48:53.257563 kernel: EDAC MC: Ver: 3.0.0 Sep 16 04:48:53.262393 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 16 04:48:53.264179 systemd[1]: Reached target sysinit.target - System Initialization. Sep 16 04:48:53.265531 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 16 04:48:53.266995 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 16 04:48:53.268433 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Sep 16 04:48:53.270120 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 16 04:48:53.271570 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 16 04:48:53.273038 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 16 04:48:53.274492 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 16 04:48:53.274559 systemd[1]: Reached target paths.target - Path Units. Sep 16 04:48:53.275636 systemd[1]: Reached target timers.target - Timer Units. Sep 16 04:48:53.277883 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 16 04:48:53.280874 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 16 04:48:53.284565 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 16 04:48:53.286243 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 16 04:48:53.287792 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 16 04:48:53.293960 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 16 04:48:53.318301 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 16 04:48:53.320925 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 16 04:48:53.323220 systemd[1]: Reached target sockets.target - Socket Units. Sep 16 04:48:53.324434 systemd[1]: Reached target basic.target - Basic System. Sep 16 04:48:53.325666 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 16 04:48:53.325704 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 16 04:48:53.327240 systemd[1]: Starting containerd.service - containerd container runtime... Sep 16 04:48:53.329891 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 16 04:48:53.332799 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 16 04:48:53.335985 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 16 04:48:53.339694 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 16 04:48:53.340929 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 16 04:48:53.342894 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Sep 16 04:48:53.345376 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 16 04:48:53.346649 jq[1574]: false Sep 16 04:48:53.348184 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 16 04:48:53.351736 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 16 04:48:53.355479 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 16 04:48:53.357017 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Refreshing passwd entry cache Sep 16 04:48:53.357033 oslogin_cache_refresh[1576]: Refreshing passwd entry cache Sep 16 04:48:53.364754 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 16 04:48:53.364990 extend-filesystems[1575]: Found /dev/vda6 Sep 16 04:48:53.369060 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Failure getting users, quitting Sep 16 04:48:53.369060 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 16 04:48:53.369060 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Refreshing group entry cache Sep 16 04:48:53.368497 oslogin_cache_refresh[1576]: Failure getting users, quitting Sep 16 04:48:53.368558 oslogin_cache_refresh[1576]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 16 04:48:53.368639 oslogin_cache_refresh[1576]: Refreshing group entry cache Sep 16 04:48:53.370207 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 16 04:48:53.371713 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 16 04:48:53.372939 systemd[1]: Starting update-engine.service - Update Engine... Sep 16 04:48:53.375251 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Failure getting groups, quitting Sep 16 04:48:53.375251 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 16 04:48:53.375237 oslogin_cache_refresh[1576]: Failure getting groups, quitting Sep 16 04:48:53.375256 oslogin_cache_refresh[1576]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 16 04:48:53.375833 extend-filesystems[1575]: Found /dev/vda9 Sep 16 04:48:53.378350 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 16 04:48:53.379750 extend-filesystems[1575]: Checking size of /dev/vda9 Sep 16 04:48:53.385479 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 16 04:48:53.387573 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 16 04:48:53.387933 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 16 04:48:53.388352 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Sep 16 04:48:53.388648 jq[1597]: true Sep 16 04:48:53.388704 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Sep 16 04:48:53.390869 systemd[1]: motdgen.service: Deactivated successfully. Sep 16 04:48:53.391235 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 16 04:48:53.395368 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 16 04:48:53.395729 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 16 04:48:53.408819 update_engine[1593]: I20250916 04:48:53.408697 1593 main.cc:92] Flatcar Update Engine starting Sep 16 04:48:53.422940 (ntainerd)[1604]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 16 04:48:53.432881 extend-filesystems[1575]: Resized partition /dev/vda9 Sep 16 04:48:53.437338 extend-filesystems[1614]: resize2fs 1.47.3 (8-Jul-2025) Sep 16 04:48:53.445306 jq[1601]: true Sep 16 04:48:53.502270 tar[1600]: linux-amd64/LICENSE Sep 16 04:48:53.502957 tar[1600]: linux-amd64/helm Sep 16 04:48:53.513418 dbus-daemon[1572]: [system] SELinux support is enabled Sep 16 04:48:53.514117 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 16 04:48:53.520154 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 16 04:48:53.520190 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 16 04:48:53.521366 update_engine[1593]: I20250916 04:48:53.521301 1593 update_check_scheduler.cc:74] Next update check in 2m46s Sep 16 04:48:53.521850 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 16 04:48:53.521879 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 16 04:48:53.523560 systemd[1]: Started update-engine.service - Update Engine. Sep 16 04:48:53.526801 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 16 04:48:53.538549 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 16 04:48:53.757276 systemd-logind[1587]: Watching system buttons on /dev/input/event2 (Power Button) Sep 16 04:48:53.759818 systemd-logind[1587]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 16 04:48:53.767932 systemd-logind[1587]: New seat seat0. Sep 16 04:48:53.790041 systemd[1]: Started systemd-logind.service - User Login Management. Sep 16 04:48:53.820100 locksmithd[1620]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 16 04:48:53.824560 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 16 04:48:53.862830 extend-filesystems[1614]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 16 04:48:53.862830 extend-filesystems[1614]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 16 04:48:53.862830 extend-filesystems[1614]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 16 04:48:53.867912 extend-filesystems[1575]: Resized filesystem in /dev/vda9 Sep 16 04:48:53.866237 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 16 04:48:53.868937 sshd_keygen[1595]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 16 04:48:53.867251 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 16 04:48:53.872634 bash[1634]: Updated "/home/core/.ssh/authorized_keys" Sep 16 04:48:53.873899 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 16 04:48:53.878753 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 16 04:48:53.913386 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 16 04:48:53.918863 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 16 04:48:53.959038 systemd[1]: issuegen.service: Deactivated successfully. Sep 16 04:48:53.959605 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 16 04:48:53.965931 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 16 04:48:54.016087 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 16 04:48:54.022767 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 16 04:48:54.026598 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 16 04:48:54.028287 systemd[1]: Reached target getty.target - Login Prompts. Sep 16 04:48:54.042608 tar[1600]: linux-amd64/README.md Sep 16 04:48:54.071763 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 16 04:48:54.159674 containerd[1604]: time="2025-09-16T04:48:54Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 16 04:48:54.160666 containerd[1604]: time="2025-09-16T04:48:54.160623037Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 16 04:48:54.174745 containerd[1604]: time="2025-09-16T04:48:54.174692187Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="9.327µs" Sep 16 04:48:54.174745 containerd[1604]: time="2025-09-16T04:48:54.174725529Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 16 04:48:54.174745 containerd[1604]: time="2025-09-16T04:48:54.174743353Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 16 04:48:54.174984 containerd[1604]: time="2025-09-16T04:48:54.174957524Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 16 04:48:54.174984 containerd[1604]: time="2025-09-16T04:48:54.174977271Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 16 04:48:54.175038 containerd[1604]: time="2025-09-16T04:48:54.175004192Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175101 containerd[1604]: time="2025-09-16T04:48:54.175075315Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175101 containerd[1604]: time="2025-09-16T04:48:54.175090714Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175448 containerd[1604]: time="2025-09-16T04:48:54.175413900Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175448 containerd[1604]: time="2025-09-16T04:48:54.175433967Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175448 containerd[1604]: time="2025-09-16T04:48:54.175444838Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175568 containerd[1604]: time="2025-09-16T04:48:54.175452783Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175630 containerd[1604]: time="2025-09-16T04:48:54.175599969Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175900 containerd[1604]: time="2025-09-16T04:48:54.175870085Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175924 containerd[1604]: time="2025-09-16T04:48:54.175905291Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 16 04:48:54.175924 containerd[1604]: time="2025-09-16T04:48:54.175914969Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 16 04:48:54.175978 containerd[1604]: time="2025-09-16T04:48:54.175954864Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 16 04:48:54.176336 containerd[1604]: time="2025-09-16T04:48:54.176258824Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 16 04:48:54.176551 containerd[1604]: time="2025-09-16T04:48:54.176491200Z" level=info msg="metadata content store policy set" policy=shared Sep 16 04:48:54.182308 containerd[1604]: time="2025-09-16T04:48:54.182220522Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 16 04:48:54.182308 containerd[1604]: time="2025-09-16T04:48:54.182302155Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182325489Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182345146Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182364352Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182377336Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182392725Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182413193Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182430406Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182444091Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182454902Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 16 04:48:54.182489 containerd[1604]: time="2025-09-16T04:48:54.182468517Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 16 04:48:54.182795 containerd[1604]: time="2025-09-16T04:48:54.182713607Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 16 04:48:54.182795 containerd[1604]: time="2025-09-16T04:48:54.182764793Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 16 04:48:54.182795 containerd[1604]: time="2025-09-16T04:48:54.182783167Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182802634Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182814606Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182825206Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182836437Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182846726Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182860011Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182873637Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 16 04:48:54.182881 containerd[1604]: time="2025-09-16T04:48:54.182884758Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 16 04:48:54.183101 containerd[1604]: time="2025-09-16T04:48:54.182969687Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 16 04:48:54.183101 containerd[1604]: time="2025-09-16T04:48:54.182985597Z" level=info msg="Start snapshots syncer" Sep 16 04:48:54.183101 containerd[1604]: time="2025-09-16T04:48:54.183016555Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 16 04:48:54.183361 containerd[1604]: time="2025-09-16T04:48:54.183300137Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 16 04:48:54.183643 containerd[1604]: time="2025-09-16T04:48:54.183369527Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 16 04:48:54.183643 containerd[1604]: time="2025-09-16T04:48:54.183456099Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 16 04:48:54.183643 containerd[1604]: time="2025-09-16T04:48:54.183629474Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 16 04:48:54.183747 containerd[1604]: time="2025-09-16T04:48:54.183664590Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 16 04:48:54.183747 containerd[1604]: time="2025-09-16T04:48:54.183681632Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 16 04:48:54.183747 containerd[1604]: time="2025-09-16T04:48:54.183693504Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 16 04:48:54.183747 containerd[1604]: time="2025-09-16T04:48:54.183706799Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 16 04:48:54.183747 containerd[1604]: time="2025-09-16T04:48:54.183730804Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 16 04:48:54.183747 containerd[1604]: time="2025-09-16T04:48:54.183747766Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183773514Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183784815Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183806887Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183837023Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183851911Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183861459Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183870456Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183878060Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183887979Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.183985341Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.184079498Z" level=info msg="runtime interface created" Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.184086431Z" level=info msg="created NRI interface" Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.184095077Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.184106098Z" level=info msg="Connect containerd service" Sep 16 04:48:54.184139 containerd[1604]: time="2025-09-16T04:48:54.184130463Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 16 04:48:54.185466 containerd[1604]: time="2025-09-16T04:48:54.185416184Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 16 04:48:54.320180 containerd[1604]: time="2025-09-16T04:48:54.320123460Z" level=info msg="Start subscribing containerd event" Sep 16 04:48:54.320180 containerd[1604]: time="2025-09-16T04:48:54.320188793Z" level=info msg="Start recovering state" Sep 16 04:48:54.320356 containerd[1604]: time="2025-09-16T04:48:54.320344044Z" level=info msg="Start event monitor" Sep 16 04:48:54.320390 containerd[1604]: time="2025-09-16T04:48:54.320370503Z" level=info msg="Start cni network conf syncer for default" Sep 16 04:48:54.320454 containerd[1604]: time="2025-09-16T04:48:54.320346699Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 16 04:48:54.320454 containerd[1604]: time="2025-09-16T04:48:54.320381654Z" level=info msg="Start streaming server" Sep 16 04:48:54.320454 containerd[1604]: time="2025-09-16T04:48:54.320434523Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 16 04:48:54.320454 containerd[1604]: time="2025-09-16T04:48:54.320443741Z" level=info msg="runtime interface starting up..." Sep 16 04:48:54.320454 containerd[1604]: time="2025-09-16T04:48:54.320451405Z" level=info msg="starting plugins..." Sep 16 04:48:54.320618 containerd[1604]: time="2025-09-16T04:48:54.320462917Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 16 04:48:54.320618 containerd[1604]: time="2025-09-16T04:48:54.320474007Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 16 04:48:54.320964 containerd[1604]: time="2025-09-16T04:48:54.320931585Z" level=info msg="containerd successfully booted in 0.161908s" Sep 16 04:48:54.321132 systemd[1]: Started containerd.service - containerd container runtime. Sep 16 04:48:55.092737 systemd-networkd[1495]: eth0: Gained IPv6LL Sep 16 04:48:55.096400 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 16 04:48:55.098721 systemd[1]: Reached target network-online.target - Network is Online. Sep 16 04:48:55.101800 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 16 04:48:55.104520 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:48:55.106925 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 16 04:48:55.133371 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 16 04:48:55.133770 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 16 04:48:55.135976 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 16 04:48:55.138861 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 16 04:48:56.471839 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:48:56.474039 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 16 04:48:56.476521 systemd[1]: Startup finished in 3.820s (kernel) + 6.463s (initrd) + 6.105s (userspace) = 16.388s. Sep 16 04:48:56.486029 (kubelet)[1705]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 16 04:48:57.155073 kubelet[1705]: E0916 04:48:57.154991 1705 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 16 04:48:57.159948 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 16 04:48:57.160162 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 16 04:48:57.160659 systemd[1]: kubelet.service: Consumed 1.833s CPU time, 266.1M memory peak. Sep 16 04:48:57.937201 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 16 04:48:57.939017 systemd[1]: Started sshd@0-10.0.0.34:22-10.0.0.1:48526.service - OpenSSH per-connection server daemon (10.0.0.1:48526). Sep 16 04:48:58.024472 sshd[1718]: Accepted publickey for core from 10.0.0.1 port 48526 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:58.026743 sshd-session[1718]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:58.036415 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 16 04:48:58.038097 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 16 04:48:58.048007 systemd-logind[1587]: New session 1 of user core. Sep 16 04:48:58.065716 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 16 04:48:58.068818 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 16 04:48:58.091763 (systemd)[1723]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 16 04:48:58.094363 systemd-logind[1587]: New session c1 of user core. Sep 16 04:48:58.258576 systemd[1723]: Queued start job for default target default.target. Sep 16 04:48:58.283333 systemd[1723]: Created slice app.slice - User Application Slice. Sep 16 04:48:58.283364 systemd[1723]: Reached target paths.target - Paths. Sep 16 04:48:58.283420 systemd[1723]: Reached target timers.target - Timers. Sep 16 04:48:58.285065 systemd[1723]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 16 04:48:58.296949 systemd[1723]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 16 04:48:58.297082 systemd[1723]: Reached target sockets.target - Sockets. Sep 16 04:48:58.297121 systemd[1723]: Reached target basic.target - Basic System. Sep 16 04:48:58.297165 systemd[1723]: Reached target default.target - Main User Target. Sep 16 04:48:58.297200 systemd[1723]: Startup finished in 194ms. Sep 16 04:48:58.297637 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 16 04:48:58.299455 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 16 04:48:58.363635 systemd[1]: Started sshd@1-10.0.0.34:22-10.0.0.1:48540.service - OpenSSH per-connection server daemon (10.0.0.1:48540). Sep 16 04:48:58.427939 sshd[1734]: Accepted publickey for core from 10.0.0.1 port 48540 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:58.429872 sshd-session[1734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:58.434771 systemd-logind[1587]: New session 2 of user core. Sep 16 04:48:58.444662 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 16 04:48:58.500085 sshd[1737]: Connection closed by 10.0.0.1 port 48540 Sep 16 04:48:58.500529 sshd-session[1734]: pam_unix(sshd:session): session closed for user core Sep 16 04:48:58.513828 systemd[1]: sshd@1-10.0.0.34:22-10.0.0.1:48540.service: Deactivated successfully. Sep 16 04:48:58.515745 systemd[1]: session-2.scope: Deactivated successfully. Sep 16 04:48:58.516587 systemd-logind[1587]: Session 2 logged out. Waiting for processes to exit. Sep 16 04:48:58.519473 systemd[1]: Started sshd@2-10.0.0.34:22-10.0.0.1:48546.service - OpenSSH per-connection server daemon (10.0.0.1:48546). Sep 16 04:48:58.520081 systemd-logind[1587]: Removed session 2. Sep 16 04:48:58.591638 sshd[1743]: Accepted publickey for core from 10.0.0.1 port 48546 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:58.593782 sshd-session[1743]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:58.598666 systemd-logind[1587]: New session 3 of user core. Sep 16 04:48:58.605691 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 16 04:48:58.655899 sshd[1747]: Connection closed by 10.0.0.1 port 48546 Sep 16 04:48:58.656292 sshd-session[1743]: pam_unix(sshd:session): session closed for user core Sep 16 04:48:58.666087 systemd[1]: sshd@2-10.0.0.34:22-10.0.0.1:48546.service: Deactivated successfully. Sep 16 04:48:58.668031 systemd[1]: session-3.scope: Deactivated successfully. Sep 16 04:48:58.668839 systemd-logind[1587]: Session 3 logged out. Waiting for processes to exit. Sep 16 04:48:58.671831 systemd[1]: Started sshd@3-10.0.0.34:22-10.0.0.1:48556.service - OpenSSH per-connection server daemon (10.0.0.1:48556). Sep 16 04:48:58.672677 systemd-logind[1587]: Removed session 3. Sep 16 04:48:58.734834 sshd[1753]: Accepted publickey for core from 10.0.0.1 port 48556 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:58.736548 sshd-session[1753]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:58.741804 systemd-logind[1587]: New session 4 of user core. Sep 16 04:48:58.752657 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 16 04:48:58.812041 sshd[1756]: Connection closed by 10.0.0.1 port 48556 Sep 16 04:48:58.812324 sshd-session[1753]: pam_unix(sshd:session): session closed for user core Sep 16 04:48:58.823903 systemd[1]: sshd@3-10.0.0.34:22-10.0.0.1:48556.service: Deactivated successfully. Sep 16 04:48:58.825877 systemd[1]: session-4.scope: Deactivated successfully. Sep 16 04:48:58.826665 systemd-logind[1587]: Session 4 logged out. Waiting for processes to exit. Sep 16 04:48:58.829628 systemd[1]: Started sshd@4-10.0.0.34:22-10.0.0.1:48560.service - OpenSSH per-connection server daemon (10.0.0.1:48560). Sep 16 04:48:58.830445 systemd-logind[1587]: Removed session 4. Sep 16 04:48:58.891853 sshd[1762]: Accepted publickey for core from 10.0.0.1 port 48560 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:58.893465 sshd-session[1762]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:58.898100 systemd-logind[1587]: New session 5 of user core. Sep 16 04:48:58.907726 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 16 04:48:58.970075 sudo[1767]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 16 04:48:58.970484 sudo[1767]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 16 04:48:58.988623 sudo[1767]: pam_unix(sudo:session): session closed for user root Sep 16 04:48:58.990638 sshd[1766]: Connection closed by 10.0.0.1 port 48560 Sep 16 04:48:58.991078 sshd-session[1762]: pam_unix(sshd:session): session closed for user core Sep 16 04:48:59.006579 systemd[1]: sshd@4-10.0.0.34:22-10.0.0.1:48560.service: Deactivated successfully. Sep 16 04:48:59.009133 systemd[1]: session-5.scope: Deactivated successfully. Sep 16 04:48:59.010889 systemd-logind[1587]: Session 5 logged out. Waiting for processes to exit. Sep 16 04:48:59.016722 systemd[1]: Started sshd@5-10.0.0.34:22-10.0.0.1:48574.service - OpenSSH per-connection server daemon (10.0.0.1:48574). Sep 16 04:48:59.017477 systemd-logind[1587]: Removed session 5. Sep 16 04:48:59.090537 sshd[1773]: Accepted publickey for core from 10.0.0.1 port 48574 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:59.092340 sshd-session[1773]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:59.097213 systemd-logind[1587]: New session 6 of user core. Sep 16 04:48:59.106649 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 16 04:48:59.162957 sudo[1778]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 16 04:48:59.163280 sudo[1778]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 16 04:48:59.259702 sudo[1778]: pam_unix(sudo:session): session closed for user root Sep 16 04:48:59.269851 sudo[1777]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 16 04:48:59.275852 sudo[1777]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 16 04:48:59.289234 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 16 04:48:59.348791 augenrules[1800]: No rules Sep 16 04:48:59.351048 systemd[1]: audit-rules.service: Deactivated successfully. Sep 16 04:48:59.351430 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 16 04:48:59.353024 sudo[1777]: pam_unix(sudo:session): session closed for user root Sep 16 04:48:59.355057 sshd[1776]: Connection closed by 10.0.0.1 port 48574 Sep 16 04:48:59.355478 sshd-session[1773]: pam_unix(sshd:session): session closed for user core Sep 16 04:48:59.373458 systemd[1]: sshd@5-10.0.0.34:22-10.0.0.1:48574.service: Deactivated successfully. Sep 16 04:48:59.375954 systemd[1]: session-6.scope: Deactivated successfully. Sep 16 04:48:59.376961 systemd-logind[1587]: Session 6 logged out. Waiting for processes to exit. Sep 16 04:48:59.380633 systemd[1]: Started sshd@6-10.0.0.34:22-10.0.0.1:48586.service - OpenSSH per-connection server daemon (10.0.0.1:48586). Sep 16 04:48:59.381448 systemd-logind[1587]: Removed session 6. Sep 16 04:48:59.450701 sshd[1809]: Accepted publickey for core from 10.0.0.1 port 48586 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:48:59.452305 sshd-session[1809]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:48:59.458566 systemd-logind[1587]: New session 7 of user core. Sep 16 04:48:59.472686 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 16 04:48:59.529595 sudo[1813]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 16 04:48:59.530048 sudo[1813]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 16 04:49:00.057196 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 16 04:49:00.083228 (dockerd)[1833]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 16 04:49:00.548069 dockerd[1833]: time="2025-09-16T04:49:00.547991942Z" level=info msg="Starting up" Sep 16 04:49:00.549308 dockerd[1833]: time="2025-09-16T04:49:00.549278013Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 16 04:49:00.721086 dockerd[1833]: time="2025-09-16T04:49:00.721011574Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 16 04:49:00.972628 dockerd[1833]: time="2025-09-16T04:49:00.972227804Z" level=info msg="Loading containers: start." Sep 16 04:49:00.984542 kernel: Initializing XFRM netlink socket Sep 16 04:49:01.303009 systemd-networkd[1495]: docker0: Link UP Sep 16 04:49:01.308337 dockerd[1833]: time="2025-09-16T04:49:01.308274211Z" level=info msg="Loading containers: done." Sep 16 04:49:01.331373 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3579224711-merged.mount: Deactivated successfully. Sep 16 04:49:01.333057 dockerd[1833]: time="2025-09-16T04:49:01.332993880Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 16 04:49:01.333152 dockerd[1833]: time="2025-09-16T04:49:01.333114116Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 16 04:49:01.333280 dockerd[1833]: time="2025-09-16T04:49:01.333253858Z" level=info msg="Initializing buildkit" Sep 16 04:49:01.371699 dockerd[1833]: time="2025-09-16T04:49:01.371614855Z" level=info msg="Completed buildkit initialization" Sep 16 04:49:01.379237 dockerd[1833]: time="2025-09-16T04:49:01.379161214Z" level=info msg="Daemon has completed initialization" Sep 16 04:49:01.379398 dockerd[1833]: time="2025-09-16T04:49:01.379295877Z" level=info msg="API listen on /run/docker.sock" Sep 16 04:49:01.379541 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 16 04:49:02.611711 containerd[1604]: time="2025-09-16T04:49:02.611640102Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Sep 16 04:49:03.346982 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1571639670.mount: Deactivated successfully. Sep 16 04:49:04.969216 containerd[1604]: time="2025-09-16T04:49:04.969137318Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:04.969935 containerd[1604]: time="2025-09-16T04:49:04.969900389Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Sep 16 04:49:04.972313 containerd[1604]: time="2025-09-16T04:49:04.972234225Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:04.975531 containerd[1604]: time="2025-09-16T04:49:04.975482706Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:04.976728 containerd[1604]: time="2025-09-16T04:49:04.976668189Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 2.364956784s" Sep 16 04:49:04.976728 containerd[1604]: time="2025-09-16T04:49:04.976715237Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Sep 16 04:49:04.977550 containerd[1604]: time="2025-09-16T04:49:04.977501491Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Sep 16 04:49:06.457890 containerd[1604]: time="2025-09-16T04:49:06.457799660Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:06.458555 containerd[1604]: time="2025-09-16T04:49:06.458515392Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Sep 16 04:49:06.459787 containerd[1604]: time="2025-09-16T04:49:06.459751249Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:06.462580 containerd[1604]: time="2025-09-16T04:49:06.462531242Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:06.463470 containerd[1604]: time="2025-09-16T04:49:06.463443342Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.485894793s" Sep 16 04:49:06.463556 containerd[1604]: time="2025-09-16T04:49:06.463481083Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Sep 16 04:49:06.464155 containerd[1604]: time="2025-09-16T04:49:06.464118688Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Sep 16 04:49:07.232139 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 16 04:49:07.236596 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:49:08.069654 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:08.428936 (kubelet)[2128]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 16 04:49:08.518906 kubelet[2128]: E0916 04:49:08.518817 2128 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 16 04:49:08.526150 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 16 04:49:08.526364 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 16 04:49:08.526943 systemd[1]: kubelet.service: Consumed 1.039s CPU time, 111M memory peak. Sep 16 04:49:09.328477 containerd[1604]: time="2025-09-16T04:49:09.328387261Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:09.329481 containerd[1604]: time="2025-09-16T04:49:09.329430757Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Sep 16 04:49:09.331067 containerd[1604]: time="2025-09-16T04:49:09.330971306Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:09.334632 containerd[1604]: time="2025-09-16T04:49:09.334571887Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:09.335682 containerd[1604]: time="2025-09-16T04:49:09.335598712Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 2.871436453s" Sep 16 04:49:09.335682 containerd[1604]: time="2025-09-16T04:49:09.335680095Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Sep 16 04:49:09.336521 containerd[1604]: time="2025-09-16T04:49:09.336484072Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Sep 16 04:49:10.927869 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2489128335.mount: Deactivated successfully. Sep 16 04:49:12.208402 containerd[1604]: time="2025-09-16T04:49:12.208280600Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:12.209459 containerd[1604]: time="2025-09-16T04:49:12.209415518Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Sep 16 04:49:12.211156 containerd[1604]: time="2025-09-16T04:49:12.211064290Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:12.213210 containerd[1604]: time="2025-09-16T04:49:12.213152785Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:12.213901 containerd[1604]: time="2025-09-16T04:49:12.213811971Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 2.877273316s" Sep 16 04:49:12.213901 containerd[1604]: time="2025-09-16T04:49:12.213871974Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Sep 16 04:49:12.214767 containerd[1604]: time="2025-09-16T04:49:12.214719062Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 16 04:49:12.791317 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1550264473.mount: Deactivated successfully. Sep 16 04:49:13.745661 containerd[1604]: time="2025-09-16T04:49:13.745595736Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:13.746399 containerd[1604]: time="2025-09-16T04:49:13.746371680Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Sep 16 04:49:13.747818 containerd[1604]: time="2025-09-16T04:49:13.747751658Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:13.750792 containerd[1604]: time="2025-09-16T04:49:13.750751372Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:13.751790 containerd[1604]: time="2025-09-16T04:49:13.751748201Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.536988333s" Sep 16 04:49:13.751851 containerd[1604]: time="2025-09-16T04:49:13.751787675Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 16 04:49:13.752340 containerd[1604]: time="2025-09-16T04:49:13.752317779Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 16 04:49:14.182408 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3438269816.mount: Deactivated successfully. Sep 16 04:49:14.189288 containerd[1604]: time="2025-09-16T04:49:14.189230601Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 16 04:49:14.190121 containerd[1604]: time="2025-09-16T04:49:14.190083590Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 16 04:49:14.191298 containerd[1604]: time="2025-09-16T04:49:14.191268141Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 16 04:49:14.193755 containerd[1604]: time="2025-09-16T04:49:14.193665757Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 16 04:49:14.194452 containerd[1604]: time="2025-09-16T04:49:14.194405223Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 442.062116ms" Sep 16 04:49:14.194452 containerd[1604]: time="2025-09-16T04:49:14.194433476Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 16 04:49:14.194984 containerd[1604]: time="2025-09-16T04:49:14.194953451Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Sep 16 04:49:14.734993 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1955761836.mount: Deactivated successfully. Sep 16 04:49:16.414841 containerd[1604]: time="2025-09-16T04:49:16.414761120Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:16.415615 containerd[1604]: time="2025-09-16T04:49:16.415588271Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Sep 16 04:49:16.417244 containerd[1604]: time="2025-09-16T04:49:16.417164416Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:16.420029 containerd[1604]: time="2025-09-16T04:49:16.419959497Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:16.421309 containerd[1604]: time="2025-09-16T04:49:16.421269383Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.226279183s" Sep 16 04:49:16.421309 containerd[1604]: time="2025-09-16T04:49:16.421309278Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Sep 16 04:49:18.731456 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 16 04:49:18.733663 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:49:18.978163 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:18.997855 (kubelet)[2286]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 16 04:49:19.052050 kubelet[2286]: E0916 04:49:19.051958 2286 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 16 04:49:19.056849 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 16 04:49:19.057094 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 16 04:49:19.057598 systemd[1]: kubelet.service: Consumed 265ms CPU time, 109.1M memory peak. Sep 16 04:49:19.322975 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:19.323192 systemd[1]: kubelet.service: Consumed 265ms CPU time, 109.1M memory peak. Sep 16 04:49:19.326103 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:49:19.351538 systemd[1]: Reload requested from client PID 2302 ('systemctl') (unit session-7.scope)... Sep 16 04:49:19.351563 systemd[1]: Reloading... Sep 16 04:49:19.456791 zram_generator::config[2347]: No configuration found. Sep 16 04:49:19.974307 systemd[1]: Reloading finished in 622 ms. Sep 16 04:49:20.050659 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 16 04:49:20.050822 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 16 04:49:20.051297 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:20.051367 systemd[1]: kubelet.service: Consumed 174ms CPU time, 98.4M memory peak. Sep 16 04:49:20.053832 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:49:20.287073 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:20.298894 (kubelet)[2392]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 16 04:49:20.354215 kubelet[2392]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 16 04:49:20.354215 kubelet[2392]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 16 04:49:20.354215 kubelet[2392]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 16 04:49:20.354215 kubelet[2392]: I0916 04:49:20.354157 2392 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 16 04:49:20.570733 kubelet[2392]: I0916 04:49:20.570590 2392 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 16 04:49:20.570733 kubelet[2392]: I0916 04:49:20.570619 2392 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 16 04:49:20.570882 kubelet[2392]: I0916 04:49:20.570861 2392 server.go:954] "Client rotation is on, will bootstrap in background" Sep 16 04:49:20.599733 kubelet[2392]: I0916 04:49:20.599694 2392 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 16 04:49:20.600275 kubelet[2392]: E0916 04:49:20.600233 2392 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.34:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:20.609659 kubelet[2392]: I0916 04:49:20.609613 2392 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 16 04:49:20.619528 kubelet[2392]: I0916 04:49:20.619457 2392 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 16 04:49:20.623183 kubelet[2392]: I0916 04:49:20.623095 2392 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 16 04:49:20.623442 kubelet[2392]: I0916 04:49:20.623171 2392 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 16 04:49:20.623695 kubelet[2392]: I0916 04:49:20.623452 2392 topology_manager.go:138] "Creating topology manager with none policy" Sep 16 04:49:20.623695 kubelet[2392]: I0916 04:49:20.623467 2392 container_manager_linux.go:304] "Creating device plugin manager" Sep 16 04:49:20.623760 kubelet[2392]: I0916 04:49:20.623709 2392 state_mem.go:36] "Initialized new in-memory state store" Sep 16 04:49:20.628000 kubelet[2392]: I0916 04:49:20.627956 2392 kubelet.go:446] "Attempting to sync node with API server" Sep 16 04:49:20.628090 kubelet[2392]: I0916 04:49:20.628064 2392 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 16 04:49:20.628154 kubelet[2392]: I0916 04:49:20.628116 2392 kubelet.go:352] "Adding apiserver pod source" Sep 16 04:49:20.628154 kubelet[2392]: I0916 04:49:20.628140 2392 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 16 04:49:20.633188 kubelet[2392]: I0916 04:49:20.633162 2392 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 16 04:49:20.633262 kubelet[2392]: W0916 04:49:20.633116 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:20.633262 kubelet[2392]: E0916 04:49:20.633235 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:20.633333 kubelet[2392]: W0916 04:49:20.633272 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.34:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:20.633333 kubelet[2392]: E0916 04:49:20.633304 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.34:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:20.633606 kubelet[2392]: I0916 04:49:20.633588 2392 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 16 04:49:20.635179 kubelet[2392]: W0916 04:49:20.635123 2392 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 16 04:49:20.637430 kubelet[2392]: I0916 04:49:20.637397 2392 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 16 04:49:20.637478 kubelet[2392]: I0916 04:49:20.637450 2392 server.go:1287] "Started kubelet" Sep 16 04:49:20.638495 kubelet[2392]: I0916 04:49:20.638444 2392 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 16 04:49:20.640037 kubelet[2392]: I0916 04:49:20.639219 2392 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 16 04:49:20.640037 kubelet[2392]: I0916 04:49:20.639423 2392 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 16 04:49:20.640037 kubelet[2392]: I0916 04:49:20.639574 2392 server.go:479] "Adding debug handlers to kubelet server" Sep 16 04:49:20.640037 kubelet[2392]: I0916 04:49:20.639987 2392 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 16 04:49:20.641453 kubelet[2392]: I0916 04:49:20.640780 2392 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 16 04:49:20.643265 kubelet[2392]: E0916 04:49:20.643240 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 16 04:49:20.643318 kubelet[2392]: I0916 04:49:20.643285 2392 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 16 04:49:20.643639 kubelet[2392]: I0916 04:49:20.643613 2392 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 16 04:49:20.643722 kubelet[2392]: I0916 04:49:20.643690 2392 reconciler.go:26] "Reconciler: start to sync state" Sep 16 04:49:20.644388 kubelet[2392]: W0916 04:49:20.644226 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:20.644388 kubelet[2392]: E0916 04:49:20.644304 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:20.644502 kubelet[2392]: E0916 04:49:20.644442 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.34:6443: connect: connection refused" interval="200ms" Sep 16 04:49:20.644844 kubelet[2392]: E0916 04:49:20.644795 2392 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 16 04:49:20.645132 kubelet[2392]: I0916 04:49:20.645109 2392 factory.go:221] Registration of the systemd container factory successfully Sep 16 04:49:20.645275 kubelet[2392]: I0916 04:49:20.645211 2392 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 16 04:49:20.646596 kubelet[2392]: I0916 04:49:20.646571 2392 factory.go:221] Registration of the containerd container factory successfully Sep 16 04:49:20.646666 kubelet[2392]: E0916 04:49:20.645027 2392 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.34:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.34:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1865aa03489557e9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-16 04:49:20.637417449 +0000 UTC m=+0.333206153,LastTimestamp:2025-09-16 04:49:20.637417449 +0000 UTC m=+0.333206153,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 16 04:49:20.661791 kubelet[2392]: I0916 04:49:20.661584 2392 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 16 04:49:20.663468 kubelet[2392]: I0916 04:49:20.663442 2392 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 16 04:49:20.663548 kubelet[2392]: I0916 04:49:20.663487 2392 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 16 04:49:20.663548 kubelet[2392]: I0916 04:49:20.663533 2392 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 16 04:49:20.663993 kubelet[2392]: I0916 04:49:20.663736 2392 kubelet.go:2382] "Starting kubelet main sync loop" Sep 16 04:49:20.663993 kubelet[2392]: E0916 04:49:20.663810 2392 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 16 04:49:20.664207 kubelet[2392]: W0916 04:49:20.664181 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:20.664306 kubelet[2392]: E0916 04:49:20.664283 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:20.664618 kubelet[2392]: I0916 04:49:20.664536 2392 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 16 04:49:20.664618 kubelet[2392]: I0916 04:49:20.664559 2392 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 16 04:49:20.664618 kubelet[2392]: I0916 04:49:20.664604 2392 state_mem.go:36] "Initialized new in-memory state store" Sep 16 04:49:20.743745 kubelet[2392]: E0916 04:49:20.743706 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 16 04:49:20.764143 kubelet[2392]: E0916 04:49:20.764077 2392 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 16 04:49:20.844550 kubelet[2392]: E0916 04:49:20.844323 2392 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 16 04:49:20.846120 kubelet[2392]: E0916 04:49:20.846087 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.34:6443: connect: connection refused" interval="400ms" Sep 16 04:49:20.886931 kubelet[2392]: I0916 04:49:20.886864 2392 policy_none.go:49] "None policy: Start" Sep 16 04:49:20.886931 kubelet[2392]: I0916 04:49:20.886925 2392 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 16 04:49:20.886931 kubelet[2392]: I0916 04:49:20.886950 2392 state_mem.go:35] "Initializing new in-memory state store" Sep 16 04:49:20.897114 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 16 04:49:20.910740 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 16 04:49:20.914874 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 16 04:49:20.923678 kubelet[2392]: I0916 04:49:20.923565 2392 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 16 04:49:20.923863 kubelet[2392]: I0916 04:49:20.923835 2392 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 16 04:49:20.923897 kubelet[2392]: I0916 04:49:20.923852 2392 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 16 04:49:20.924100 kubelet[2392]: I0916 04:49:20.924083 2392 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 16 04:49:20.925058 kubelet[2392]: E0916 04:49:20.925029 2392 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 16 04:49:20.925120 kubelet[2392]: E0916 04:49:20.925074 2392 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 16 04:49:20.975409 systemd[1]: Created slice kubepods-burstable-pod1403266a9792debaa127cd8df7a81c3c.slice - libcontainer container kubepods-burstable-pod1403266a9792debaa127cd8df7a81c3c.slice. Sep 16 04:49:20.986719 kubelet[2392]: E0916 04:49:20.986646 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:20.989552 systemd[1]: Created slice kubepods-burstable-pod72a30db4fc25e4da65a3b99eba43be94.slice - libcontainer container kubepods-burstable-pod72a30db4fc25e4da65a3b99eba43be94.slice. Sep 16 04:49:21.008986 kubelet[2392]: E0916 04:49:21.008936 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:21.012325 systemd[1]: Created slice kubepods-burstable-podfd193e1910e5dda003618658db09ac54.slice - libcontainer container kubepods-burstable-podfd193e1910e5dda003618658db09ac54.slice. Sep 16 04:49:21.014192 kubelet[2392]: E0916 04:49:21.014162 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:21.025236 kubelet[2392]: I0916 04:49:21.025216 2392 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 16 04:49:21.025639 kubelet[2392]: E0916 04:49:21.025607 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.34:6443/api/v1/nodes\": dial tcp 10.0.0.34:6443: connect: connection refused" node="localhost" Sep 16 04:49:21.046038 kubelet[2392]: I0916 04:49:21.046003 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:21.046096 kubelet[2392]: I0916 04:49:21.046040 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/72a30db4fc25e4da65a3b99eba43be94-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"72a30db4fc25e4da65a3b99eba43be94\") " pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:21.046096 kubelet[2392]: I0916 04:49:21.046067 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fd193e1910e5dda003618658db09ac54-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fd193e1910e5dda003618658db09ac54\") " pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:21.046096 kubelet[2392]: I0916 04:49:21.046092 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fd193e1910e5dda003618658db09ac54-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fd193e1910e5dda003618658db09ac54\") " pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:21.046310 kubelet[2392]: I0916 04:49:21.046116 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fd193e1910e5dda003618658db09ac54-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fd193e1910e5dda003618658db09ac54\") " pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:21.046310 kubelet[2392]: I0916 04:49:21.046138 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:21.046404 kubelet[2392]: I0916 04:49:21.046171 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:21.046465 kubelet[2392]: I0916 04:49:21.046408 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:21.046535 kubelet[2392]: I0916 04:49:21.046450 2392 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:21.227624 kubelet[2392]: I0916 04:49:21.227466 2392 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 16 04:49:21.227922 kubelet[2392]: E0916 04:49:21.227890 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.34:6443/api/v1/nodes\": dial tcp 10.0.0.34:6443: connect: connection refused" node="localhost" Sep 16 04:49:21.246850 kubelet[2392]: E0916 04:49:21.246790 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.34:6443: connect: connection refused" interval="800ms" Sep 16 04:49:21.288298 kubelet[2392]: E0916 04:49:21.288234 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:21.289088 containerd[1604]: time="2025-09-16T04:49:21.289032207Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:1403266a9792debaa127cd8df7a81c3c,Namespace:kube-system,Attempt:0,}" Sep 16 04:49:21.310263 kubelet[2392]: E0916 04:49:21.310233 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:21.310840 containerd[1604]: time="2025-09-16T04:49:21.310759716Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:72a30db4fc25e4da65a3b99eba43be94,Namespace:kube-system,Attempt:0,}" Sep 16 04:49:21.314956 kubelet[2392]: E0916 04:49:21.314930 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:21.315564 containerd[1604]: time="2025-09-16T04:49:21.315521935Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:fd193e1910e5dda003618658db09ac54,Namespace:kube-system,Attempt:0,}" Sep 16 04:49:21.364457 containerd[1604]: time="2025-09-16T04:49:21.364385795Z" level=info msg="connecting to shim 7b48e5cba52c6321aef048b433ea30631b286cd212595968547d85d048250584" address="unix:///run/containerd/s/b21837cfcb748e4b5a262c9b2c70e867516b996b514dcd045156f63b6e9ea78e" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:21.368578 containerd[1604]: time="2025-09-16T04:49:21.368522151Z" level=info msg="connecting to shim ae478c2a03e2a922e2aaf37f7f25ae5cb9ae77563098dc62242701bf588e2830" address="unix:///run/containerd/s/040f552da9facee361052232549b8681e8643b2f1c6a231282f01ba493b22f56" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:21.412549 containerd[1604]: time="2025-09-16T04:49:21.412088938Z" level=info msg="connecting to shim eeed65716202d87ba57926b6312ac43bbfa4691f2cb3b6abd995420da576b204" address="unix:///run/containerd/s/a94e87121f82dd17c1798d3610e3a3f33ee2cf4b882a14237ba8d64a99bc0e80" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:21.445896 systemd[1]: Started cri-containerd-7b48e5cba52c6321aef048b433ea30631b286cd212595968547d85d048250584.scope - libcontainer container 7b48e5cba52c6321aef048b433ea30631b286cd212595968547d85d048250584. Sep 16 04:49:21.558716 kubelet[2392]: W0916 04:49:21.558623 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:21.558716 kubelet[2392]: E0916 04:49:21.558727 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:21.629933 kubelet[2392]: I0916 04:49:21.629882 2392 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 16 04:49:21.630494 kubelet[2392]: E0916 04:49:21.630422 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.34:6443/api/v1/nodes\": dial tcp 10.0.0.34:6443: connect: connection refused" node="localhost" Sep 16 04:49:21.645413 systemd[1]: Started cri-containerd-ae478c2a03e2a922e2aaf37f7f25ae5cb9ae77563098dc62242701bf588e2830.scope - libcontainer container ae478c2a03e2a922e2aaf37f7f25ae5cb9ae77563098dc62242701bf588e2830. Sep 16 04:49:21.648749 systemd[1]: Started cri-containerd-eeed65716202d87ba57926b6312ac43bbfa4691f2cb3b6abd995420da576b204.scope - libcontainer container eeed65716202d87ba57926b6312ac43bbfa4691f2cb3b6abd995420da576b204. Sep 16 04:49:21.832891 kubelet[2392]: W0916 04:49:21.832709 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:21.832891 kubelet[2392]: E0916 04:49:21.832780 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:21.913488 kubelet[2392]: W0916 04:49:21.913375 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:21.913488 kubelet[2392]: E0916 04:49:21.913475 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:21.995320 kubelet[2392]: W0916 04:49:21.995209 2392 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.34:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.34:6443: connect: connection refused Sep 16 04:49:21.995320 kubelet[2392]: E0916 04:49:21.995299 2392 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.34:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.34:6443: connect: connection refused" logger="UnhandledError" Sep 16 04:49:22.047642 containerd[1604]: time="2025-09-16T04:49:22.047587598Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:fd193e1910e5dda003618658db09ac54,Namespace:kube-system,Attempt:0,} returns sandbox id \"7b48e5cba52c6321aef048b433ea30631b286cd212595968547d85d048250584\"" Sep 16 04:49:22.047789 kubelet[2392]: E0916 04:49:22.047677 2392 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.34:6443: connect: connection refused" interval="1.6s" Sep 16 04:49:22.048827 kubelet[2392]: E0916 04:49:22.048790 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:22.049828 containerd[1604]: time="2025-09-16T04:49:22.049791771Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:72a30db4fc25e4da65a3b99eba43be94,Namespace:kube-system,Attempt:0,} returns sandbox id \"eeed65716202d87ba57926b6312ac43bbfa4691f2cb3b6abd995420da576b204\"" Sep 16 04:49:22.050373 kubelet[2392]: E0916 04:49:22.050322 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:22.050697 containerd[1604]: time="2025-09-16T04:49:22.050662343Z" level=info msg="CreateContainer within sandbox \"7b48e5cba52c6321aef048b433ea30631b286cd212595968547d85d048250584\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 16 04:49:22.051772 containerd[1604]: time="2025-09-16T04:49:22.051747558Z" level=info msg="CreateContainer within sandbox \"eeed65716202d87ba57926b6312ac43bbfa4691f2cb3b6abd995420da576b204\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 16 04:49:22.053256 containerd[1604]: time="2025-09-16T04:49:22.053210832Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:1403266a9792debaa127cd8df7a81c3c,Namespace:kube-system,Attempt:0,} returns sandbox id \"ae478c2a03e2a922e2aaf37f7f25ae5cb9ae77563098dc62242701bf588e2830\"" Sep 16 04:49:22.053718 kubelet[2392]: E0916 04:49:22.053690 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:22.055075 containerd[1604]: time="2025-09-16T04:49:22.055047005Z" level=info msg="CreateContainer within sandbox \"ae478c2a03e2a922e2aaf37f7f25ae5cb9ae77563098dc62242701bf588e2830\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 16 04:49:22.067672 containerd[1604]: time="2025-09-16T04:49:22.067593439Z" level=info msg="Container 2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:22.071299 containerd[1604]: time="2025-09-16T04:49:22.071253552Z" level=info msg="Container d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:22.075195 containerd[1604]: time="2025-09-16T04:49:22.075142694Z" level=info msg="Container b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:22.138741 containerd[1604]: time="2025-09-16T04:49:22.138555583Z" level=info msg="CreateContainer within sandbox \"eeed65716202d87ba57926b6312ac43bbfa4691f2cb3b6abd995420da576b204\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95\"" Sep 16 04:49:22.139771 containerd[1604]: time="2025-09-16T04:49:22.139577830Z" level=info msg="StartContainer for \"d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95\"" Sep 16 04:49:22.142089 containerd[1604]: time="2025-09-16T04:49:22.142039596Z" level=info msg="connecting to shim d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95" address="unix:///run/containerd/s/a94e87121f82dd17c1798d3610e3a3f33ee2cf4b882a14237ba8d64a99bc0e80" protocol=ttrpc version=3 Sep 16 04:49:22.146218 containerd[1604]: time="2025-09-16T04:49:22.146183706Z" level=info msg="CreateContainer within sandbox \"7b48e5cba52c6321aef048b433ea30631b286cd212595968547d85d048250584\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f\"" Sep 16 04:49:22.146756 containerd[1604]: time="2025-09-16T04:49:22.146719791Z" level=info msg="StartContainer for \"2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f\"" Sep 16 04:49:22.147973 containerd[1604]: time="2025-09-16T04:49:22.147945008Z" level=info msg="connecting to shim 2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f" address="unix:///run/containerd/s/b21837cfcb748e4b5a262c9b2c70e867516b996b514dcd045156f63b6e9ea78e" protocol=ttrpc version=3 Sep 16 04:49:22.151527 containerd[1604]: time="2025-09-16T04:49:22.151399886Z" level=info msg="CreateContainer within sandbox \"ae478c2a03e2a922e2aaf37f7f25ae5cb9ae77563098dc62242701bf588e2830\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8\"" Sep 16 04:49:22.153738 containerd[1604]: time="2025-09-16T04:49:22.153708114Z" level=info msg="StartContainer for \"b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8\"" Sep 16 04:49:22.156183 containerd[1604]: time="2025-09-16T04:49:22.156149482Z" level=info msg="connecting to shim b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8" address="unix:///run/containerd/s/040f552da9facee361052232549b8681e8643b2f1c6a231282f01ba493b22f56" protocol=ttrpc version=3 Sep 16 04:49:22.258804 systemd[1]: Started cri-containerd-d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95.scope - libcontainer container d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95. Sep 16 04:49:22.264209 systemd[1]: Started cri-containerd-2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f.scope - libcontainer container 2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f. Sep 16 04:49:22.269045 systemd[1]: Started cri-containerd-b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8.scope - libcontainer container b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8. Sep 16 04:49:22.366124 containerd[1604]: time="2025-09-16T04:49:22.366038303Z" level=info msg="StartContainer for \"b424093cda3d3fafa3fdeac44c0aeef7005989d23c5b21a5220ed1cc0d751ec8\" returns successfully" Sep 16 04:49:22.367026 containerd[1604]: time="2025-09-16T04:49:22.366944713Z" level=info msg="StartContainer for \"2b46c095a92300045f58c9a08d93434703535d7ce18e228b2109d0b7fecd120f\" returns successfully" Sep 16 04:49:22.432851 kubelet[2392]: I0916 04:49:22.432699 2392 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 16 04:49:22.433498 kubelet[2392]: E0916 04:49:22.433463 2392 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.34:6443/api/v1/nodes\": dial tcp 10.0.0.34:6443: connect: connection refused" node="localhost" Sep 16 04:49:22.680719 kubelet[2392]: E0916 04:49:22.680664 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:22.681234 kubelet[2392]: E0916 04:49:22.680871 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:22.684389 kubelet[2392]: E0916 04:49:22.684253 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:22.684389 kubelet[2392]: E0916 04:49:22.684381 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:22.735806 containerd[1604]: time="2025-09-16T04:49:22.735720497Z" level=info msg="StartContainer for \"d94076e9594833e091e469b61b5162250d6ab4d3397795a70cc4ce6570e09a95\" returns successfully" Sep 16 04:49:23.696558 kubelet[2392]: E0916 04:49:23.695610 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:23.696558 kubelet[2392]: E0916 04:49:23.695748 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:23.696558 kubelet[2392]: E0916 04:49:23.695752 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:23.696558 kubelet[2392]: E0916 04:49:23.695886 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:24.035995 kubelet[2392]: I0916 04:49:24.035880 2392 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 16 04:49:24.709395 kubelet[2392]: E0916 04:49:24.709123 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:24.709395 kubelet[2392]: E0916 04:49:24.709308 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:24.710203 kubelet[2392]: E0916 04:49:24.710055 2392 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 16 04:49:24.710203 kubelet[2392]: E0916 04:49:24.710157 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:24.826977 kubelet[2392]: E0916 04:49:24.826863 2392 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 16 04:49:24.861145 kubelet[2392]: E0916 04:49:24.860965 2392 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1865aa03489557e9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-16 04:49:20.637417449 +0000 UTC m=+0.333206153,LastTimestamp:2025-09-16 04:49:20.637417449 +0000 UTC m=+0.333206153,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 16 04:49:24.919578 kubelet[2392]: E0916 04:49:24.919418 2392 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1865aa034905c434 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-16 04:49:20.644785204 +0000 UTC m=+0.340573907,LastTimestamp:2025-09-16 04:49:20.644785204 +0000 UTC m=+0.340573907,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 16 04:49:24.924319 kubelet[2392]: I0916 04:49:24.924186 2392 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 16 04:49:24.944323 kubelet[2392]: I0916 04:49:24.944261 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:24.974629 kubelet[2392]: E0916 04:49:24.974286 2392 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1865aa034a2ce4b9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node localhost status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-16 04:49:20.664126649 +0000 UTC m=+0.359915352,LastTimestamp:2025-09-16 04:49:20.664126649 +0000 UTC m=+0.359915352,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 16 04:49:25.036480 kubelet[2392]: E0916 04:49:25.036360 2392 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1865aa034a2cf887 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node localhost status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-16 04:49:20.664131719 +0000 UTC m=+0.359920422,LastTimestamp:2025-09-16 04:49:20.664131719 +0000 UTC m=+0.359920422,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 16 04:49:25.039778 kubelet[2392]: E0916 04:49:25.039716 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:25.039778 kubelet[2392]: I0916 04:49:25.039772 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:25.044049 kubelet[2392]: E0916 04:49:25.044006 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:25.044049 kubelet[2392]: I0916 04:49:25.044041 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:25.045605 kubelet[2392]: E0916 04:49:25.045539 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:25.632012 kubelet[2392]: I0916 04:49:25.631927 2392 apiserver.go:52] "Watching apiserver" Sep 16 04:49:25.644111 kubelet[2392]: I0916 04:49:25.644057 2392 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 16 04:49:25.707111 kubelet[2392]: I0916 04:49:25.707053 2392 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:25.709324 kubelet[2392]: E0916 04:49:25.709282 2392 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:25.709466 kubelet[2392]: E0916 04:49:25.709443 2392 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:26.737206 systemd[1]: Reload requested from client PID 2668 ('systemctl') (unit session-7.scope)... Sep 16 04:49:26.737224 systemd[1]: Reloading... Sep 16 04:49:26.975621 zram_generator::config[2717]: No configuration found. Sep 16 04:49:27.221142 systemd[1]: Reloading finished in 483 ms. Sep 16 04:49:27.256500 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:49:27.282887 systemd[1]: kubelet.service: Deactivated successfully. Sep 16 04:49:27.283185 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:27.283250 systemd[1]: kubelet.service: Consumed 944ms CPU time, 132M memory peak. Sep 16 04:49:27.285308 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 16 04:49:27.511013 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 16 04:49:27.522910 (kubelet)[2756]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 16 04:49:27.584184 kubelet[2756]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 16 04:49:27.584184 kubelet[2756]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 16 04:49:27.584184 kubelet[2756]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 16 04:49:27.584649 kubelet[2756]: I0916 04:49:27.584218 2756 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 16 04:49:27.592528 kubelet[2756]: I0916 04:49:27.592470 2756 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 16 04:49:27.592528 kubelet[2756]: I0916 04:49:27.592497 2756 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 16 04:49:27.592812 kubelet[2756]: I0916 04:49:27.592781 2756 server.go:954] "Client rotation is on, will bootstrap in background" Sep 16 04:49:27.593995 kubelet[2756]: I0916 04:49:27.593964 2756 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 16 04:49:27.600423 kubelet[2756]: I0916 04:49:27.600385 2756 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 16 04:49:27.604774 kubelet[2756]: I0916 04:49:27.604748 2756 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 16 04:49:27.611259 kubelet[2756]: I0916 04:49:27.611227 2756 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 16 04:49:27.611556 kubelet[2756]: I0916 04:49:27.611493 2756 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 16 04:49:27.611763 kubelet[2756]: I0916 04:49:27.611558 2756 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 16 04:49:27.611866 kubelet[2756]: I0916 04:49:27.611769 2756 topology_manager.go:138] "Creating topology manager with none policy" Sep 16 04:49:27.611866 kubelet[2756]: I0916 04:49:27.611778 2756 container_manager_linux.go:304] "Creating device plugin manager" Sep 16 04:49:27.611866 kubelet[2756]: I0916 04:49:27.611829 2756 state_mem.go:36] "Initialized new in-memory state store" Sep 16 04:49:27.612039 kubelet[2756]: I0916 04:49:27.612000 2756 kubelet.go:446] "Attempting to sync node with API server" Sep 16 04:49:27.612186 kubelet[2756]: I0916 04:49:27.612090 2756 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 16 04:49:27.612186 kubelet[2756]: I0916 04:49:27.612119 2756 kubelet.go:352] "Adding apiserver pod source" Sep 16 04:49:27.612186 kubelet[2756]: I0916 04:49:27.612131 2756 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 16 04:49:27.613678 kubelet[2756]: I0916 04:49:27.613653 2756 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 16 04:49:27.615526 kubelet[2756]: I0916 04:49:27.614091 2756 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 16 04:49:27.615712 kubelet[2756]: I0916 04:49:27.615682 2756 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 16 04:49:27.615763 kubelet[2756]: I0916 04:49:27.615717 2756 server.go:1287] "Started kubelet" Sep 16 04:49:27.617537 kubelet[2756]: I0916 04:49:27.615860 2756 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 16 04:49:27.617537 kubelet[2756]: I0916 04:49:27.616145 2756 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 16 04:49:27.617537 kubelet[2756]: I0916 04:49:27.616436 2756 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 16 04:49:27.619066 kubelet[2756]: I0916 04:49:27.619037 2756 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 16 04:49:27.619972 kubelet[2756]: E0916 04:49:27.619934 2756 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 16 04:49:27.620058 kubelet[2756]: I0916 04:49:27.620044 2756 server.go:479] "Adding debug handlers to kubelet server" Sep 16 04:49:27.624244 kubelet[2756]: I0916 04:49:27.624216 2756 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 16 04:49:27.626597 kubelet[2756]: E0916 04:49:27.626568 2756 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 16 04:49:27.626597 kubelet[2756]: I0916 04:49:27.626600 2756 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 16 04:49:27.626806 kubelet[2756]: I0916 04:49:27.626781 2756 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 16 04:49:27.626947 kubelet[2756]: I0916 04:49:27.626926 2756 reconciler.go:26] "Reconciler: start to sync state" Sep 16 04:49:27.628746 kubelet[2756]: I0916 04:49:27.628716 2756 factory.go:221] Registration of the systemd container factory successfully Sep 16 04:49:27.628925 kubelet[2756]: I0916 04:49:27.628904 2756 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 16 04:49:27.631417 kubelet[2756]: I0916 04:49:27.631383 2756 factory.go:221] Registration of the containerd container factory successfully Sep 16 04:49:27.639259 kubelet[2756]: I0916 04:49:27.639214 2756 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 16 04:49:27.642094 kubelet[2756]: I0916 04:49:27.642063 2756 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 16 04:49:27.642158 kubelet[2756]: I0916 04:49:27.642140 2756 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 16 04:49:27.642270 kubelet[2756]: I0916 04:49:27.642174 2756 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 16 04:49:27.642270 kubelet[2756]: I0916 04:49:27.642269 2756 kubelet.go:2382] "Starting kubelet main sync loop" Sep 16 04:49:27.642637 kubelet[2756]: E0916 04:49:27.642601 2756 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 16 04:49:27.676410 kubelet[2756]: I0916 04:49:27.676370 2756 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 16 04:49:27.676410 kubelet[2756]: I0916 04:49:27.676391 2756 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 16 04:49:27.676410 kubelet[2756]: I0916 04:49:27.676415 2756 state_mem.go:36] "Initialized new in-memory state store" Sep 16 04:49:27.676632 kubelet[2756]: I0916 04:49:27.676598 2756 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 16 04:49:27.676632 kubelet[2756]: I0916 04:49:27.676610 2756 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 16 04:49:27.676632 kubelet[2756]: I0916 04:49:27.676629 2756 policy_none.go:49] "None policy: Start" Sep 16 04:49:27.676711 kubelet[2756]: I0916 04:49:27.676643 2756 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 16 04:49:27.676711 kubelet[2756]: I0916 04:49:27.676657 2756 state_mem.go:35] "Initializing new in-memory state store" Sep 16 04:49:27.676786 kubelet[2756]: I0916 04:49:27.676765 2756 state_mem.go:75] "Updated machine memory state" Sep 16 04:49:27.681247 kubelet[2756]: I0916 04:49:27.681202 2756 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 16 04:49:27.681416 kubelet[2756]: I0916 04:49:27.681389 2756 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 16 04:49:27.681445 kubelet[2756]: I0916 04:49:27.681406 2756 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 16 04:49:27.681596 kubelet[2756]: I0916 04:49:27.681579 2756 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 16 04:49:27.682430 kubelet[2756]: E0916 04:49:27.682404 2756 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 16 04:49:27.743906 kubelet[2756]: I0916 04:49:27.743629 2756 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:27.743906 kubelet[2756]: I0916 04:49:27.743689 2756 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:27.743906 kubelet[2756]: I0916 04:49:27.743720 2756 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:27.786956 kubelet[2756]: I0916 04:49:27.786814 2756 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 16 04:49:27.828608 kubelet[2756]: I0916 04:49:27.828543 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:27.828608 kubelet[2756]: I0916 04:49:27.828588 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:27.828608 kubelet[2756]: I0916 04:49:27.828611 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:27.828847 kubelet[2756]: I0916 04:49:27.828631 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fd193e1910e5dda003618658db09ac54-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fd193e1910e5dda003618658db09ac54\") " pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:27.828847 kubelet[2756]: I0916 04:49:27.828653 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fd193e1910e5dda003618658db09ac54-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fd193e1910e5dda003618658db09ac54\") " pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:27.828847 kubelet[2756]: I0916 04:49:27.828697 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:27.828847 kubelet[2756]: I0916 04:49:27.828767 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/1403266a9792debaa127cd8df7a81c3c-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"1403266a9792debaa127cd8df7a81c3c\") " pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:27.828847 kubelet[2756]: I0916 04:49:27.828815 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/72a30db4fc25e4da65a3b99eba43be94-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"72a30db4fc25e4da65a3b99eba43be94\") " pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:27.828976 kubelet[2756]: I0916 04:49:27.828850 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fd193e1910e5dda003618658db09ac54-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fd193e1910e5dda003618658db09ac54\") " pod="kube-system/kube-apiserver-localhost" Sep 16 04:49:27.925696 kubelet[2756]: I0916 04:49:27.925644 2756 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 16 04:49:27.925875 kubelet[2756]: I0916 04:49:27.925776 2756 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 16 04:49:28.222353 kubelet[2756]: E0916 04:49:28.222172 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:28.224652 kubelet[2756]: E0916 04:49:28.224618 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:28.224761 kubelet[2756]: E0916 04:49:28.224626 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:28.613456 kubelet[2756]: I0916 04:49:28.613401 2756 apiserver.go:52] "Watching apiserver" Sep 16 04:49:28.627854 kubelet[2756]: I0916 04:49:28.627816 2756 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 16 04:49:28.653358 kubelet[2756]: I0916 04:49:28.652987 2756 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:28.653536 kubelet[2756]: E0916 04:49:28.653029 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:28.653717 kubelet[2756]: I0916 04:49:28.653641 2756 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:28.659229 kubelet[2756]: E0916 04:49:28.659201 2756 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 16 04:49:28.659364 kubelet[2756]: E0916 04:49:28.659331 2756 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 16 04:49:28.659411 kubelet[2756]: E0916 04:49:28.659371 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:28.659549 kubelet[2756]: E0916 04:49:28.659493 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:28.760565 kubelet[2756]: I0916 04:49:28.760476 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.760446166 podStartE2EDuration="1.760446166s" podCreationTimestamp="2025-09-16 04:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:49:28.735116445 +0000 UTC m=+1.207876014" watchObservedRunningTime="2025-09-16 04:49:28.760446166 +0000 UTC m=+1.233205735" Sep 16 04:49:28.770844 kubelet[2756]: I0916 04:49:28.770765 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.770733117 podStartE2EDuration="1.770733117s" podCreationTimestamp="2025-09-16 04:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:49:28.76106456 +0000 UTC m=+1.233824129" watchObservedRunningTime="2025-09-16 04:49:28.770733117 +0000 UTC m=+1.243492677" Sep 16 04:49:29.654671 kubelet[2756]: E0916 04:49:29.654621 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:29.655117 kubelet[2756]: E0916 04:49:29.655092 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:29.655254 kubelet[2756]: E0916 04:49:29.655233 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:30.657349 kubelet[2756]: E0916 04:49:30.657283 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:31.531579 kubelet[2756]: I0916 04:49:31.531482 2756 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 16 04:49:31.532052 containerd[1604]: time="2025-09-16T04:49:31.532007712Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 16 04:49:31.532472 kubelet[2756]: I0916 04:49:31.532191 2756 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 16 04:49:32.571965 kubelet[2756]: I0916 04:49:32.571883 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=5.571860775 podStartE2EDuration="5.571860775s" podCreationTimestamp="2025-09-16 04:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:49:28.770952508 +0000 UTC m=+1.243712077" watchObservedRunningTime="2025-09-16 04:49:32.571860775 +0000 UTC m=+5.044620344" Sep 16 04:49:32.582124 systemd[1]: Created slice kubepods-besteffort-poda8a7fc31_0e41_487f_a19a_e5cb03cf8231.slice - libcontainer container kubepods-besteffort-poda8a7fc31_0e41_487f_a19a_e5cb03cf8231.slice. Sep 16 04:49:32.658297 kubelet[2756]: I0916 04:49:32.658201 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/a8a7fc31-0e41-487f-a19a-e5cb03cf8231-kube-proxy\") pod \"kube-proxy-hmtgr\" (UID: \"a8a7fc31-0e41-487f-a19a-e5cb03cf8231\") " pod="kube-system/kube-proxy-hmtgr" Sep 16 04:49:32.658297 kubelet[2756]: I0916 04:49:32.658269 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a8a7fc31-0e41-487f-a19a-e5cb03cf8231-xtables-lock\") pod \"kube-proxy-hmtgr\" (UID: \"a8a7fc31-0e41-487f-a19a-e5cb03cf8231\") " pod="kube-system/kube-proxy-hmtgr" Sep 16 04:49:32.658297 kubelet[2756]: I0916 04:49:32.658295 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a8a7fc31-0e41-487f-a19a-e5cb03cf8231-lib-modules\") pod \"kube-proxy-hmtgr\" (UID: \"a8a7fc31-0e41-487f-a19a-e5cb03cf8231\") " pod="kube-system/kube-proxy-hmtgr" Sep 16 04:49:32.658712 kubelet[2756]: I0916 04:49:32.658318 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2g29\" (UniqueName: \"kubernetes.io/projected/a8a7fc31-0e41-487f-a19a-e5cb03cf8231-kube-api-access-p2g29\") pod \"kube-proxy-hmtgr\" (UID: \"a8a7fc31-0e41-487f-a19a-e5cb03cf8231\") " pod="kube-system/kube-proxy-hmtgr" Sep 16 04:49:32.694022 systemd[1]: Created slice kubepods-besteffort-pod856e9adf_d82f_4a95_b780_45936375246d.slice - libcontainer container kubepods-besteffort-pod856e9adf_d82f_4a95_b780_45936375246d.slice. Sep 16 04:49:32.758664 kubelet[2756]: I0916 04:49:32.758607 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/856e9adf-d82f-4a95-b780-45936375246d-var-lib-calico\") pod \"tigera-operator-755d956888-wwl8x\" (UID: \"856e9adf-d82f-4a95-b780-45936375246d\") " pod="tigera-operator/tigera-operator-755d956888-wwl8x" Sep 16 04:49:32.758664 kubelet[2756]: I0916 04:49:32.758667 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-875zl\" (UniqueName: \"kubernetes.io/projected/856e9adf-d82f-4a95-b780-45936375246d-kube-api-access-875zl\") pod \"tigera-operator-755d956888-wwl8x\" (UID: \"856e9adf-d82f-4a95-b780-45936375246d\") " pod="tigera-operator/tigera-operator-755d956888-wwl8x" Sep 16 04:49:32.895123 kubelet[2756]: E0916 04:49:32.895023 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:32.895892 containerd[1604]: time="2025-09-16T04:49:32.895803559Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-hmtgr,Uid:a8a7fc31-0e41-487f-a19a-e5cb03cf8231,Namespace:kube-system,Attempt:0,}" Sep 16 04:49:32.918966 containerd[1604]: time="2025-09-16T04:49:32.918906369Z" level=info msg="connecting to shim 8d6600bf9070959afc252769c61506d01dd8c339354158f02aace5753ffa45e4" address="unix:///run/containerd/s/2e8f40ccfe791c92a86acd9eb496fa0c3dbdb3748886197a80eb33ff96a9ac6f" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:32.998809 containerd[1604]: time="2025-09-16T04:49:32.998652520Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-wwl8x,Uid:856e9adf-d82f-4a95-b780-45936375246d,Namespace:tigera-operator,Attempt:0,}" Sep 16 04:49:33.053051 kubelet[2756]: E0916 04:49:33.027229 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:33.244190 containerd[1604]: time="2025-09-16T04:49:33.243913572Z" level=info msg="connecting to shim a1a659248814c7630f315891daadc01a0db222eeab0ef6592ca8dfa955c7e44c" address="unix:///run/containerd/s/373bec2a34acd5749422303e91abe4fc41085d9fa1ce558660f9f4a79af03b96" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:33.281777 systemd[1]: Started cri-containerd-8d6600bf9070959afc252769c61506d01dd8c339354158f02aace5753ffa45e4.scope - libcontainer container 8d6600bf9070959afc252769c61506d01dd8c339354158f02aace5753ffa45e4. Sep 16 04:49:33.286237 systemd[1]: Started cri-containerd-a1a659248814c7630f315891daadc01a0db222eeab0ef6592ca8dfa955c7e44c.scope - libcontainer container a1a659248814c7630f315891daadc01a0db222eeab0ef6592ca8dfa955c7e44c. Sep 16 04:49:33.330702 containerd[1604]: time="2025-09-16T04:49:33.330489781Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-hmtgr,Uid:a8a7fc31-0e41-487f-a19a-e5cb03cf8231,Namespace:kube-system,Attempt:0,} returns sandbox id \"8d6600bf9070959afc252769c61506d01dd8c339354158f02aace5753ffa45e4\"" Sep 16 04:49:33.333317 kubelet[2756]: E0916 04:49:33.333266 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:33.337894 containerd[1604]: time="2025-09-16T04:49:33.337843994Z" level=info msg="CreateContainer within sandbox \"8d6600bf9070959afc252769c61506d01dd8c339354158f02aace5753ffa45e4\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 16 04:49:33.355858 containerd[1604]: time="2025-09-16T04:49:33.355806200Z" level=info msg="Container 0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:33.358617 containerd[1604]: time="2025-09-16T04:49:33.358563136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-wwl8x,Uid:856e9adf-d82f-4a95-b780-45936375246d,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"a1a659248814c7630f315891daadc01a0db222eeab0ef6592ca8dfa955c7e44c\"" Sep 16 04:49:33.362156 containerd[1604]: time="2025-09-16T04:49:33.361992150Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 16 04:49:33.367165 containerd[1604]: time="2025-09-16T04:49:33.367110038Z" level=info msg="CreateContainer within sandbox \"8d6600bf9070959afc252769c61506d01dd8c339354158f02aace5753ffa45e4\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29\"" Sep 16 04:49:33.368063 containerd[1604]: time="2025-09-16T04:49:33.368020280Z" level=info msg="StartContainer for \"0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29\"" Sep 16 04:49:33.369485 containerd[1604]: time="2025-09-16T04:49:33.369448568Z" level=info msg="connecting to shim 0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29" address="unix:///run/containerd/s/2e8f40ccfe791c92a86acd9eb496fa0c3dbdb3748886197a80eb33ff96a9ac6f" protocol=ttrpc version=3 Sep 16 04:49:33.392702 systemd[1]: Started cri-containerd-0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29.scope - libcontainer container 0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29. Sep 16 04:49:33.441425 containerd[1604]: time="2025-09-16T04:49:33.441376275Z" level=info msg="StartContainer for \"0971ab68de95fe86cb61b03b8c65849cdd471180c3ae58703b364356b83bca29\" returns successfully" Sep 16 04:49:33.520704 kubelet[2756]: E0916 04:49:33.520533 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:33.668281 kubelet[2756]: E0916 04:49:33.668125 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:33.670140 kubelet[2756]: E0916 04:49:33.670098 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:33.671181 kubelet[2756]: E0916 04:49:33.670693 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:33.785173 kubelet[2756]: I0916 04:49:33.784861 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-hmtgr" podStartSLOduration=1.784834682 podStartE2EDuration="1.784834682s" podCreationTimestamp="2025-09-16 04:49:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:49:33.784467433 +0000 UTC m=+6.257227003" watchObservedRunningTime="2025-09-16 04:49:33.784834682 +0000 UTC m=+6.257594251" Sep 16 04:49:34.669794 kubelet[2756]: E0916 04:49:34.669712 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:34.669794 kubelet[2756]: E0916 04:49:34.669716 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:35.257497 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2958566861.mount: Deactivated successfully. Sep 16 04:49:38.343262 containerd[1604]: time="2025-09-16T04:49:38.343171010Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:38.345020 containerd[1604]: time="2025-09-16T04:49:38.344988795Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Sep 16 04:49:38.346819 containerd[1604]: time="2025-09-16T04:49:38.346752186Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:38.350454 containerd[1604]: time="2025-09-16T04:49:38.350388728Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:38.351546 containerd[1604]: time="2025-09-16T04:49:38.351472732Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 4.989442489s" Sep 16 04:49:38.351546 containerd[1604]: time="2025-09-16T04:49:38.351535030Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Sep 16 04:49:38.354146 containerd[1604]: time="2025-09-16T04:49:38.354104510Z" level=info msg="CreateContainer within sandbox \"a1a659248814c7630f315891daadc01a0db222eeab0ef6592ca8dfa955c7e44c\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 16 04:49:38.368524 containerd[1604]: time="2025-09-16T04:49:38.368407765Z" level=info msg="Container ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:38.379535 containerd[1604]: time="2025-09-16T04:49:38.378161469Z" level=info msg="CreateContainer within sandbox \"a1a659248814c7630f315891daadc01a0db222eeab0ef6592ca8dfa955c7e44c\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890\"" Sep 16 04:49:38.380418 containerd[1604]: time="2025-09-16T04:49:38.380326151Z" level=info msg="StartContainer for \"ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890\"" Sep 16 04:49:38.382661 containerd[1604]: time="2025-09-16T04:49:38.382594900Z" level=info msg="connecting to shim ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890" address="unix:///run/containerd/s/373bec2a34acd5749422303e91abe4fc41085d9fa1ce558660f9f4a79af03b96" protocol=ttrpc version=3 Sep 16 04:49:38.448789 systemd[1]: Started cri-containerd-ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890.scope - libcontainer container ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890. Sep 16 04:49:38.476087 kubelet[2756]: E0916 04:49:38.476047 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:38.506853 containerd[1604]: time="2025-09-16T04:49:38.506789797Z" level=info msg="StartContainer for \"ac734079ac1173651acae2783013ef00c1e111cb1261fb118779be1938400890\" returns successfully" Sep 16 04:49:39.088374 update_engine[1593]: I20250916 04:49:39.088227 1593 update_attempter.cc:509] Updating boot flags... Sep 16 04:49:44.044562 sudo[1813]: pam_unix(sudo:session): session closed for user root Sep 16 04:49:44.047567 sshd[1812]: Connection closed by 10.0.0.1 port 48586 Sep 16 04:49:44.059358 sshd-session[1809]: pam_unix(sshd:session): session closed for user core Sep 16 04:49:44.070426 systemd[1]: sshd@6-10.0.0.34:22-10.0.0.1:48586.service: Deactivated successfully. Sep 16 04:49:44.073806 systemd[1]: session-7.scope: Deactivated successfully. Sep 16 04:49:44.074193 systemd[1]: session-7.scope: Consumed 5.828s CPU time, 226.1M memory peak. Sep 16 04:49:44.076128 systemd-logind[1587]: Session 7 logged out. Waiting for processes to exit. Sep 16 04:49:44.078102 systemd-logind[1587]: Removed session 7. Sep 16 04:49:47.346182 kubelet[2756]: I0916 04:49:47.346085 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-wwl8x" podStartSLOduration=10.353557963 podStartE2EDuration="15.34603142s" podCreationTimestamp="2025-09-16 04:49:32 +0000 UTC" firstStartedPulling="2025-09-16 04:49:33.360098899 +0000 UTC m=+5.832858468" lastFinishedPulling="2025-09-16 04:49:38.352572346 +0000 UTC m=+10.825331925" observedRunningTime="2025-09-16 04:49:38.687883781 +0000 UTC m=+11.160643350" watchObservedRunningTime="2025-09-16 04:49:47.34603142 +0000 UTC m=+19.818790999" Sep 16 04:49:47.365257 systemd[1]: Created slice kubepods-besteffort-pod5af5d4dc_cd63_4307_97f2_06fe93bd1f22.slice - libcontainer container kubepods-besteffort-pod5af5d4dc_cd63_4307_97f2_06fe93bd1f22.slice. Sep 16 04:49:47.517700 kubelet[2756]: I0916 04:49:47.517621 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5af5d4dc-cd63-4307-97f2-06fe93bd1f22-tigera-ca-bundle\") pod \"calico-typha-754f9f798f-7b9h5\" (UID: \"5af5d4dc-cd63-4307-97f2-06fe93bd1f22\") " pod="calico-system/calico-typha-754f9f798f-7b9h5" Sep 16 04:49:47.517700 kubelet[2756]: I0916 04:49:47.517676 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p67h\" (UniqueName: \"kubernetes.io/projected/5af5d4dc-cd63-4307-97f2-06fe93bd1f22-kube-api-access-4p67h\") pod \"calico-typha-754f9f798f-7b9h5\" (UID: \"5af5d4dc-cd63-4307-97f2-06fe93bd1f22\") " pod="calico-system/calico-typha-754f9f798f-7b9h5" Sep 16 04:49:47.517935 kubelet[2756]: I0916 04:49:47.517722 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/5af5d4dc-cd63-4307-97f2-06fe93bd1f22-typha-certs\") pod \"calico-typha-754f9f798f-7b9h5\" (UID: \"5af5d4dc-cd63-4307-97f2-06fe93bd1f22\") " pod="calico-system/calico-typha-754f9f798f-7b9h5" Sep 16 04:49:47.670314 kubelet[2756]: E0916 04:49:47.670118 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:47.671553 containerd[1604]: time="2025-09-16T04:49:47.671453087Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-754f9f798f-7b9h5,Uid:5af5d4dc-cd63-4307-97f2-06fe93bd1f22,Namespace:calico-system,Attempt:0,}" Sep 16 04:49:47.738379 systemd[1]: Created slice kubepods-besteffort-pod207cc75a_bfc7_4563_a4f8_4d07510b5d82.slice - libcontainer container kubepods-besteffort-pod207cc75a_bfc7_4563_a4f8_4d07510b5d82.slice. Sep 16 04:49:47.755274 containerd[1604]: time="2025-09-16T04:49:47.755205413Z" level=info msg="connecting to shim 5b3aeca916bcabcec536b537a2d33e917f18db52b6f19251aea57b317d36773f" address="unix:///run/containerd/s/ff8da3b836b08fa75c0405819076a7491951c66ef42fe4e8871e14cd1e296f28" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:47.788789 systemd[1]: Started cri-containerd-5b3aeca916bcabcec536b537a2d33e917f18db52b6f19251aea57b317d36773f.scope - libcontainer container 5b3aeca916bcabcec536b537a2d33e917f18db52b6f19251aea57b317d36773f. Sep 16 04:49:47.846979 containerd[1604]: time="2025-09-16T04:49:47.846905390Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-754f9f798f-7b9h5,Uid:5af5d4dc-cd63-4307-97f2-06fe93bd1f22,Namespace:calico-system,Attempt:0,} returns sandbox id \"5b3aeca916bcabcec536b537a2d33e917f18db52b6f19251aea57b317d36773f\"" Sep 16 04:49:47.848035 kubelet[2756]: E0916 04:49:47.848001 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:47.849190 containerd[1604]: time="2025-09-16T04:49:47.849145124Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 16 04:49:47.921013 kubelet[2756]: I0916 04:49:47.920866 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-var-lib-calico\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921013 kubelet[2756]: I0916 04:49:47.920915 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmzv5\" (UniqueName: \"kubernetes.io/projected/207cc75a-bfc7-4563-a4f8-4d07510b5d82-kube-api-access-mmzv5\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921013 kubelet[2756]: I0916 04:49:47.920946 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/207cc75a-bfc7-4563-a4f8-4d07510b5d82-tigera-ca-bundle\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921013 kubelet[2756]: I0916 04:49:47.920968 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-xtables-lock\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921013 kubelet[2756]: I0916 04:49:47.920997 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-flexvol-driver-host\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921282 kubelet[2756]: I0916 04:49:47.921018 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-cni-bin-dir\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921282 kubelet[2756]: I0916 04:49:47.921037 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-cni-net-dir\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921282 kubelet[2756]: I0916 04:49:47.921057 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-policysync\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921282 kubelet[2756]: I0916 04:49:47.921078 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-var-run-calico\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921282 kubelet[2756]: I0916 04:49:47.921099 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-lib-modules\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921487 kubelet[2756]: I0916 04:49:47.921130 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/207cc75a-bfc7-4563-a4f8-4d07510b5d82-node-certs\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:47.921487 kubelet[2756]: I0916 04:49:47.921160 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/207cc75a-bfc7-4563-a4f8-4d07510b5d82-cni-log-dir\") pod \"calico-node-pswhm\" (UID: \"207cc75a-bfc7-4563-a4f8-4d07510b5d82\") " pod="calico-system/calico-node-pswhm" Sep 16 04:49:48.027617 kubelet[2756]: E0916 04:49:48.027266 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:49:48.033517 kubelet[2756]: E0916 04:49:48.033306 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.033517 kubelet[2756]: W0916 04:49:48.033344 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.033517 kubelet[2756]: E0916 04:49:48.033387 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.051315 kubelet[2756]: E0916 04:49:48.051253 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.051315 kubelet[2756]: W0916 04:49:48.051300 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.051315 kubelet[2756]: E0916 04:49:48.051336 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.122951 kubelet[2756]: E0916 04:49:48.122891 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.122951 kubelet[2756]: W0916 04:49:48.122927 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.122951 kubelet[2756]: E0916 04:49:48.122962 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.123472 kubelet[2756]: E0916 04:49:48.123394 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.123472 kubelet[2756]: W0916 04:49:48.123466 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.123617 kubelet[2756]: E0916 04:49:48.123557 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.124088 kubelet[2756]: E0916 04:49:48.124068 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.124088 kubelet[2756]: W0916 04:49:48.124085 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.124181 kubelet[2756]: E0916 04:49:48.124099 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.124598 kubelet[2756]: E0916 04:49:48.124577 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.124598 kubelet[2756]: W0916 04:49:48.124592 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.124598 kubelet[2756]: E0916 04:49:48.124605 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.125067 kubelet[2756]: E0916 04:49:48.125045 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.125067 kubelet[2756]: W0916 04:49:48.125065 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.125200 kubelet[2756]: E0916 04:49:48.125078 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.125418 kubelet[2756]: E0916 04:49:48.125398 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.125418 kubelet[2756]: W0916 04:49:48.125414 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.125479 kubelet[2756]: E0916 04:49:48.125428 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.125732 kubelet[2756]: E0916 04:49:48.125701 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.125732 kubelet[2756]: W0916 04:49:48.125724 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.125827 kubelet[2756]: E0916 04:49:48.125747 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.126107 kubelet[2756]: E0916 04:49:48.126009 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.126107 kubelet[2756]: W0916 04:49:48.126026 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.126107 kubelet[2756]: E0916 04:49:48.126038 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.126322 kubelet[2756]: E0916 04:49:48.126260 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.126322 kubelet[2756]: W0916 04:49:48.126273 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.126322 kubelet[2756]: E0916 04:49:48.126285 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.126598 kubelet[2756]: E0916 04:49:48.126493 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.126598 kubelet[2756]: W0916 04:49:48.126548 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.126598 kubelet[2756]: E0916 04:49:48.126562 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.127652 kubelet[2756]: E0916 04:49:48.126776 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.127652 kubelet[2756]: W0916 04:49:48.126786 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.127652 kubelet[2756]: E0916 04:49:48.126800 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.127652 kubelet[2756]: E0916 04:49:48.126993 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.127652 kubelet[2756]: W0916 04:49:48.127003 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.127652 kubelet[2756]: E0916 04:49:48.127013 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.127652 kubelet[2756]: E0916 04:49:48.127327 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.127652 kubelet[2756]: W0916 04:49:48.127339 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.127652 kubelet[2756]: E0916 04:49:48.127351 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.128234 kubelet[2756]: E0916 04:49:48.128105 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.128234 kubelet[2756]: W0916 04:49:48.128120 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.128234 kubelet[2756]: E0916 04:49:48.128130 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.128354 kubelet[2756]: E0916 04:49:48.128295 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.128354 kubelet[2756]: W0916 04:49:48.128303 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.128354 kubelet[2756]: E0916 04:49:48.128310 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.128706 kubelet[2756]: E0916 04:49:48.128456 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.128706 kubelet[2756]: W0916 04:49:48.128466 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.128706 kubelet[2756]: E0916 04:49:48.128474 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.128706 kubelet[2756]: E0916 04:49:48.128692 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.128706 kubelet[2756]: W0916 04:49:48.128700 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.128706 kubelet[2756]: E0916 04:49:48.128708 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.128907 kubelet[2756]: E0916 04:49:48.128876 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.128907 kubelet[2756]: W0916 04:49:48.128885 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.128907 kubelet[2756]: E0916 04:49:48.128894 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.129236 kubelet[2756]: E0916 04:49:48.129057 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.129236 kubelet[2756]: W0916 04:49:48.129071 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.129236 kubelet[2756]: E0916 04:49:48.129081 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.129598 kubelet[2756]: E0916 04:49:48.129259 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.129598 kubelet[2756]: W0916 04:49:48.129267 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.129598 kubelet[2756]: E0916 04:49:48.129276 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.129718 kubelet[2756]: E0916 04:49:48.129600 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.129718 kubelet[2756]: W0916 04:49:48.129611 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.129718 kubelet[2756]: E0916 04:49:48.129635 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.129807 kubelet[2756]: I0916 04:49:48.129676 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4-registration-dir\") pod \"csi-node-driver-9872f\" (UID: \"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4\") " pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:48.130222 kubelet[2756]: E0916 04:49:48.129939 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.130222 kubelet[2756]: W0916 04:49:48.129954 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.130222 kubelet[2756]: E0916 04:49:48.129965 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.130222 kubelet[2756]: I0916 04:49:48.129981 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqv9v\" (UniqueName: \"kubernetes.io/projected/7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4-kube-api-access-tqv9v\") pod \"csi-node-driver-9872f\" (UID: \"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4\") " pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:48.130222 kubelet[2756]: E0916 04:49:48.130195 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.130222 kubelet[2756]: W0916 04:49:48.130204 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.130222 kubelet[2756]: E0916 04:49:48.130214 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.130222 kubelet[2756]: I0916 04:49:48.130228 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4-socket-dir\") pod \"csi-node-driver-9872f\" (UID: \"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4\") " pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:48.130910 kubelet[2756]: E0916 04:49:48.130463 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.130910 kubelet[2756]: W0916 04:49:48.130472 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.130910 kubelet[2756]: E0916 04:49:48.130491 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.130910 kubelet[2756]: I0916 04:49:48.130521 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4-varrun\") pod \"csi-node-driver-9872f\" (UID: \"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4\") " pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:48.131719 kubelet[2756]: E0916 04:49:48.131631 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.131719 kubelet[2756]: W0916 04:49:48.131689 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.131821 kubelet[2756]: E0916 04:49:48.131752 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.132942 kubelet[2756]: E0916 04:49:48.132767 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.132942 kubelet[2756]: W0916 04:49:48.132792 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.132942 kubelet[2756]: E0916 04:49:48.132850 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.133441 kubelet[2756]: E0916 04:49:48.133423 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.133543 kubelet[2756]: W0916 04:49:48.133526 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.133696 kubelet[2756]: E0916 04:49:48.133644 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.134037 kubelet[2756]: E0916 04:49:48.133919 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.134037 kubelet[2756]: W0916 04:49:48.133935 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.134037 kubelet[2756]: E0916 04:49:48.133972 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.134421 kubelet[2756]: E0916 04:49:48.134281 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.134421 kubelet[2756]: W0916 04:49:48.134294 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.134421 kubelet[2756]: E0916 04:49:48.134328 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.134421 kubelet[2756]: I0916 04:49:48.134369 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4-kubelet-dir\") pod \"csi-node-driver-9872f\" (UID: \"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4\") " pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:48.134919 kubelet[2756]: E0916 04:49:48.134801 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.134919 kubelet[2756]: W0916 04:49:48.134815 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.134919 kubelet[2756]: E0916 04:49:48.134854 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.135273 kubelet[2756]: E0916 04:49:48.135183 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.135273 kubelet[2756]: W0916 04:49:48.135198 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.135273 kubelet[2756]: E0916 04:49:48.135215 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.136458 kubelet[2756]: E0916 04:49:48.136168 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.136458 kubelet[2756]: W0916 04:49:48.136191 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.136458 kubelet[2756]: E0916 04:49:48.136216 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.136458 kubelet[2756]: E0916 04:49:48.136382 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.136458 kubelet[2756]: W0916 04:49:48.136391 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.136458 kubelet[2756]: E0916 04:49:48.136399 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.137428 kubelet[2756]: E0916 04:49:48.137390 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.137428 kubelet[2756]: W0916 04:49:48.137415 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.137428 kubelet[2756]: E0916 04:49:48.137428 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.138354 kubelet[2756]: E0916 04:49:48.137705 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.138354 kubelet[2756]: W0916 04:49:48.137717 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.138354 kubelet[2756]: E0916 04:49:48.137730 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.238729 kubelet[2756]: E0916 04:49:48.238662 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.238729 kubelet[2756]: W0916 04:49:48.238701 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.238729 kubelet[2756]: E0916 04:49:48.238728 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.239061 kubelet[2756]: E0916 04:49:48.239039 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.239061 kubelet[2756]: W0916 04:49:48.239052 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.239135 kubelet[2756]: E0916 04:49:48.239071 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.239348 kubelet[2756]: E0916 04:49:48.239308 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.239348 kubelet[2756]: W0916 04:49:48.239325 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.239348 kubelet[2756]: E0916 04:49:48.239345 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.239695 kubelet[2756]: E0916 04:49:48.239643 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.239695 kubelet[2756]: W0916 04:49:48.239667 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.239764 kubelet[2756]: E0916 04:49:48.239714 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.239996 kubelet[2756]: E0916 04:49:48.239974 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.239996 kubelet[2756]: W0916 04:49:48.239990 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.240060 kubelet[2756]: E0916 04:49:48.240006 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.240269 kubelet[2756]: E0916 04:49:48.240229 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.240269 kubelet[2756]: W0916 04:49:48.240250 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.240269 kubelet[2756]: E0916 04:49:48.240268 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.240486 kubelet[2756]: E0916 04:49:48.240467 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.240486 kubelet[2756]: W0916 04:49:48.240481 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.240567 kubelet[2756]: E0916 04:49:48.240529 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.240720 kubelet[2756]: E0916 04:49:48.240701 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.240720 kubelet[2756]: W0916 04:49:48.240715 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.240769 kubelet[2756]: E0916 04:49:48.240749 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.240928 kubelet[2756]: E0916 04:49:48.240910 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.240928 kubelet[2756]: W0916 04:49:48.240924 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.240986 kubelet[2756]: E0916 04:49:48.240955 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.241130 kubelet[2756]: E0916 04:49:48.241112 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.241130 kubelet[2756]: W0916 04:49:48.241126 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.241198 kubelet[2756]: E0916 04:49:48.241158 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.241327 kubelet[2756]: E0916 04:49:48.241306 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.241327 kubelet[2756]: W0916 04:49:48.241319 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.241390 kubelet[2756]: E0916 04:49:48.241350 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.241531 kubelet[2756]: E0916 04:49:48.241499 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.241531 kubelet[2756]: W0916 04:49:48.241528 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.241596 kubelet[2756]: E0916 04:49:48.241546 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.241784 kubelet[2756]: E0916 04:49:48.241766 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.241784 kubelet[2756]: W0916 04:49:48.241780 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.241838 kubelet[2756]: E0916 04:49:48.241796 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.241987 kubelet[2756]: E0916 04:49:48.241970 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.241987 kubelet[2756]: W0916 04:49:48.241983 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.242034 kubelet[2756]: E0916 04:49:48.241997 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.242231 kubelet[2756]: E0916 04:49:48.242208 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.242231 kubelet[2756]: W0916 04:49:48.242227 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.242299 kubelet[2756]: E0916 04:49:48.242245 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.242446 kubelet[2756]: E0916 04:49:48.242427 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.242446 kubelet[2756]: W0916 04:49:48.242443 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.242493 kubelet[2756]: E0916 04:49:48.242474 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.242670 kubelet[2756]: E0916 04:49:48.242652 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.242670 kubelet[2756]: W0916 04:49:48.242667 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.242733 kubelet[2756]: E0916 04:49:48.242711 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.242886 kubelet[2756]: E0916 04:49:48.242869 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.242886 kubelet[2756]: W0916 04:49:48.242883 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.242934 kubelet[2756]: E0916 04:49:48.242914 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.243268 kubelet[2756]: E0916 04:49:48.243238 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.243268 kubelet[2756]: W0916 04:49:48.243257 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.243327 kubelet[2756]: E0916 04:49:48.243298 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.243523 kubelet[2756]: E0916 04:49:48.243477 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.243523 kubelet[2756]: W0916 04:49:48.243494 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.243574 kubelet[2756]: E0916 04:49:48.243556 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.243741 kubelet[2756]: E0916 04:49:48.243723 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.243741 kubelet[2756]: W0916 04:49:48.243738 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.243800 kubelet[2756]: E0916 04:49:48.243751 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.244027 kubelet[2756]: E0916 04:49:48.244003 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.244027 kubelet[2756]: W0916 04:49:48.244019 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.244109 kubelet[2756]: E0916 04:49:48.244040 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.244262 kubelet[2756]: E0916 04:49:48.244240 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.244262 kubelet[2756]: W0916 04:49:48.244256 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.244339 kubelet[2756]: E0916 04:49:48.244289 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.244610 kubelet[2756]: E0916 04:49:48.244592 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.244610 kubelet[2756]: W0916 04:49:48.244605 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.244883 kubelet[2756]: E0916 04:49:48.244750 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.244883 kubelet[2756]: E0916 04:49:48.244844 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.244883 kubelet[2756]: W0916 04:49:48.244863 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.244883 kubelet[2756]: E0916 04:49:48.244878 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.254471 kubelet[2756]: E0916 04:49:48.254428 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:48.254471 kubelet[2756]: W0916 04:49:48.254451 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:48.254471 kubelet[2756]: E0916 04:49:48.254475 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:48.343174 containerd[1604]: time="2025-09-16T04:49:48.343107888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-pswhm,Uid:207cc75a-bfc7-4563-a4f8-4d07510b5d82,Namespace:calico-system,Attempt:0,}" Sep 16 04:49:48.374701 containerd[1604]: time="2025-09-16T04:49:48.374623231Z" level=info msg="connecting to shim 3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7" address="unix:///run/containerd/s/0a935b1dba4442345aa8a2636b425529a3234faea10e38d26c4f540786a8b101" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:49:48.404833 systemd[1]: Started cri-containerd-3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7.scope - libcontainer container 3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7. Sep 16 04:49:48.438652 containerd[1604]: time="2025-09-16T04:49:48.438588418Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-pswhm,Uid:207cc75a-bfc7-4563-a4f8-4d07510b5d82,Namespace:calico-system,Attempt:0,} returns sandbox id \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\"" Sep 16 04:49:49.420236 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1463922450.mount: Deactivated successfully. Sep 16 04:49:49.643322 kubelet[2756]: E0916 04:49:49.643216 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:49:51.043779 containerd[1604]: time="2025-09-16T04:49:51.043707693Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:51.044601 containerd[1604]: time="2025-09-16T04:49:51.044561071Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Sep 16 04:49:51.045855 containerd[1604]: time="2025-09-16T04:49:51.045820413Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:51.047854 containerd[1604]: time="2025-09-16T04:49:51.047818306Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:51.048297 containerd[1604]: time="2025-09-16T04:49:51.048270107Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 3.199088744s" Sep 16 04:49:51.048348 containerd[1604]: time="2025-09-16T04:49:51.048299273Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Sep 16 04:49:51.049406 containerd[1604]: time="2025-09-16T04:49:51.049223183Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 16 04:49:51.061064 containerd[1604]: time="2025-09-16T04:49:51.061014614Z" level=info msg="CreateContainer within sandbox \"5b3aeca916bcabcec536b537a2d33e917f18db52b6f19251aea57b317d36773f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 16 04:49:51.070800 containerd[1604]: time="2025-09-16T04:49:51.070741947Z" level=info msg="Container a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:51.081189 containerd[1604]: time="2025-09-16T04:49:51.081121950Z" level=info msg="CreateContainer within sandbox \"5b3aeca916bcabcec536b537a2d33e917f18db52b6f19251aea57b317d36773f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b\"" Sep 16 04:49:51.081728 containerd[1604]: time="2025-09-16T04:49:51.081698385Z" level=info msg="StartContainer for \"a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b\"" Sep 16 04:49:51.083050 containerd[1604]: time="2025-09-16T04:49:51.082997552Z" level=info msg="connecting to shim a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b" address="unix:///run/containerd/s/ff8da3b836b08fa75c0405819076a7491951c66ef42fe4e8871e14cd1e296f28" protocol=ttrpc version=3 Sep 16 04:49:51.108819 systemd[1]: Started cri-containerd-a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b.scope - libcontainer container a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b. Sep 16 04:49:51.173880 containerd[1604]: time="2025-09-16T04:49:51.173816014Z" level=info msg="StartContainer for \"a757a5779a63394cc6a6759dddc2f8cbfcc8a5fee3553e5bd13504913ad2067b\" returns successfully" Sep 16 04:49:51.643594 kubelet[2756]: E0916 04:49:51.643475 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:49:51.704875 kubelet[2756]: E0916 04:49:51.704779 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:51.717059 kubelet[2756]: I0916 04:49:51.716984 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-754f9f798f-7b9h5" podStartSLOduration=1.516769526 podStartE2EDuration="4.716963203s" podCreationTimestamp="2025-09-16 04:49:47 +0000 UTC" firstStartedPulling="2025-09-16 04:49:47.84877712 +0000 UTC m=+20.321536689" lastFinishedPulling="2025-09-16 04:49:51.048970797 +0000 UTC m=+23.521730366" observedRunningTime="2025-09-16 04:49:51.716172133 +0000 UTC m=+24.188931692" watchObservedRunningTime="2025-09-16 04:49:51.716963203 +0000 UTC m=+24.189722772" Sep 16 04:49:51.752976 kubelet[2756]: E0916 04:49:51.752826 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.752976 kubelet[2756]: W0916 04:49:51.752861 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.752976 kubelet[2756]: E0916 04:49:51.752889 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.753395 kubelet[2756]: E0916 04:49:51.753357 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.753395 kubelet[2756]: W0916 04:49:51.753371 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.753395 kubelet[2756]: E0916 04:49:51.753381 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.753770 kubelet[2756]: E0916 04:49:51.753645 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.753770 kubelet[2756]: W0916 04:49:51.753683 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.753922 kubelet[2756]: E0916 04:49:51.753764 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.754362 kubelet[2756]: E0916 04:49:51.754331 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.754362 kubelet[2756]: W0916 04:49:51.754351 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.754362 kubelet[2756]: E0916 04:49:51.754363 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.754703 kubelet[2756]: E0916 04:49:51.754682 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.754703 kubelet[2756]: W0916 04:49:51.754696 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.754703 kubelet[2756]: E0916 04:49:51.754706 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.755050 kubelet[2756]: E0916 04:49:51.754864 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.755050 kubelet[2756]: W0916 04:49:51.754872 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.755050 kubelet[2756]: E0916 04:49:51.754881 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.755050 kubelet[2756]: E0916 04:49:51.755032 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.755050 kubelet[2756]: W0916 04:49:51.755039 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.755050 kubelet[2756]: E0916 04:49:51.755048 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.755597 kubelet[2756]: E0916 04:49:51.755217 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.755597 kubelet[2756]: W0916 04:49:51.755226 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.755597 kubelet[2756]: E0916 04:49:51.755234 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.755597 kubelet[2756]: E0916 04:49:51.755541 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.755597 kubelet[2756]: W0916 04:49:51.755551 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.755597 kubelet[2756]: E0916 04:49:51.755561 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.755893 kubelet[2756]: E0916 04:49:51.755824 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.755893 kubelet[2756]: W0916 04:49:51.755838 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.755893 kubelet[2756]: E0916 04:49:51.755851 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.756125 kubelet[2756]: E0916 04:49:51.756102 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.756125 kubelet[2756]: W0916 04:49:51.756120 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.756230 kubelet[2756]: E0916 04:49:51.756131 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.756363 kubelet[2756]: E0916 04:49:51.756342 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.756363 kubelet[2756]: W0916 04:49:51.756355 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.756363 kubelet[2756]: E0916 04:49:51.756365 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.756656 kubelet[2756]: E0916 04:49:51.756605 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.756656 kubelet[2756]: W0916 04:49:51.756631 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.756656 kubelet[2756]: E0916 04:49:51.756641 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.757261 kubelet[2756]: E0916 04:49:51.756808 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.757261 kubelet[2756]: W0916 04:49:51.756816 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.757261 kubelet[2756]: E0916 04:49:51.756824 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.757261 kubelet[2756]: E0916 04:49:51.756961 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.757261 kubelet[2756]: W0916 04:49:51.756968 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.757261 kubelet[2756]: E0916 04:49:51.756976 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.763776 kubelet[2756]: E0916 04:49:51.763737 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.763776 kubelet[2756]: W0916 04:49:51.763759 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.763776 kubelet[2756]: E0916 04:49:51.763775 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.764153 kubelet[2756]: E0916 04:49:51.764102 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.764153 kubelet[2756]: W0916 04:49:51.764141 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.764237 kubelet[2756]: E0916 04:49:51.764195 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.764591 kubelet[2756]: E0916 04:49:51.764539 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.764591 kubelet[2756]: W0916 04:49:51.764580 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.764692 kubelet[2756]: E0916 04:49:51.764616 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.764891 kubelet[2756]: E0916 04:49:51.764856 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.764891 kubelet[2756]: W0916 04:49:51.764873 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.764891 kubelet[2756]: E0916 04:49:51.764893 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.765197 kubelet[2756]: E0916 04:49:51.765176 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.765197 kubelet[2756]: W0916 04:49:51.765191 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.765282 kubelet[2756]: E0916 04:49:51.765208 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.765465 kubelet[2756]: E0916 04:49:51.765441 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.765465 kubelet[2756]: W0916 04:49:51.765454 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.765465 kubelet[2756]: E0916 04:49:51.765470 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.765829 kubelet[2756]: E0916 04:49:51.765796 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.765829 kubelet[2756]: W0916 04:49:51.765811 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.765913 kubelet[2756]: E0916 04:49:51.765855 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.766036 kubelet[2756]: E0916 04:49:51.766013 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.766036 kubelet[2756]: W0916 04:49:51.766026 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.766130 kubelet[2756]: E0916 04:49:51.766059 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.766246 kubelet[2756]: E0916 04:49:51.766223 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.766246 kubelet[2756]: W0916 04:49:51.766238 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.766331 kubelet[2756]: E0916 04:49:51.766275 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.766461 kubelet[2756]: E0916 04:49:51.766440 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.766461 kubelet[2756]: W0916 04:49:51.766452 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.766587 kubelet[2756]: E0916 04:49:51.766468 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.766737 kubelet[2756]: E0916 04:49:51.766715 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.766737 kubelet[2756]: W0916 04:49:51.766727 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.766821 kubelet[2756]: E0916 04:49:51.766742 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.767074 kubelet[2756]: E0916 04:49:51.767049 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.767074 kubelet[2756]: W0916 04:49:51.767067 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.767164 kubelet[2756]: E0916 04:49:51.767091 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.767320 kubelet[2756]: E0916 04:49:51.767300 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.767320 kubelet[2756]: W0916 04:49:51.767315 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.767393 kubelet[2756]: E0916 04:49:51.767333 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.767726 kubelet[2756]: E0916 04:49:51.767700 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.767726 kubelet[2756]: W0916 04:49:51.767717 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.767823 kubelet[2756]: E0916 04:49:51.767735 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.768052 kubelet[2756]: E0916 04:49:51.768029 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.768052 kubelet[2756]: W0916 04:49:51.768044 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.768126 kubelet[2756]: E0916 04:49:51.768064 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.768370 kubelet[2756]: E0916 04:49:51.768347 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.768370 kubelet[2756]: W0916 04:49:51.768364 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.768456 kubelet[2756]: E0916 04:49:51.768383 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.768713 kubelet[2756]: E0916 04:49:51.768689 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.768713 kubelet[2756]: W0916 04:49:51.768705 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.768783 kubelet[2756]: E0916 04:49:51.768717 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:51.768932 kubelet[2756]: E0916 04:49:51.768911 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:51.768932 kubelet[2756]: W0916 04:49:51.768927 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:51.769015 kubelet[2756]: E0916 04:49:51.768940 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.706493 kubelet[2756]: I0916 04:49:52.706454 2756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 16 04:49:52.707493 kubelet[2756]: E0916 04:49:52.707454 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:52.740469 containerd[1604]: time="2025-09-16T04:49:52.740399438Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:52.741496 containerd[1604]: time="2025-09-16T04:49:52.741459895Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Sep 16 04:49:52.743071 containerd[1604]: time="2025-09-16T04:49:52.742978014Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:52.745167 containerd[1604]: time="2025-09-16T04:49:52.745099028Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:52.745853 containerd[1604]: time="2025-09-16T04:49:52.745780781Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 1.696524015s" Sep 16 04:49:52.745853 containerd[1604]: time="2025-09-16T04:49:52.745843510Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Sep 16 04:49:52.748578 containerd[1604]: time="2025-09-16T04:49:52.748375328Z" level=info msg="CreateContainer within sandbox \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 16 04:49:52.761701 containerd[1604]: time="2025-09-16T04:49:52.760013102Z" level=info msg="Container f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:52.762529 kubelet[2756]: E0916 04:49:52.762399 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.762529 kubelet[2756]: W0916 04:49:52.762428 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.762529 kubelet[2756]: E0916 04:49:52.762462 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.763241 kubelet[2756]: E0916 04:49:52.762912 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.763241 kubelet[2756]: W0916 04:49:52.762936 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.763241 kubelet[2756]: E0916 04:49:52.762950 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.763241 kubelet[2756]: E0916 04:49:52.763185 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.763241 kubelet[2756]: W0916 04:49:52.763211 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.763241 kubelet[2756]: E0916 04:49:52.763222 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.763468 kubelet[2756]: E0916 04:49:52.763451 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.763468 kubelet[2756]: W0916 04:49:52.763466 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.763475 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.763680 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.764131 kubelet[2756]: W0916 04:49:52.763688 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.763695 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.763843 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.764131 kubelet[2756]: W0916 04:49:52.763851 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.763859 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.764022 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.764131 kubelet[2756]: W0916 04:49:52.764031 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.764131 kubelet[2756]: E0916 04:49:52.764040 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.764428 kubelet[2756]: E0916 04:49:52.764223 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.764428 kubelet[2756]: W0916 04:49:52.764232 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.764428 kubelet[2756]: E0916 04:49:52.764242 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.764428 kubelet[2756]: E0916 04:49:52.764423 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.764428 kubelet[2756]: W0916 04:49:52.764432 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.764441 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.764653 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.766167 kubelet[2756]: W0916 04:49:52.764663 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.764672 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.764911 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.766167 kubelet[2756]: W0916 04:49:52.764943 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.764984 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.765286 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.766167 kubelet[2756]: W0916 04:49:52.765297 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766167 kubelet[2756]: E0916 04:49:52.765307 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.766396 kubelet[2756]: E0916 04:49:52.765601 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.766396 kubelet[2756]: W0916 04:49:52.765620 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766396 kubelet[2756]: E0916 04:49:52.765631 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.766396 kubelet[2756]: E0916 04:49:52.765940 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.766396 kubelet[2756]: W0916 04:49:52.765951 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766396 kubelet[2756]: E0916 04:49:52.765964 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.766396 kubelet[2756]: E0916 04:49:52.766175 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.766396 kubelet[2756]: W0916 04:49:52.766187 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.766396 kubelet[2756]: E0916 04:49:52.766199 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.770681 kubelet[2756]: E0916 04:49:52.770645 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.770681 kubelet[2756]: W0916 04:49:52.770667 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.770789 kubelet[2756]: E0916 04:49:52.770689 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.770919 kubelet[2756]: E0916 04:49:52.770899 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.770919 kubelet[2756]: W0916 04:49:52.770917 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.770972 kubelet[2756]: E0916 04:49:52.770930 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.771190 kubelet[2756]: E0916 04:49:52.771171 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.771245 kubelet[2756]: W0916 04:49:52.771185 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.771245 kubelet[2756]: E0916 04:49:52.771221 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.771438 kubelet[2756]: E0916 04:49:52.771422 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.771438 kubelet[2756]: W0916 04:49:52.771433 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.771538 kubelet[2756]: E0916 04:49:52.771446 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.771677 containerd[1604]: time="2025-09-16T04:49:52.771600273Z" level=info msg="CreateContainer within sandbox \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\"" Sep 16 04:49:52.771822 kubelet[2756]: E0916 04:49:52.771753 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.771822 kubelet[2756]: W0916 04:49:52.771780 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.771910 kubelet[2756]: E0916 04:49:52.771835 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.772170 kubelet[2756]: E0916 04:49:52.772147 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.772170 kubelet[2756]: W0916 04:49:52.772166 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.772346 kubelet[2756]: E0916 04:49:52.772195 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.772479 containerd[1604]: time="2025-09-16T04:49:52.772446507Z" level=info msg="StartContainer for \"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\"" Sep 16 04:49:52.772660 kubelet[2756]: E0916 04:49:52.772634 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.772660 kubelet[2756]: W0916 04:49:52.772654 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.772740 kubelet[2756]: E0916 04:49:52.772678 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.773237 kubelet[2756]: E0916 04:49:52.773201 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.773237 kubelet[2756]: W0916 04:49:52.773219 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.773584 kubelet[2756]: E0916 04:49:52.773563 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.773927 kubelet[2756]: E0916 04:49:52.773894 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.773927 kubelet[2756]: W0916 04:49:52.773914 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.774003 kubelet[2756]: E0916 04:49:52.773975 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.774203 kubelet[2756]: E0916 04:49:52.774179 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.774203 kubelet[2756]: W0916 04:49:52.774196 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.774349 kubelet[2756]: E0916 04:49:52.774239 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.774454 kubelet[2756]: E0916 04:49:52.774433 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.774454 kubelet[2756]: W0916 04:49:52.774447 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.774573 kubelet[2756]: E0916 04:49:52.774553 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.775369 kubelet[2756]: E0916 04:49:52.774697 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.775369 kubelet[2756]: W0916 04:49:52.774713 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.775369 kubelet[2756]: E0916 04:49:52.774731 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.775369 kubelet[2756]: E0916 04:49:52.774964 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.775369 kubelet[2756]: W0916 04:49:52.774975 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.775369 kubelet[2756]: E0916 04:49:52.774990 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.775369 kubelet[2756]: E0916 04:49:52.775178 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.775369 kubelet[2756]: W0916 04:49:52.775186 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.775369 kubelet[2756]: E0916 04:49:52.775200 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.775644 containerd[1604]: time="2025-09-16T04:49:52.775273691Z" level=info msg="connecting to shim f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23" address="unix:///run/containerd/s/0a935b1dba4442345aa8a2636b425529a3234faea10e38d26c4f540786a8b101" protocol=ttrpc version=3 Sep 16 04:49:52.775678 kubelet[2756]: E0916 04:49:52.775386 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.775678 kubelet[2756]: W0916 04:49:52.775430 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.775678 kubelet[2756]: E0916 04:49:52.775440 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.775945 kubelet[2756]: E0916 04:49:52.775917 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.775988 kubelet[2756]: W0916 04:49:52.775944 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.776099 kubelet[2756]: E0916 04:49:52.776070 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.776358 kubelet[2756]: E0916 04:49:52.776335 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.776358 kubelet[2756]: W0916 04:49:52.776355 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.776645 kubelet[2756]: E0916 04:49:52.776580 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.777049 kubelet[2756]: E0916 04:49:52.777016 2756 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 16 04:49:52.777049 kubelet[2756]: W0916 04:49:52.777031 2756 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 16 04:49:52.777049 kubelet[2756]: E0916 04:49:52.777041 2756 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 16 04:49:52.799708 systemd[1]: Started cri-containerd-f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23.scope - libcontainer container f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23. Sep 16 04:49:52.854087 containerd[1604]: time="2025-09-16T04:49:52.853995915Z" level=info msg="StartContainer for \"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\" returns successfully" Sep 16 04:49:52.867173 systemd[1]: cri-containerd-f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23.scope: Deactivated successfully. Sep 16 04:49:52.870127 containerd[1604]: time="2025-09-16T04:49:52.870073801Z" level=info msg="received exit event container_id:\"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\" id:\"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\" pid:3500 exited_at:{seconds:1757998192 nanos:869575142}" Sep 16 04:49:52.870263 containerd[1604]: time="2025-09-16T04:49:52.870241657Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\" id:\"f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23\" pid:3500 exited_at:{seconds:1757998192 nanos:869575142}" Sep 16 04:49:52.899499 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f61970852c199b96438433e926720b245ae234d584b8ed47f6ecaf9969078a23-rootfs.mount: Deactivated successfully. Sep 16 04:49:53.645762 kubelet[2756]: E0916 04:49:53.645705 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:49:53.713031 containerd[1604]: time="2025-09-16T04:49:53.712975461Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 16 04:49:54.991209 kubelet[2756]: I0916 04:49:54.991154 2756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 16 04:49:54.991793 kubelet[2756]: E0916 04:49:54.991627 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:55.643857 kubelet[2756]: E0916 04:49:55.643766 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:49:55.715420 kubelet[2756]: E0916 04:49:55.715364 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:57.647042 kubelet[2756]: E0916 04:49:57.646996 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:49:57.821304 containerd[1604]: time="2025-09-16T04:49:57.821243152Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:57.822045 containerd[1604]: time="2025-09-16T04:49:57.821980960Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Sep 16 04:49:57.823088 containerd[1604]: time="2025-09-16T04:49:57.823053076Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:57.825323 containerd[1604]: time="2025-09-16T04:49:57.825242354Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:49:57.825888 containerd[1604]: time="2025-09-16T04:49:57.825850347Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 4.112829682s" Sep 16 04:49:57.825888 containerd[1604]: time="2025-09-16T04:49:57.825882759Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Sep 16 04:49:57.827881 containerd[1604]: time="2025-09-16T04:49:57.827853836Z" level=info msg="CreateContainer within sandbox \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 16 04:49:57.838334 containerd[1604]: time="2025-09-16T04:49:57.838251392Z" level=info msg="Container 71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:49:57.848818 containerd[1604]: time="2025-09-16T04:49:57.848778223Z" level=info msg="CreateContainer within sandbox \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\"" Sep 16 04:49:57.850096 containerd[1604]: time="2025-09-16T04:49:57.849393770Z" level=info msg="StartContainer for \"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\"" Sep 16 04:49:57.850868 containerd[1604]: time="2025-09-16T04:49:57.850833298Z" level=info msg="connecting to shim 71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114" address="unix:///run/containerd/s/0a935b1dba4442345aa8a2636b425529a3234faea10e38d26c4f540786a8b101" protocol=ttrpc version=3 Sep 16 04:49:57.877724 systemd[1]: Started cri-containerd-71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114.scope - libcontainer container 71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114. Sep 16 04:49:57.927499 containerd[1604]: time="2025-09-16T04:49:57.927335349Z" level=info msg="StartContainer for \"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\" returns successfully" Sep 16 04:49:58.778622 kubelet[2756]: I0916 04:49:58.778583 2756 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 16 04:49:58.780378 systemd[1]: cri-containerd-71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114.scope: Deactivated successfully. Sep 16 04:49:58.780846 systemd[1]: cri-containerd-71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114.scope: Consumed 553ms CPU time, 179.5M memory peak, 3.4M read from disk, 171.3M written to disk. Sep 16 04:49:58.786669 containerd[1604]: time="2025-09-16T04:49:58.786613514Z" level=info msg="received exit event container_id:\"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\" id:\"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\" pid:3561 exited_at:{seconds:1757998198 nanos:786146757}" Sep 16 04:49:58.786864 containerd[1604]: time="2025-09-16T04:49:58.786613293Z" level=info msg="TaskExit event in podsandbox handler container_id:\"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\" id:\"71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114\" pid:3561 exited_at:{seconds:1757998198 nanos:786146757}" Sep 16 04:49:58.828560 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-71e14a8e3f86be3239278c45c61b52efd4625e0efc7f1d429bb299114387d114-rootfs.mount: Deactivated successfully. Sep 16 04:49:58.840965 systemd[1]: Created slice kubepods-besteffort-pod735decdf_fb16_4265_96d7_4da1d18f1679.slice - libcontainer container kubepods-besteffort-pod735decdf_fb16_4265_96d7_4da1d18f1679.slice. Sep 16 04:49:58.847111 systemd[1]: Created slice kubepods-burstable-pod1b3a3c6b_e0ed_437a_9826_214673136666.slice - libcontainer container kubepods-burstable-pod1b3a3c6b_e0ed_437a_9826_214673136666.slice. Sep 16 04:49:58.856242 systemd[1]: Created slice kubepods-besteffort-pod61a5328c_f236_4ba6_9b78_7bb5afd28a2f.slice - libcontainer container kubepods-besteffort-pod61a5328c_f236_4ba6_9b78_7bb5afd28a2f.slice. Sep 16 04:49:58.863013 systemd[1]: Created slice kubepods-burstable-pod805af22b_e7a5_4562_8fa7_b331a6a81d08.slice - libcontainer container kubepods-burstable-pod805af22b_e7a5_4562_8fa7_b331a6a81d08.slice. Sep 16 04:49:58.870918 systemd[1]: Created slice kubepods-besteffort-pod7da891ca_d373_47fb_8d2b_a8bb2752d2c0.slice - libcontainer container kubepods-besteffort-pod7da891ca_d373_47fb_8d2b_a8bb2752d2c0.slice. Sep 16 04:49:58.883104 systemd[1]: Created slice kubepods-besteffort-podd7bf725a_401b_4090_b1b5_801486b6473f.slice - libcontainer container kubepods-besteffort-podd7bf725a_401b_4090_b1b5_801486b6473f.slice. Sep 16 04:49:58.889158 systemd[1]: Created slice kubepods-besteffort-pod3e572577_68f4_49a7_9725_3f0cb8aef839.slice - libcontainer container kubepods-besteffort-pod3e572577_68f4_49a7_9725_3f0cb8aef839.slice. Sep 16 04:49:58.914381 kubelet[2756]: I0916 04:49:58.914313 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j5gw\" (UniqueName: \"kubernetes.io/projected/3e572577-68f4-49a7-9725-3f0cb8aef839-kube-api-access-9j5gw\") pod \"calico-apiserver-6b89c956bf-6bfxs\" (UID: \"3e572577-68f4-49a7-9725-3f0cb8aef839\") " pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" Sep 16 04:49:58.914381 kubelet[2756]: I0916 04:49:58.914363 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/d7bf725a-401b-4090-b1b5-801486b6473f-calico-apiserver-certs\") pod \"calico-apiserver-6b89c956bf-92rwh\" (UID: \"d7bf725a-401b-4090-b1b5-801486b6473f\") " pod="calico-apiserver/calico-apiserver-6b89c956bf-92rwh" Sep 16 04:49:58.914381 kubelet[2756]: I0916 04:49:58.914392 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr6bt\" (UniqueName: \"kubernetes.io/projected/d7bf725a-401b-4090-b1b5-801486b6473f-kube-api-access-pr6bt\") pod \"calico-apiserver-6b89c956bf-92rwh\" (UID: \"d7bf725a-401b-4090-b1b5-801486b6473f\") " pod="calico-apiserver/calico-apiserver-6b89c956bf-92rwh" Sep 16 04:49:58.914732 kubelet[2756]: I0916 04:49:58.914414 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lq5z\" (UniqueName: \"kubernetes.io/projected/805af22b-e7a5-4562-8fa7-b331a6a81d08-kube-api-access-2lq5z\") pod \"coredns-668d6bf9bc-s4499\" (UID: \"805af22b-e7a5-4562-8fa7-b331a6a81d08\") " pod="kube-system/coredns-668d6bf9bc-s4499" Sep 16 04:49:58.914732 kubelet[2756]: I0916 04:49:58.914456 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/7da891ca-d373-47fb-8d2b-a8bb2752d2c0-goldmane-key-pair\") pod \"goldmane-54d579b49d-mdn5g\" (UID: \"7da891ca-d373-47fb-8d2b-a8bb2752d2c0\") " pod="calico-system/goldmane-54d579b49d-mdn5g" Sep 16 04:49:58.914732 kubelet[2756]: I0916 04:49:58.914528 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-ca-bundle\") pod \"whisker-66c667cbbc-gbsmd\" (UID: \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\") " pod="calico-system/whisker-66c667cbbc-gbsmd" Sep 16 04:49:58.914732 kubelet[2756]: I0916 04:49:58.914556 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp66b\" (UniqueName: \"kubernetes.io/projected/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-kube-api-access-pp66b\") pod \"whisker-66c667cbbc-gbsmd\" (UID: \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\") " pod="calico-system/whisker-66c667cbbc-gbsmd" Sep 16 04:49:58.914732 kubelet[2756]: I0916 04:49:58.914579 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da891ca-d373-47fb-8d2b-a8bb2752d2c0-config\") pod \"goldmane-54d579b49d-mdn5g\" (UID: \"7da891ca-d373-47fb-8d2b-a8bb2752d2c0\") " pod="calico-system/goldmane-54d579b49d-mdn5g" Sep 16 04:49:58.914910 kubelet[2756]: I0916 04:49:58.914604 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/735decdf-fb16-4265-96d7-4da1d18f1679-tigera-ca-bundle\") pod \"calico-kube-controllers-b8d5c47bf-nrnjv\" (UID: \"735decdf-fb16-4265-96d7-4da1d18f1679\") " pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" Sep 16 04:49:58.914910 kubelet[2756]: I0916 04:49:58.914634 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3e572577-68f4-49a7-9725-3f0cb8aef839-calico-apiserver-certs\") pod \"calico-apiserver-6b89c956bf-6bfxs\" (UID: \"3e572577-68f4-49a7-9725-3f0cb8aef839\") " pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" Sep 16 04:49:58.914910 kubelet[2756]: I0916 04:49:58.914658 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c29z\" (UniqueName: \"kubernetes.io/projected/1b3a3c6b-e0ed-437a-9826-214673136666-kube-api-access-9c29z\") pod \"coredns-668d6bf9bc-2wvgh\" (UID: \"1b3a3c6b-e0ed-437a-9826-214673136666\") " pod="kube-system/coredns-668d6bf9bc-2wvgh" Sep 16 04:49:58.914910 kubelet[2756]: I0916 04:49:58.914678 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xf67\" (UniqueName: \"kubernetes.io/projected/735decdf-fb16-4265-96d7-4da1d18f1679-kube-api-access-5xf67\") pod \"calico-kube-controllers-b8d5c47bf-nrnjv\" (UID: \"735decdf-fb16-4265-96d7-4da1d18f1679\") " pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" Sep 16 04:49:58.914910 kubelet[2756]: I0916 04:49:58.914698 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-backend-key-pair\") pod \"whisker-66c667cbbc-gbsmd\" (UID: \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\") " pod="calico-system/whisker-66c667cbbc-gbsmd" Sep 16 04:49:58.915092 kubelet[2756]: I0916 04:49:58.914712 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da891ca-d373-47fb-8d2b-a8bb2752d2c0-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-mdn5g\" (UID: \"7da891ca-d373-47fb-8d2b-a8bb2752d2c0\") " pod="calico-system/goldmane-54d579b49d-mdn5g" Sep 16 04:49:58.915092 kubelet[2756]: I0916 04:49:58.914725 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/805af22b-e7a5-4562-8fa7-b331a6a81d08-config-volume\") pod \"coredns-668d6bf9bc-s4499\" (UID: \"805af22b-e7a5-4562-8fa7-b331a6a81d08\") " pod="kube-system/coredns-668d6bf9bc-s4499" Sep 16 04:49:58.915092 kubelet[2756]: I0916 04:49:58.914742 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr29n\" (UniqueName: \"kubernetes.io/projected/7da891ca-d373-47fb-8d2b-a8bb2752d2c0-kube-api-access-tr29n\") pod \"goldmane-54d579b49d-mdn5g\" (UID: \"7da891ca-d373-47fb-8d2b-a8bb2752d2c0\") " pod="calico-system/goldmane-54d579b49d-mdn5g" Sep 16 04:49:58.915092 kubelet[2756]: I0916 04:49:58.914756 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b3a3c6b-e0ed-437a-9826-214673136666-config-volume\") pod \"coredns-668d6bf9bc-2wvgh\" (UID: \"1b3a3c6b-e0ed-437a-9826-214673136666\") " pod="kube-system/coredns-668d6bf9bc-2wvgh" Sep 16 04:49:59.192838 kubelet[2756]: E0916 04:49:59.192596 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:59.192838 kubelet[2756]: E0916 04:49:59.192828 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:49:59.193008 containerd[1604]: time="2025-09-16T04:49:59.192711602Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-mdn5g,Uid:7da891ca-d373-47fb-8d2b-a8bb2752d2c0,Namespace:calico-system,Attempt:0,}" Sep 16 04:49:59.193366 containerd[1604]: time="2025-09-16T04:49:59.193301872Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s4499,Uid:805af22b-e7a5-4562-8fa7-b331a6a81d08,Namespace:kube-system,Attempt:0,}" Sep 16 04:49:59.193525 containerd[1604]: time="2025-09-16T04:49:59.193402100Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-b8d5c47bf-nrnjv,Uid:735decdf-fb16-4265-96d7-4da1d18f1679,Namespace:calico-system,Attempt:0,}" Sep 16 04:49:59.193666 containerd[1604]: time="2025-09-16T04:49:59.193636121Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-92rwh,Uid:d7bf725a-401b-4090-b1b5-801486b6473f,Namespace:calico-apiserver,Attempt:0,}" Sep 16 04:49:59.193841 containerd[1604]: time="2025-09-16T04:49:59.193807964Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-2wvgh,Uid:1b3a3c6b-e0ed-437a-9826-214673136666,Namespace:kube-system,Attempt:0,}" Sep 16 04:49:59.194286 containerd[1604]: time="2025-09-16T04:49:59.193972713Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-66c667cbbc-gbsmd,Uid:61a5328c-f236-4ba6-9b78-7bb5afd28a2f,Namespace:calico-system,Attempt:0,}" Sep 16 04:49:59.198448 containerd[1604]: time="2025-09-16T04:49:59.198316410Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-6bfxs,Uid:3e572577-68f4-49a7-9725-3f0cb8aef839,Namespace:calico-apiserver,Attempt:0,}" Sep 16 04:49:59.334112 containerd[1604]: time="2025-09-16T04:49:59.333950015Z" level=error msg="Failed to destroy network for sandbox \"dee38da718b2fab6e5f80696975e66627accdbb3a64e9198ee53d70fd0b5403e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.342837 containerd[1604]: time="2025-09-16T04:49:59.342773555Z" level=error msg="Failed to destroy network for sandbox \"8a62a52d87b781e0cc6a275bae29735a5328ae735fabae6a2f1d0c0cf875e278\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.343021 containerd[1604]: time="2025-09-16T04:49:59.342989601Z" level=error msg="Failed to destroy network for sandbox \"05612b3cb5e060a236c79ca2f666d554d95bf13f6a331b4f924fd2296cebd62c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.343237 containerd[1604]: time="2025-09-16T04:49:59.343184107Z" level=error msg="Failed to destroy network for sandbox \"8f5add85341627848560882de5f95f1ffcbbc2c2d960f39073484db40b34d6d3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.343337 containerd[1604]: time="2025-09-16T04:49:59.343304052Z" level=error msg="Failed to destroy network for sandbox \"0118c560c86cc8ebf93f82e59d97b57268024af486b40e919a74a9cf34c2aa63\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.354783 containerd[1604]: time="2025-09-16T04:49:59.354727460Z" level=error msg="Failed to destroy network for sandbox \"c29be128c644912c775f70c228078073e6181221e8d3cfde9fb6371b117075f8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.367147 containerd[1604]: time="2025-09-16T04:49:59.367102067Z" level=error msg="Failed to destroy network for sandbox \"2dd6c9f7117acc51335b1a19564e9bd5eb4f8aec17eb7966f2c406be5ed39f09\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.392365 containerd[1604]: time="2025-09-16T04:49:59.392298800Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s4499,Uid:805af22b-e7a5-4562-8fa7-b331a6a81d08,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee38da718b2fab6e5f80696975e66627accdbb3a64e9198ee53d70fd0b5403e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.401498 kubelet[2756]: E0916 04:49:59.401411 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee38da718b2fab6e5f80696975e66627accdbb3a64e9198ee53d70fd0b5403e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.401604 kubelet[2756]: E0916 04:49:59.401544 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee38da718b2fab6e5f80696975e66627accdbb3a64e9198ee53d70fd0b5403e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s4499" Sep 16 04:49:59.401604 kubelet[2756]: E0916 04:49:59.401580 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee38da718b2fab6e5f80696975e66627accdbb3a64e9198ee53d70fd0b5403e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s4499" Sep 16 04:49:59.401704 kubelet[2756]: E0916 04:49:59.401651 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s4499_kube-system(805af22b-e7a5-4562-8fa7-b331a6a81d08)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s4499_kube-system(805af22b-e7a5-4562-8fa7-b331a6a81d08)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dee38da718b2fab6e5f80696975e66627accdbb3a64e9198ee53d70fd0b5403e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s4499" podUID="805af22b-e7a5-4562-8fa7-b331a6a81d08" Sep 16 04:49:59.458067 containerd[1604]: time="2025-09-16T04:49:59.457870076Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-2wvgh,Uid:1b3a3c6b-e0ed-437a-9826-214673136666,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a62a52d87b781e0cc6a275bae29735a5328ae735fabae6a2f1d0c0cf875e278\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.458249 kubelet[2756]: E0916 04:49:59.458174 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a62a52d87b781e0cc6a275bae29735a5328ae735fabae6a2f1d0c0cf875e278\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.458312 kubelet[2756]: E0916 04:49:59.458296 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a62a52d87b781e0cc6a275bae29735a5328ae735fabae6a2f1d0c0cf875e278\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-2wvgh" Sep 16 04:49:59.458343 kubelet[2756]: E0916 04:49:59.458330 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a62a52d87b781e0cc6a275bae29735a5328ae735fabae6a2f1d0c0cf875e278\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-2wvgh" Sep 16 04:49:59.458444 kubelet[2756]: E0916 04:49:59.458384 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-2wvgh_kube-system(1b3a3c6b-e0ed-437a-9826-214673136666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-2wvgh_kube-system(1b3a3c6b-e0ed-437a-9826-214673136666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8a62a52d87b781e0cc6a275bae29735a5328ae735fabae6a2f1d0c0cf875e278\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-2wvgh" podUID="1b3a3c6b-e0ed-437a-9826-214673136666" Sep 16 04:49:59.478303 containerd[1604]: time="2025-09-16T04:49:59.478205419Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-b8d5c47bf-nrnjv,Uid:735decdf-fb16-4265-96d7-4da1d18f1679,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"05612b3cb5e060a236c79ca2f666d554d95bf13f6a331b4f924fd2296cebd62c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.479286 kubelet[2756]: E0916 04:49:59.478595 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"05612b3cb5e060a236c79ca2f666d554d95bf13f6a331b4f924fd2296cebd62c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.479286 kubelet[2756]: E0916 04:49:59.478661 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"05612b3cb5e060a236c79ca2f666d554d95bf13f6a331b4f924fd2296cebd62c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" Sep 16 04:49:59.479286 kubelet[2756]: E0916 04:49:59.478685 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"05612b3cb5e060a236c79ca2f666d554d95bf13f6a331b4f924fd2296cebd62c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" Sep 16 04:49:59.479445 kubelet[2756]: E0916 04:49:59.478722 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-b8d5c47bf-nrnjv_calico-system(735decdf-fb16-4265-96d7-4da1d18f1679)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-b8d5c47bf-nrnjv_calico-system(735decdf-fb16-4265-96d7-4da1d18f1679)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"05612b3cb5e060a236c79ca2f666d554d95bf13f6a331b4f924fd2296cebd62c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" podUID="735decdf-fb16-4265-96d7-4da1d18f1679" Sep 16 04:49:59.480177 containerd[1604]: time="2025-09-16T04:49:59.480102106Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-mdn5g,Uid:7da891ca-d373-47fb-8d2b-a8bb2752d2c0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8f5add85341627848560882de5f95f1ffcbbc2c2d960f39073484db40b34d6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.480496 kubelet[2756]: E0916 04:49:59.480457 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8f5add85341627848560882de5f95f1ffcbbc2c2d960f39073484db40b34d6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.480581 kubelet[2756]: E0916 04:49:59.480526 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8f5add85341627848560882de5f95f1ffcbbc2c2d960f39073484db40b34d6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-mdn5g" Sep 16 04:49:59.480581 kubelet[2756]: E0916 04:49:59.480543 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8f5add85341627848560882de5f95f1ffcbbc2c2d960f39073484db40b34d6d3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-mdn5g" Sep 16 04:49:59.480651 kubelet[2756]: E0916 04:49:59.480578 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-mdn5g_calico-system(7da891ca-d373-47fb-8d2b-a8bb2752d2c0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-mdn5g_calico-system(7da891ca-d373-47fb-8d2b-a8bb2752d2c0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8f5add85341627848560882de5f95f1ffcbbc2c2d960f39073484db40b34d6d3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-mdn5g" podUID="7da891ca-d373-47fb-8d2b-a8bb2752d2c0" Sep 16 04:49:59.482196 containerd[1604]: time="2025-09-16T04:49:59.482116514Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-66c667cbbc-gbsmd,Uid:61a5328c-f236-4ba6-9b78-7bb5afd28a2f,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0118c560c86cc8ebf93f82e59d97b57268024af486b40e919a74a9cf34c2aa63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.482320 kubelet[2756]: E0916 04:49:59.482290 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0118c560c86cc8ebf93f82e59d97b57268024af486b40e919a74a9cf34c2aa63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.482362 kubelet[2756]: E0916 04:49:59.482336 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0118c560c86cc8ebf93f82e59d97b57268024af486b40e919a74a9cf34c2aa63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-66c667cbbc-gbsmd" Sep 16 04:49:59.482362 kubelet[2756]: E0916 04:49:59.482352 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0118c560c86cc8ebf93f82e59d97b57268024af486b40e919a74a9cf34c2aa63\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-66c667cbbc-gbsmd" Sep 16 04:49:59.482419 kubelet[2756]: E0916 04:49:59.482382 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-66c667cbbc-gbsmd_calico-system(61a5328c-f236-4ba6-9b78-7bb5afd28a2f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-66c667cbbc-gbsmd_calico-system(61a5328c-f236-4ba6-9b78-7bb5afd28a2f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0118c560c86cc8ebf93f82e59d97b57268024af486b40e919a74a9cf34c2aa63\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-66c667cbbc-gbsmd" podUID="61a5328c-f236-4ba6-9b78-7bb5afd28a2f" Sep 16 04:49:59.483689 containerd[1604]: time="2025-09-16T04:49:59.483624209Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-6bfxs,Uid:3e572577-68f4-49a7-9725-3f0cb8aef839,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c29be128c644912c775f70c228078073e6181221e8d3cfde9fb6371b117075f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.484535 kubelet[2756]: E0916 04:49:59.483863 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c29be128c644912c775f70c228078073e6181221e8d3cfde9fb6371b117075f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.484535 kubelet[2756]: E0916 04:49:59.483895 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c29be128c644912c775f70c228078073e6181221e8d3cfde9fb6371b117075f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" Sep 16 04:49:59.484535 kubelet[2756]: E0916 04:49:59.483910 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c29be128c644912c775f70c228078073e6181221e8d3cfde9fb6371b117075f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" Sep 16 04:49:59.484647 kubelet[2756]: E0916 04:49:59.483942 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6b89c956bf-6bfxs_calico-apiserver(3e572577-68f4-49a7-9725-3f0cb8aef839)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6b89c956bf-6bfxs_calico-apiserver(3e572577-68f4-49a7-9725-3f0cb8aef839)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c29be128c644912c775f70c228078073e6181221e8d3cfde9fb6371b117075f8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" podUID="3e572577-68f4-49a7-9725-3f0cb8aef839" Sep 16 04:49:59.485163 containerd[1604]: time="2025-09-16T04:49:59.485115251Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-92rwh,Uid:d7bf725a-401b-4090-b1b5-801486b6473f,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd6c9f7117acc51335b1a19564e9bd5eb4f8aec17eb7966f2c406be5ed39f09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.485434 kubelet[2756]: E0916 04:49:59.485257 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd6c9f7117acc51335b1a19564e9bd5eb4f8aec17eb7966f2c406be5ed39f09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.485434 kubelet[2756]: E0916 04:49:59.485284 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd6c9f7117acc51335b1a19564e9bd5eb4f8aec17eb7966f2c406be5ed39f09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b89c956bf-92rwh" Sep 16 04:49:59.485434 kubelet[2756]: E0916 04:49:59.485298 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd6c9f7117acc51335b1a19564e9bd5eb4f8aec17eb7966f2c406be5ed39f09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b89c956bf-92rwh" Sep 16 04:49:59.485634 kubelet[2756]: E0916 04:49:59.485348 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6b89c956bf-92rwh_calico-apiserver(d7bf725a-401b-4090-b1b5-801486b6473f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6b89c956bf-92rwh_calico-apiserver(d7bf725a-401b-4090-b1b5-801486b6473f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2dd6c9f7117acc51335b1a19564e9bd5eb4f8aec17eb7966f2c406be5ed39f09\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b89c956bf-92rwh" podUID="d7bf725a-401b-4090-b1b5-801486b6473f" Sep 16 04:49:59.653024 systemd[1]: Created slice kubepods-besteffort-pod7fb6a3cc_bcfc_4ebe_8396_a7a64f08a3e4.slice - libcontainer container kubepods-besteffort-pod7fb6a3cc_bcfc_4ebe_8396_a7a64f08a3e4.slice. Sep 16 04:49:59.655894 containerd[1604]: time="2025-09-16T04:49:59.655848129Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9872f,Uid:7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4,Namespace:calico-system,Attempt:0,}" Sep 16 04:49:59.733648 containerd[1604]: time="2025-09-16T04:49:59.733601218Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 16 04:49:59.814716 containerd[1604]: time="2025-09-16T04:49:59.814645285Z" level=error msg="Failed to destroy network for sandbox \"2b8952ce05082f675bcfc013e41c38697c373c0783bc76b1b65a5c8f1b3f2076\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.816222 containerd[1604]: time="2025-09-16T04:49:59.816175973Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9872f,Uid:7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b8952ce05082f675bcfc013e41c38697c373c0783bc76b1b65a5c8f1b3f2076\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.816561 kubelet[2756]: E0916 04:49:59.816442 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b8952ce05082f675bcfc013e41c38697c373c0783bc76b1b65a5c8f1b3f2076\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:49:59.817047 kubelet[2756]: E0916 04:49:59.816566 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b8952ce05082f675bcfc013e41c38697c373c0783bc76b1b65a5c8f1b3f2076\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:59.817047 kubelet[2756]: E0916 04:49:59.816596 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b8952ce05082f675bcfc013e41c38697c373c0783bc76b1b65a5c8f1b3f2076\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-9872f" Sep 16 04:49:59.817047 kubelet[2756]: E0916 04:49:59.816664 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-9872f_calico-system(7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-9872f_calico-system(7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2b8952ce05082f675bcfc013e41c38697c373c0783bc76b1b65a5c8f1b3f2076\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-9872f" podUID="7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4" Sep 16 04:50:00.028863 systemd[1]: run-netns-cni\x2df548fabc\x2daded\x2d1c3c\x2d8c69\x2d218b451af7f3.mount: Deactivated successfully. Sep 16 04:50:00.029002 systemd[1]: run-netns-cni\x2d2d41dbd8\x2d0b94\x2d78b1\x2db68b\x2d6462d7af909e.mount: Deactivated successfully. Sep 16 04:50:00.029098 systemd[1]: run-netns-cni\x2d763da780\x2db8fd\x2dbb2b\x2d1b2b\x2d167c4a69aa5b.mount: Deactivated successfully. Sep 16 04:50:09.641826 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1776106583.mount: Deactivated successfully. Sep 16 04:50:10.644138 containerd[1604]: time="2025-09-16T04:50:10.644072204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-b8d5c47bf-nrnjv,Uid:735decdf-fb16-4265-96d7-4da1d18f1679,Namespace:calico-system,Attempt:0,}" Sep 16 04:50:10.679944 containerd[1604]: time="2025-09-16T04:50:10.679872581Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:10.815674 containerd[1604]: time="2025-09-16T04:50:10.786659718Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Sep 16 04:50:10.949930 containerd[1604]: time="2025-09-16T04:50:10.949747042Z" level=error msg="Failed to destroy network for sandbox \"ecac552ac8e73a4000ccd75c3f2710dc4b010c6444e8a1482e25eba36540e1ba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:50:10.950274 containerd[1604]: time="2025-09-16T04:50:10.950079917Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:10.953049 systemd[1]: run-netns-cni\x2d73fddae7\x2dcc01\x2d3f9f\x2de091\x2dc94398df826c.mount: Deactivated successfully. Sep 16 04:50:11.220699 containerd[1604]: time="2025-09-16T04:50:11.220467126Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-b8d5c47bf-nrnjv,Uid:735decdf-fb16-4265-96d7-4da1d18f1679,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ecac552ac8e73a4000ccd75c3f2710dc4b010c6444e8a1482e25eba36540e1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:50:11.221082 kubelet[2756]: E0916 04:50:11.220967 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ecac552ac8e73a4000ccd75c3f2710dc4b010c6444e8a1482e25eba36540e1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:50:11.221082 kubelet[2756]: E0916 04:50:11.221083 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ecac552ac8e73a4000ccd75c3f2710dc4b010c6444e8a1482e25eba36540e1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" Sep 16 04:50:11.221862 kubelet[2756]: E0916 04:50:11.221132 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ecac552ac8e73a4000ccd75c3f2710dc4b010c6444e8a1482e25eba36540e1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" Sep 16 04:50:11.221862 kubelet[2756]: E0916 04:50:11.221205 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-b8d5c47bf-nrnjv_calico-system(735decdf-fb16-4265-96d7-4da1d18f1679)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-b8d5c47bf-nrnjv_calico-system(735decdf-fb16-4265-96d7-4da1d18f1679)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ecac552ac8e73a4000ccd75c3f2710dc4b010c6444e8a1482e25eba36540e1ba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" podUID="735decdf-fb16-4265-96d7-4da1d18f1679" Sep 16 04:50:11.268790 containerd[1604]: time="2025-09-16T04:50:11.268707421Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:11.269537 containerd[1604]: time="2025-09-16T04:50:11.269465905Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 11.535817229s" Sep 16 04:50:11.269593 containerd[1604]: time="2025-09-16T04:50:11.269546807Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Sep 16 04:50:11.283076 containerd[1604]: time="2025-09-16T04:50:11.283026994Z" level=info msg="CreateContainer within sandbox \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 16 04:50:11.512208 containerd[1604]: time="2025-09-16T04:50:11.511311573Z" level=info msg="Container 85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:11.629640 systemd[1]: Started sshd@7-10.0.0.34:22-10.0.0.1:43806.service - OpenSSH per-connection server daemon (10.0.0.1:43806). Sep 16 04:50:11.643893 containerd[1604]: time="2025-09-16T04:50:11.643821534Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-6bfxs,Uid:3e572577-68f4-49a7-9725-3f0cb8aef839,Namespace:calico-apiserver,Attempt:0,}" Sep 16 04:50:11.712145 containerd[1604]: time="2025-09-16T04:50:11.712070935Z" level=info msg="CreateContainer within sandbox \"3f8747321696052e2b37960cbd7a2618d0dd37c71c45139a55a887592d9e71a7\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\"" Sep 16 04:50:11.712841 containerd[1604]: time="2025-09-16T04:50:11.712773384Z" level=info msg="StartContainer for \"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\"" Sep 16 04:50:11.714482 containerd[1604]: time="2025-09-16T04:50:11.714414485Z" level=info msg="connecting to shim 85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95" address="unix:///run/containerd/s/0a935b1dba4442345aa8a2636b425529a3234faea10e38d26c4f540786a8b101" protocol=ttrpc version=3 Sep 16 04:50:11.738698 systemd[1]: Started cri-containerd-85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95.scope - libcontainer container 85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95. Sep 16 04:50:11.767096 sshd[3905]: Accepted publickey for core from 10.0.0.1 port 43806 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:11.772365 sshd-session[3905]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:11.784893 systemd-logind[1587]: New session 8 of user core. Sep 16 04:50:11.792894 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 16 04:50:11.866562 containerd[1604]: time="2025-09-16T04:50:11.865702560Z" level=error msg="Failed to destroy network for sandbox \"84fb8d42cd69ad7a68507ac6d3cbb108e1b0fd1f64d73deef54b4e56fc50cbd1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:50:11.870857 systemd[1]: run-netns-cni\x2d15ca9eec\x2d4661\x2d337e\x2d3c54\x2de09b04b49a5a.mount: Deactivated successfully. Sep 16 04:50:11.880012 containerd[1604]: time="2025-09-16T04:50:11.879943726Z" level=info msg="StartContainer for \"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\" returns successfully" Sep 16 04:50:11.934746 containerd[1604]: time="2025-09-16T04:50:11.934648198Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-6bfxs,Uid:3e572577-68f4-49a7-9725-3f0cb8aef839,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"84fb8d42cd69ad7a68507ac6d3cbb108e1b0fd1f64d73deef54b4e56fc50cbd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:50:11.935022 kubelet[2756]: E0916 04:50:11.934966 2756 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84fb8d42cd69ad7a68507ac6d3cbb108e1b0fd1f64d73deef54b4e56fc50cbd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 16 04:50:11.935091 kubelet[2756]: E0916 04:50:11.935040 2756 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84fb8d42cd69ad7a68507ac6d3cbb108e1b0fd1f64d73deef54b4e56fc50cbd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" Sep 16 04:50:11.935091 kubelet[2756]: E0916 04:50:11.935063 2756 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84fb8d42cd69ad7a68507ac6d3cbb108e1b0fd1f64d73deef54b4e56fc50cbd1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" Sep 16 04:50:11.935183 kubelet[2756]: E0916 04:50:11.935107 2756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6b89c956bf-6bfxs_calico-apiserver(3e572577-68f4-49a7-9725-3f0cb8aef839)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6b89c956bf-6bfxs_calico-apiserver(3e572577-68f4-49a7-9725-3f0cb8aef839)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"84fb8d42cd69ad7a68507ac6d3cbb108e1b0fd1f64d73deef54b4e56fc50cbd1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" podUID="3e572577-68f4-49a7-9725-3f0cb8aef839" Sep 16 04:50:11.977840 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 16 04:50:11.978013 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 16 04:50:11.993641 sshd[3945]: Connection closed by 10.0.0.1 port 43806 Sep 16 04:50:11.993794 sshd-session[3905]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:12.002192 systemd[1]: sshd@7-10.0.0.34:22-10.0.0.1:43806.service: Deactivated successfully. Sep 16 04:50:12.005995 systemd[1]: session-8.scope: Deactivated successfully. Sep 16 04:50:12.007299 systemd-logind[1587]: Session 8 logged out. Waiting for processes to exit. Sep 16 04:50:12.010624 systemd-logind[1587]: Removed session 8. Sep 16 04:50:12.405476 kubelet[2756]: I0916 04:50:12.405394 2756 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp66b\" (UniqueName: \"kubernetes.io/projected/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-kube-api-access-pp66b\") pod \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\" (UID: \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\") " Sep 16 04:50:12.405476 kubelet[2756]: I0916 04:50:12.405480 2756 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-ca-bundle\") pod \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\" (UID: \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\") " Sep 16 04:50:12.406279 kubelet[2756]: I0916 04:50:12.405552 2756 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-backend-key-pair\") pod \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\" (UID: \"61a5328c-f236-4ba6-9b78-7bb5afd28a2f\") " Sep 16 04:50:12.406555 kubelet[2756]: I0916 04:50:12.406480 2756 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "61a5328c-f236-4ba6-9b78-7bb5afd28a2f" (UID: "61a5328c-f236-4ba6-9b78-7bb5afd28a2f"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 16 04:50:12.410342 kubelet[2756]: I0916 04:50:12.410285 2756 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "61a5328c-f236-4ba6-9b78-7bb5afd28a2f" (UID: "61a5328c-f236-4ba6-9b78-7bb5afd28a2f"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 16 04:50:12.410709 kubelet[2756]: I0916 04:50:12.410676 2756 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-kube-api-access-pp66b" (OuterVolumeSpecName: "kube-api-access-pp66b") pod "61a5328c-f236-4ba6-9b78-7bb5afd28a2f" (UID: "61a5328c-f236-4ba6-9b78-7bb5afd28a2f"). InnerVolumeSpecName "kube-api-access-pp66b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 16 04:50:12.412154 systemd[1]: var-lib-kubelet-pods-61a5328c\x2df236\x2d4ba6\x2d9b78\x2d7bb5afd28a2f-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dpp66b.mount: Deactivated successfully. Sep 16 04:50:12.412327 systemd[1]: var-lib-kubelet-pods-61a5328c\x2df236\x2d4ba6\x2d9b78\x2d7bb5afd28a2f-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 16 04:50:12.507549 kubelet[2756]: I0916 04:50:12.505865 2756 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Sep 16 04:50:12.507549 kubelet[2756]: I0916 04:50:12.505924 2756 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Sep 16 04:50:12.507549 kubelet[2756]: I0916 04:50:12.505934 2756 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pp66b\" (UniqueName: \"kubernetes.io/projected/61a5328c-f236-4ba6-9b78-7bb5afd28a2f-kube-api-access-pp66b\") on node \"localhost\" DevicePath \"\"" Sep 16 04:50:12.792669 systemd[1]: Removed slice kubepods-besteffort-pod61a5328c_f236_4ba6_9b78_7bb5afd28a2f.slice - libcontainer container kubepods-besteffort-pod61a5328c_f236_4ba6_9b78_7bb5afd28a2f.slice. Sep 16 04:50:12.804143 kubelet[2756]: I0916 04:50:12.803971 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-pswhm" podStartSLOduration=2.973262914 podStartE2EDuration="25.803936497s" podCreationTimestamp="2025-09-16 04:49:47 +0000 UTC" firstStartedPulling="2025-09-16 04:49:48.440040676 +0000 UTC m=+20.912800245" lastFinishedPulling="2025-09-16 04:50:11.270714259 +0000 UTC m=+43.743473828" observedRunningTime="2025-09-16 04:50:12.802568288 +0000 UTC m=+45.275327877" watchObservedRunningTime="2025-09-16 04:50:12.803936497 +0000 UTC m=+45.276696066" Sep 16 04:50:12.862765 systemd[1]: Created slice kubepods-besteffort-pod164847fe_3858_42e5_bd8c_5fa72a31fdea.slice - libcontainer container kubepods-besteffort-pod164847fe_3858_42e5_bd8c_5fa72a31fdea.slice. Sep 16 04:50:12.909472 kubelet[2756]: I0916 04:50:12.909308 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164847fe-3858-42e5-bd8c-5fa72a31fdea-whisker-ca-bundle\") pod \"whisker-79dc8ddbd6-2vn8p\" (UID: \"164847fe-3858-42e5-bd8c-5fa72a31fdea\") " pod="calico-system/whisker-79dc8ddbd6-2vn8p" Sep 16 04:50:12.910654 kubelet[2756]: I0916 04:50:12.909749 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4rj7\" (UniqueName: \"kubernetes.io/projected/164847fe-3858-42e5-bd8c-5fa72a31fdea-kube-api-access-p4rj7\") pod \"whisker-79dc8ddbd6-2vn8p\" (UID: \"164847fe-3858-42e5-bd8c-5fa72a31fdea\") " pod="calico-system/whisker-79dc8ddbd6-2vn8p" Sep 16 04:50:12.910654 kubelet[2756]: I0916 04:50:12.910581 2756 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/164847fe-3858-42e5-bd8c-5fa72a31fdea-whisker-backend-key-pair\") pod \"whisker-79dc8ddbd6-2vn8p\" (UID: \"164847fe-3858-42e5-bd8c-5fa72a31fdea\") " pod="calico-system/whisker-79dc8ddbd6-2vn8p" Sep 16 04:50:12.946054 containerd[1604]: time="2025-09-16T04:50:12.945992108Z" level=info msg="TaskExit event in podsandbox handler container_id:\"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\" id:\"52d2cdeec165f9261c891b98c3a37cd5a475e1174acb23fd4818c6f6617af0a8\" pid:4032 exit_status:1 exited_at:{seconds:1757998212 nanos:945599010}" Sep 16 04:50:13.170752 containerd[1604]: time="2025-09-16T04:50:13.170602091Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-79dc8ddbd6-2vn8p,Uid:164847fe-3858-42e5-bd8c-5fa72a31fdea,Namespace:calico-system,Attempt:0,}" Sep 16 04:50:13.459177 systemd-networkd[1495]: cali5f6f863d32b: Link UP Sep 16 04:50:13.459720 systemd-networkd[1495]: cali5f6f863d32b: Gained carrier Sep 16 04:50:13.481008 containerd[1604]: 2025-09-16 04:50:13.220 [INFO][4047] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 16 04:50:13.481008 containerd[1604]: 2025-09-16 04:50:13.262 [INFO][4047] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0 whisker-79dc8ddbd6- calico-system 164847fe-3858-42e5-bd8c-5fa72a31fdea 988 0 2025-09-16 04:50:12 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:79dc8ddbd6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-79dc8ddbd6-2vn8p eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali5f6f863d32b [] [] }} ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-" Sep 16 04:50:13.481008 containerd[1604]: 2025-09-16 04:50:13.262 [INFO][4047] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.481008 containerd[1604]: 2025-09-16 04:50:13.402 [INFO][4061] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" HandleID="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Workload="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.404 [INFO][4061] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" HandleID="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Workload="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fa70), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-79dc8ddbd6-2vn8p", "timestamp":"2025-09-16 04:50:13.402308025 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.404 [INFO][4061] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.406 [INFO][4061] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.406 [INFO][4061] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.416 [INFO][4061] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" host="localhost" Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.423 [INFO][4061] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.428 [INFO][4061] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.429 [INFO][4061] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.432 [INFO][4061] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:13.481363 containerd[1604]: 2025-09-16 04:50:13.432 [INFO][4061] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" host="localhost" Sep 16 04:50:13.481721 containerd[1604]: 2025-09-16 04:50:13.434 [INFO][4061] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650 Sep 16 04:50:13.481721 containerd[1604]: 2025-09-16 04:50:13.439 [INFO][4061] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" host="localhost" Sep 16 04:50:13.481721 containerd[1604]: 2025-09-16 04:50:13.445 [INFO][4061] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" host="localhost" Sep 16 04:50:13.481721 containerd[1604]: 2025-09-16 04:50:13.445 [INFO][4061] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" host="localhost" Sep 16 04:50:13.481721 containerd[1604]: 2025-09-16 04:50:13.445 [INFO][4061] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:13.481721 containerd[1604]: 2025-09-16 04:50:13.445 [INFO][4061] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" HandleID="k8s-pod-network.1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Workload="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.481891 containerd[1604]: 2025-09-16 04:50:13.449 [INFO][4047] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0", GenerateName:"whisker-79dc8ddbd6-", Namespace:"calico-system", SelfLink:"", UID:"164847fe-3858-42e5-bd8c-5fa72a31fdea", ResourceVersion:"988", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 50, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"79dc8ddbd6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-79dc8ddbd6-2vn8p", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5f6f863d32b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:13.481891 containerd[1604]: 2025-09-16 04:50:13.449 [INFO][4047] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.481997 containerd[1604]: 2025-09-16 04:50:13.449 [INFO][4047] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5f6f863d32b ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.481997 containerd[1604]: 2025-09-16 04:50:13.460 [INFO][4047] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.482058 containerd[1604]: 2025-09-16 04:50:13.461 [INFO][4047] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0", GenerateName:"whisker-79dc8ddbd6-", Namespace:"calico-system", SelfLink:"", UID:"164847fe-3858-42e5-bd8c-5fa72a31fdea", ResourceVersion:"988", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 50, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"79dc8ddbd6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650", Pod:"whisker-79dc8ddbd6-2vn8p", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali5f6f863d32b", MAC:"b2:0c:0f:7c:c2:bd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:13.482125 containerd[1604]: 2025-09-16 04:50:13.475 [INFO][4047] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" Namespace="calico-system" Pod="whisker-79dc8ddbd6-2vn8p" WorkloadEndpoint="localhost-k8s-whisker--79dc8ddbd6--2vn8p-eth0" Sep 16 04:50:13.534560 containerd[1604]: time="2025-09-16T04:50:13.534473863Z" level=info msg="connecting to shim 1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650" address="unix:///run/containerd/s/114ea8b83d913ec925f55f13dc091246f4c440cfac994aeccaddf406ade42d12" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:13.572739 systemd[1]: Started cri-containerd-1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650.scope - libcontainer container 1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650. Sep 16 04:50:13.586997 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:13.618121 containerd[1604]: time="2025-09-16T04:50:13.618065632Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-79dc8ddbd6-2vn8p,Uid:164847fe-3858-42e5-bd8c-5fa72a31fdea,Namespace:calico-system,Attempt:0,} returns sandbox id \"1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650\"" Sep 16 04:50:13.619650 containerd[1604]: time="2025-09-16T04:50:13.619615261Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 16 04:50:13.643928 kubelet[2756]: E0916 04:50:13.643620 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:13.643928 kubelet[2756]: E0916 04:50:13.643785 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:13.645363 containerd[1604]: time="2025-09-16T04:50:13.645315272Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-2wvgh,Uid:1b3a3c6b-e0ed-437a-9826-214673136666,Namespace:kube-system,Attempt:0,}" Sep 16 04:50:13.645573 containerd[1604]: time="2025-09-16T04:50:13.645379463Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-92rwh,Uid:d7bf725a-401b-4090-b1b5-801486b6473f,Namespace:calico-apiserver,Attempt:0,}" Sep 16 04:50:13.645573 containerd[1604]: time="2025-09-16T04:50:13.645403429Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s4499,Uid:805af22b-e7a5-4562-8fa7-b331a6a81d08,Namespace:kube-system,Attempt:0,}" Sep 16 04:50:13.645573 containerd[1604]: time="2025-09-16T04:50:13.645328467Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9872f,Uid:7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4,Namespace:calico-system,Attempt:0,}" Sep 16 04:50:13.647110 kubelet[2756]: I0916 04:50:13.647005 2756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61a5328c-f236-4ba6-9b78-7bb5afd28a2f" path="/var/lib/kubelet/pods/61a5328c-f236-4ba6-9b78-7bb5afd28a2f/volumes" Sep 16 04:50:13.839751 systemd-networkd[1495]: cali1101966f210: Link UP Sep 16 04:50:13.841047 systemd-networkd[1495]: cali1101966f210: Gained carrier Sep 16 04:50:13.877188 containerd[1604]: 2025-09-16 04:50:13.704 [INFO][4135] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 16 04:50:13.877188 containerd[1604]: 2025-09-16 04:50:13.727 [INFO][4135] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--9872f-eth0 csi-node-driver- calico-system 7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4 750 0 2025-09-16 04:49:48 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-9872f eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali1101966f210 [] [] }} ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-" Sep 16 04:50:13.877188 containerd[1604]: 2025-09-16 04:50:13.727 [INFO][4135] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:13.877188 containerd[1604]: 2025-09-16 04:50:13.777 [INFO][4183] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" HandleID="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Workload="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.779 [INFO][4183] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" HandleID="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Workload="localhost-k8s-csi--node--driver--9872f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f600), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-9872f", "timestamp":"2025-09-16 04:50:13.777964869 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.779 [INFO][4183] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.779 [INFO][4183] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.779 [INFO][4183] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.785 [INFO][4183] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" host="localhost" Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.793 [INFO][4183] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.799 [INFO][4183] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.802 [INFO][4183] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.805 [INFO][4183] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:13.877653 containerd[1604]: 2025-09-16 04:50:13.805 [INFO][4183] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" host="localhost" Sep 16 04:50:13.878716 containerd[1604]: 2025-09-16 04:50:13.810 [INFO][4183] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4 Sep 16 04:50:13.878716 containerd[1604]: 2025-09-16 04:50:13.816 [INFO][4183] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" host="localhost" Sep 16 04:50:13.878716 containerd[1604]: 2025-09-16 04:50:13.823 [INFO][4183] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" host="localhost" Sep 16 04:50:13.878716 containerd[1604]: 2025-09-16 04:50:13.823 [INFO][4183] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" host="localhost" Sep 16 04:50:13.878716 containerd[1604]: 2025-09-16 04:50:13.825 [INFO][4183] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:13.878716 containerd[1604]: 2025-09-16 04:50:13.825 [INFO][4183] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" HandleID="k8s-pod-network.28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Workload="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:13.878908 containerd[1604]: 2025-09-16 04:50:13.833 [INFO][4135] cni-plugin/k8s.go 418: Populated endpoint ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--9872f-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4", ResourceVersion:"750", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-9872f", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali1101966f210", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:13.878988 containerd[1604]: 2025-09-16 04:50:13.833 [INFO][4135] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:13.878988 containerd[1604]: 2025-09-16 04:50:13.833 [INFO][4135] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1101966f210 ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:13.878988 containerd[1604]: 2025-09-16 04:50:13.842 [INFO][4135] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:13.879092 containerd[1604]: 2025-09-16 04:50:13.842 [INFO][4135] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--9872f-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4", ResourceVersion:"750", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4", Pod:"csi-node-driver-9872f", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali1101966f210", MAC:"ee:99:2c:7b:50:80", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:13.879165 containerd[1604]: 2025-09-16 04:50:13.864 [INFO][4135] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" Namespace="calico-system" Pod="csi-node-driver-9872f" WorkloadEndpoint="localhost-k8s-csi--node--driver--9872f-eth0" Sep 16 04:50:14.029787 systemd-networkd[1495]: cali2abca3fdc34: Link UP Sep 16 04:50:14.031653 systemd-networkd[1495]: cali2abca3fdc34: Gained carrier Sep 16 04:50:14.061835 containerd[1604]: 2025-09-16 04:50:13.734 [INFO][4158] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 16 04:50:14.061835 containerd[1604]: 2025-09-16 04:50:13.756 [INFO][4158] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0 coredns-668d6bf9bc- kube-system 1b3a3c6b-e0ed-437a-9826-214673136666 867 0 2025-09-16 04:49:32 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-2wvgh eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2abca3fdc34 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-" Sep 16 04:50:14.061835 containerd[1604]: 2025-09-16 04:50:13.756 [INFO][4158] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.061835 containerd[1604]: 2025-09-16 04:50:13.808 [INFO][4203] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" HandleID="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Workload="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.808 [INFO][4203] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" HandleID="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Workload="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e75f0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-2wvgh", "timestamp":"2025-09-16 04:50:13.808712264 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.808 [INFO][4203] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.825 [INFO][4203] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.825 [INFO][4203] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.891 [INFO][4203] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" host="localhost" Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.900 [INFO][4203] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.910 [INFO][4203] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.914 [INFO][4203] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.916 [INFO][4203] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:14.063454 containerd[1604]: 2025-09-16 04:50:13.916 [INFO][4203] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" host="localhost" Sep 16 04:50:14.064129 containerd[1604]: 2025-09-16 04:50:13.919 [INFO][4203] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5 Sep 16 04:50:14.064129 containerd[1604]: 2025-09-16 04:50:13.929 [INFO][4203] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" host="localhost" Sep 16 04:50:14.064129 containerd[1604]: 2025-09-16 04:50:14.016 [INFO][4203] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" host="localhost" Sep 16 04:50:14.064129 containerd[1604]: 2025-09-16 04:50:14.016 [INFO][4203] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" host="localhost" Sep 16 04:50:14.064129 containerd[1604]: 2025-09-16 04:50:14.017 [INFO][4203] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:14.064129 containerd[1604]: 2025-09-16 04:50:14.017 [INFO][4203] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" HandleID="k8s-pod-network.6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Workload="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.064336 containerd[1604]: 2025-09-16 04:50:14.022 [INFO][4158] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1b3a3c6b-e0ed-437a-9826-214673136666", ResourceVersion:"867", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-2wvgh", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2abca3fdc34", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:14.064445 containerd[1604]: 2025-09-16 04:50:14.022 [INFO][4158] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.064445 containerd[1604]: 2025-09-16 04:50:14.022 [INFO][4158] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2abca3fdc34 ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.064445 containerd[1604]: 2025-09-16 04:50:14.031 [INFO][4158] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.066692 containerd[1604]: 2025-09-16 04:50:14.036 [INFO][4158] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1b3a3c6b-e0ed-437a-9826-214673136666", ResourceVersion:"867", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5", Pod:"coredns-668d6bf9bc-2wvgh", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2abca3fdc34", MAC:"ce:0f:f7:60:9e:95", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:14.066692 containerd[1604]: 2025-09-16 04:50:14.050 [INFO][4158] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" Namespace="kube-system" Pod="coredns-668d6bf9bc-2wvgh" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--2wvgh-eth0" Sep 16 04:50:14.081005 containerd[1604]: time="2025-09-16T04:50:14.080261648Z" level=info msg="connecting to shim 28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4" address="unix:///run/containerd/s/3f7e5dec723747c230d72b31987fe5ff9ba6ab1a59059e555d6f8b2de5ca1c32" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:14.127819 systemd[1]: Started cri-containerd-28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4.scope - libcontainer container 28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4. Sep 16 04:50:14.139709 containerd[1604]: time="2025-09-16T04:50:14.139652182Z" level=info msg="TaskExit event in podsandbox handler container_id:\"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\" id:\"7684044081c74d21c785c237eb3fb8e495dac14784bb9bb818e1b6d8a70421c9\" pid:4241 exit_status:1 exited_at:{seconds:1757998214 nanos:138316796}" Sep 16 04:50:14.148131 systemd-networkd[1495]: cali030ea07f401: Link UP Sep 16 04:50:14.154581 systemd-networkd[1495]: cali030ea07f401: Gained carrier Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:13.703 [INFO][4146] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:13.728 [INFO][4146] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--s4499-eth0 coredns-668d6bf9bc- kube-system 805af22b-e7a5-4562-8fa7-b331a6a81d08 871 0 2025-09-16 04:49:32 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-s4499 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali030ea07f401 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:13.728 [INFO][4146] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:13.799 [INFO][4186] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" HandleID="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Workload="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:13.804 [INFO][4186] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" HandleID="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Workload="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000324920), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-s4499", "timestamp":"2025-09-16 04:50:13.799680901 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:13.810 [INFO][4186] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.016 [INFO][4186] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.017 [INFO][4186] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.036 [INFO][4186] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.051 [INFO][4186] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.067 [INFO][4186] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.075 [INFO][4186] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.087 [INFO][4186] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.088 [INFO][4186] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.090 [INFO][4186] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1 Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.124 [INFO][4186] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.135 [INFO][4186] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.135 [INFO][4186] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" host="localhost" Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.136 [INFO][4186] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:14.176852 containerd[1604]: 2025-09-16 04:50:14.136 [INFO][4186] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" HandleID="k8s-pod-network.f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Workload="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.177642 containerd[1604]: 2025-09-16 04:50:14.144 [INFO][4146] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--s4499-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"805af22b-e7a5-4562-8fa7-b331a6a81d08", ResourceVersion:"871", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-s4499", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali030ea07f401", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:14.177642 containerd[1604]: 2025-09-16 04:50:14.144 [INFO][4146] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.177642 containerd[1604]: 2025-09-16 04:50:14.144 [INFO][4146] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali030ea07f401 ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.177642 containerd[1604]: 2025-09-16 04:50:14.156 [INFO][4146] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.177642 containerd[1604]: 2025-09-16 04:50:14.157 [INFO][4146] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--s4499-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"805af22b-e7a5-4562-8fa7-b331a6a81d08", ResourceVersion:"871", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1", Pod:"coredns-668d6bf9bc-s4499", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali030ea07f401", MAC:"3e:58:d1:37:61:72", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:14.177642 containerd[1604]: 2025-09-16 04:50:14.170 [INFO][4146] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" Namespace="kube-system" Pod="coredns-668d6bf9bc-s4499" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s4499-eth0" Sep 16 04:50:14.182593 containerd[1604]: time="2025-09-16T04:50:14.182544269Z" level=info msg="connecting to shim 6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5" address="unix:///run/containerd/s/55aad366e45af6034a91d3cb81bd3b428f95d7d4111c81289bdd1ffd398cd2ec" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:14.188771 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:14.217669 containerd[1604]: time="2025-09-16T04:50:14.217608684Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9872f,Uid:7fb6a3cc-bcfc-4ebe-8396-a7a64f08a3e4,Namespace:calico-system,Attempt:0,} returns sandbox id \"28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4\"" Sep 16 04:50:14.223939 containerd[1604]: time="2025-09-16T04:50:14.223886206Z" level=info msg="connecting to shim f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1" address="unix:///run/containerd/s/43189539bd0235e5193baf3df2831a8133d774fcba26f70688e759f21adefd62" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:14.235744 systemd[1]: Started cri-containerd-6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5.scope - libcontainer container 6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5. Sep 16 04:50:14.262467 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:14.264806 systemd[1]: Started cri-containerd-f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1.scope - libcontainer container f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1. Sep 16 04:50:14.305425 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:14.314643 systemd-networkd[1495]: cali3b02bd177a8: Link UP Sep 16 04:50:14.316341 systemd-networkd[1495]: cali3b02bd177a8: Gained carrier Sep 16 04:50:14.375925 containerd[1604]: time="2025-09-16T04:50:14.375850057Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-2wvgh,Uid:1b3a3c6b-e0ed-437a-9826-214673136666,Namespace:kube-system,Attempt:0,} returns sandbox id \"6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5\"" Sep 16 04:50:14.378035 kubelet[2756]: E0916 04:50:14.377111 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:14.382244 containerd[1604]: time="2025-09-16T04:50:14.381947822Z" level=info msg="CreateContainer within sandbox \"6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:13.711 [INFO][4124] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:13.734 [INFO][4124] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0 calico-apiserver-6b89c956bf- calico-apiserver d7bf725a-401b-4090-b1b5-801486b6473f 870 0 2025-09-16 04:49:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6b89c956bf projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-6b89c956bf-92rwh eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali3b02bd177a8 [] [] }} ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:13.734 [INFO][4124] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:13.815 [INFO][4190] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" HandleID="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Workload="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:13.816 [INFO][4190] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" HandleID="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Workload="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000c01c0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-6b89c956bf-92rwh", "timestamp":"2025-09-16 04:50:13.815872615 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:13.816 [INFO][4190] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.136 [INFO][4190] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.137 [INFO][4190] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.165 [INFO][4190] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.249 [INFO][4190] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.259 [INFO][4190] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.262 [INFO][4190] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.268 [INFO][4190] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.269 [INFO][4190] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.273 [INFO][4190] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463 Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.282 [INFO][4190] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.298 [INFO][4190] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.298 [INFO][4190] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" host="localhost" Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.298 [INFO][4190] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:14.389017 containerd[1604]: 2025-09-16 04:50:14.298 [INFO][4190] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" HandleID="k8s-pod-network.1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Workload="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.390535 containerd[1604]: 2025-09-16 04:50:14.306 [INFO][4124] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0", GenerateName:"calico-apiserver-6b89c956bf-", Namespace:"calico-apiserver", SelfLink:"", UID:"d7bf725a-401b-4090-b1b5-801486b6473f", ResourceVersion:"870", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b89c956bf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-6b89c956bf-92rwh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3b02bd177a8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:14.390535 containerd[1604]: 2025-09-16 04:50:14.306 [INFO][4124] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.390535 containerd[1604]: 2025-09-16 04:50:14.307 [INFO][4124] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3b02bd177a8 ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.390535 containerd[1604]: 2025-09-16 04:50:14.320 [INFO][4124] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.390535 containerd[1604]: 2025-09-16 04:50:14.320 [INFO][4124] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0", GenerateName:"calico-apiserver-6b89c956bf-", Namespace:"calico-apiserver", SelfLink:"", UID:"d7bf725a-401b-4090-b1b5-801486b6473f", ResourceVersion:"870", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b89c956bf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463", Pod:"calico-apiserver-6b89c956bf-92rwh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3b02bd177a8", MAC:"9a:21:96:16:aa:12", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:14.390535 containerd[1604]: 2025-09-16 04:50:14.381 [INFO][4124] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-92rwh" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--92rwh-eth0" Sep 16 04:50:14.390535 containerd[1604]: time="2025-09-16T04:50:14.389309550Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s4499,Uid:805af22b-e7a5-4562-8fa7-b331a6a81d08,Namespace:kube-system,Attempt:0,} returns sandbox id \"f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1\"" Sep 16 04:50:14.391455 kubelet[2756]: E0916 04:50:14.391410 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:14.397257 containerd[1604]: time="2025-09-16T04:50:14.397213946Z" level=info msg="CreateContainer within sandbox \"f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 16 04:50:14.410017 containerd[1604]: time="2025-09-16T04:50:14.409942818Z" level=info msg="Container 7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:14.422986 containerd[1604]: time="2025-09-16T04:50:14.422918283Z" level=info msg="Container 38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:14.429628 containerd[1604]: time="2025-09-16T04:50:14.429572873Z" level=info msg="CreateContainer within sandbox \"6864417b813c1c52b9fa49029d1dba06097c3a398722c3c2c56785b3fa0851e5\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18\"" Sep 16 04:50:14.430833 containerd[1604]: time="2025-09-16T04:50:14.430497108Z" level=info msg="StartContainer for \"7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18\"" Sep 16 04:50:14.434011 containerd[1604]: time="2025-09-16T04:50:14.433962522Z" level=info msg="connecting to shim 7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18" address="unix:///run/containerd/s/55aad366e45af6034a91d3cb81bd3b428f95d7d4111c81289bdd1ffd398cd2ec" protocol=ttrpc version=3 Sep 16 04:50:14.436685 containerd[1604]: time="2025-09-16T04:50:14.436640719Z" level=info msg="CreateContainer within sandbox \"f9047ce25cb04b9dd833543570a015c81ace133efe240f016360077606b557a1\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882\"" Sep 16 04:50:14.438685 containerd[1604]: time="2025-09-16T04:50:14.438646083Z" level=info msg="StartContainer for \"38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882\"" Sep 16 04:50:14.440103 containerd[1604]: time="2025-09-16T04:50:14.440065687Z" level=info msg="connecting to shim 38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882" address="unix:///run/containerd/s/43189539bd0235e5193baf3df2831a8133d774fcba26f70688e759f21adefd62" protocol=ttrpc version=3 Sep 16 04:50:14.442314 containerd[1604]: time="2025-09-16T04:50:14.442243606Z" level=info msg="connecting to shim 1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463" address="unix:///run/containerd/s/6d3529eeb255681f3dc6302a300ce917b6fb139d16ad61a5af3ff3065463a0b5" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:14.468736 systemd[1]: Started cri-containerd-38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882.scope - libcontainer container 38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882. Sep 16 04:50:14.470721 systemd[1]: Started cri-containerd-7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18.scope - libcontainer container 7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18. Sep 16 04:50:14.475879 systemd[1]: Started cri-containerd-1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463.scope - libcontainer container 1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463. Sep 16 04:50:14.498476 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:14.535754 containerd[1604]: time="2025-09-16T04:50:14.535430374Z" level=info msg="StartContainer for \"38b9cb206bec419f32675dad72a3079ce7ceb1f57cb6b7c61ff8754b3d713882\" returns successfully" Sep 16 04:50:14.535754 containerd[1604]: time="2025-09-16T04:50:14.535692356Z" level=info msg="StartContainer for \"7c7822ae86922c4c09c4af868a05d58dac4d5cfffe45baf404f4b3664255ae18\" returns successfully" Sep 16 04:50:14.594545 containerd[1604]: time="2025-09-16T04:50:14.594468967Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-92rwh,Uid:d7bf725a-401b-4090-b1b5-801486b6473f,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463\"" Sep 16 04:50:14.643628 containerd[1604]: time="2025-09-16T04:50:14.643420948Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-mdn5g,Uid:7da891ca-d373-47fb-8d2b-a8bb2752d2c0,Namespace:calico-system,Attempt:0,}" Sep 16 04:50:14.772417 systemd-networkd[1495]: vxlan.calico: Link UP Sep 16 04:50:14.772425 systemd-networkd[1495]: vxlan.calico: Gained carrier Sep 16 04:50:14.808339 kubelet[2756]: E0916 04:50:14.808270 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:14.816069 kubelet[2756]: E0916 04:50:14.816026 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:14.964726 systemd-networkd[1495]: cali5f6f863d32b: Gained IPv6LL Sep 16 04:50:14.980445 kubelet[2756]: I0916 04:50:14.980362 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-2wvgh" podStartSLOduration=42.980337969 podStartE2EDuration="42.980337969s" podCreationTimestamp="2025-09-16 04:49:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:50:14.978880754 +0000 UTC m=+47.451640343" watchObservedRunningTime="2025-09-16 04:50:14.980337969 +0000 UTC m=+47.453097539" Sep 16 04:50:15.006168 containerd[1604]: time="2025-09-16T04:50:15.006105744Z" level=info msg="TaskExit event in podsandbox handler container_id:\"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\" id:\"f8404c82294a66e46f4c9ca304621281ac328508ef12b402192956c5f88b3576\" pid:4726 exit_status:1 exited_at:{seconds:1757998215 nanos:5491972}" Sep 16 04:50:15.031533 systemd-networkd[1495]: cali1101966f210: Gained IPv6LL Sep 16 04:50:15.042911 systemd-networkd[1495]: cali3be8d7d930e: Link UP Sep 16 04:50:15.043725 systemd-networkd[1495]: cali3be8d7d930e: Gained carrier Sep 16 04:50:15.076428 kubelet[2756]: I0916 04:50:15.076125 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-s4499" podStartSLOduration=43.076100032 podStartE2EDuration="43.076100032s" podCreationTimestamp="2025-09-16 04:49:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:50:15.074650101 +0000 UTC m=+47.547409680" watchObservedRunningTime="2025-09-16 04:50:15.076100032 +0000 UTC m=+47.548859591" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.706 [INFO][4647] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--54d579b49d--mdn5g-eth0 goldmane-54d579b49d- calico-system 7da891ca-d373-47fb-8d2b-a8bb2752d2c0 872 0 2025-09-16 04:49:46 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-54d579b49d-mdn5g eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali3be8d7d930e [] [] }} ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.707 [INFO][4647] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.763 [INFO][4676] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" HandleID="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Workload="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.764 [INFO][4676] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" HandleID="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Workload="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0005153f0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-54d579b49d-mdn5g", "timestamp":"2025-09-16 04:50:14.763802009 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.764 [INFO][4676] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.765 [INFO][4676] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.765 [INFO][4676] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.778 [INFO][4676] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.786 [INFO][4676] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.793 [INFO][4676] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.796 [INFO][4676] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.801 [INFO][4676] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.802 [INFO][4676] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.811 [INFO][4676] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877 Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:14.976 [INFO][4676] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:15.031 [INFO][4676] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:15.032 [INFO][4676] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" host="localhost" Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:15.032 [INFO][4676] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:15.148558 containerd[1604]: 2025-09-16 04:50:15.032 [INFO][4676] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" HandleID="k8s-pod-network.de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Workload="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.149955 containerd[1604]: 2025-09-16 04:50:15.039 [INFO][4647] cni-plugin/k8s.go 418: Populated endpoint ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--mdn5g-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"7da891ca-d373-47fb-8d2b-a8bb2752d2c0", ResourceVersion:"872", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-54d579b49d-mdn5g", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3be8d7d930e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:15.149955 containerd[1604]: 2025-09-16 04:50:15.039 [INFO][4647] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.149955 containerd[1604]: 2025-09-16 04:50:15.039 [INFO][4647] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3be8d7d930e ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.149955 containerd[1604]: 2025-09-16 04:50:15.044 [INFO][4647] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.149955 containerd[1604]: 2025-09-16 04:50:15.044 [INFO][4647] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--54d579b49d--mdn5g-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"7da891ca-d373-47fb-8d2b-a8bb2752d2c0", ResourceVersion:"872", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877", Pod:"goldmane-54d579b49d-mdn5g", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3be8d7d930e", MAC:"86:e5:d9:98:72:b5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:15.149955 containerd[1604]: 2025-09-16 04:50:15.141 [INFO][4647] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" Namespace="calico-system" Pod="goldmane-54d579b49d-mdn5g" WorkloadEndpoint="localhost-k8s-goldmane--54d579b49d--mdn5g-eth0" Sep 16 04:50:15.198565 containerd[1604]: time="2025-09-16T04:50:15.198450264Z" level=info msg="connecting to shim de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877" address="unix:///run/containerd/s/a110fe41a7a7d24cfbce6d3e9ad1e16744976a32bf9ce38b580c30bfbb969fb7" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:15.252685 systemd[1]: Started cri-containerd-de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877.scope - libcontainer container de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877. Sep 16 04:50:15.274090 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:15.328876 containerd[1604]: time="2025-09-16T04:50:15.328730960Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-mdn5g,Uid:7da891ca-d373-47fb-8d2b-a8bb2752d2c0,Namespace:calico-system,Attempt:0,} returns sandbox id \"de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877\"" Sep 16 04:50:15.605691 systemd-networkd[1495]: cali2abca3fdc34: Gained IPv6LL Sep 16 04:50:15.796936 systemd-networkd[1495]: cali3b02bd177a8: Gained IPv6LL Sep 16 04:50:15.822792 kubelet[2756]: E0916 04:50:15.822748 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:15.823633 kubelet[2756]: E0916 04:50:15.822971 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:16.010385 containerd[1604]: time="2025-09-16T04:50:16.010317509Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:16.011417 containerd[1604]: time="2025-09-16T04:50:16.011362110Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Sep 16 04:50:16.012688 containerd[1604]: time="2025-09-16T04:50:16.012659474Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:16.014966 containerd[1604]: time="2025-09-16T04:50:16.014895962Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:16.015483 containerd[1604]: time="2025-09-16T04:50:16.015439672Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 2.395793213s" Sep 16 04:50:16.015483 containerd[1604]: time="2025-09-16T04:50:16.015469278Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Sep 16 04:50:16.016581 containerd[1604]: time="2025-09-16T04:50:16.016417227Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 16 04:50:16.017953 containerd[1604]: time="2025-09-16T04:50:16.017896142Z" level=info msg="CreateContainer within sandbox \"1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 16 04:50:16.029152 containerd[1604]: time="2025-09-16T04:50:16.029046970Z" level=info msg="Container faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:16.038434 containerd[1604]: time="2025-09-16T04:50:16.038379865Z" level=info msg="CreateContainer within sandbox \"1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71\"" Sep 16 04:50:16.040537 containerd[1604]: time="2025-09-16T04:50:16.038976074Z" level=info msg="StartContainer for \"faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71\"" Sep 16 04:50:16.040537 containerd[1604]: time="2025-09-16T04:50:16.040249584Z" level=info msg="connecting to shim faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71" address="unix:///run/containerd/s/114ea8b83d913ec925f55f13dc091246f4c440cfac994aeccaddf406ade42d12" protocol=ttrpc version=3 Sep 16 04:50:16.064744 systemd[1]: Started cri-containerd-faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71.scope - libcontainer container faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71. Sep 16 04:50:16.118754 containerd[1604]: time="2025-09-16T04:50:16.118693031Z" level=info msg="StartContainer for \"faa5d602a6b94d51c2180305615c7c485e0af0dc524bdbc821a2cb4beb706b71\" returns successfully" Sep 16 04:50:16.180840 systemd-networkd[1495]: cali030ea07f401: Gained IPv6LL Sep 16 04:50:16.564825 systemd-networkd[1495]: cali3be8d7d930e: Gained IPv6LL Sep 16 04:50:16.692727 systemd-networkd[1495]: vxlan.calico: Gained IPv6LL Sep 16 04:50:16.827402 kubelet[2756]: E0916 04:50:16.827264 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:16.827402 kubelet[2756]: E0916 04:50:16.827360 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:17.017441 systemd[1]: Started sshd@8-10.0.0.34:22-10.0.0.1:43816.service - OpenSSH per-connection server daemon (10.0.0.1:43816). Sep 16 04:50:17.090674 sshd[4877]: Accepted publickey for core from 10.0.0.1 port 43816 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:17.092703 sshd-session[4877]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:17.098322 systemd-logind[1587]: New session 9 of user core. Sep 16 04:50:17.106688 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 16 04:50:17.252198 sshd[4881]: Connection closed by 10.0.0.1 port 43816 Sep 16 04:50:17.252615 sshd-session[4877]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:17.258086 systemd[1]: sshd@8-10.0.0.34:22-10.0.0.1:43816.service: Deactivated successfully. Sep 16 04:50:17.260434 systemd[1]: session-9.scope: Deactivated successfully. Sep 16 04:50:17.261450 systemd-logind[1587]: Session 9 logged out. Waiting for processes to exit. Sep 16 04:50:17.262753 systemd-logind[1587]: Removed session 9. Sep 16 04:50:17.906962 containerd[1604]: time="2025-09-16T04:50:17.906766407Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:17.907815 containerd[1604]: time="2025-09-16T04:50:17.907744583Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Sep 16 04:50:17.908900 containerd[1604]: time="2025-09-16T04:50:17.908849366Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:17.911116 containerd[1604]: time="2025-09-16T04:50:17.911067789Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:17.911490 containerd[1604]: time="2025-09-16T04:50:17.911442923Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 1.894995218s" Sep 16 04:50:17.911490 containerd[1604]: time="2025-09-16T04:50:17.911474352Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Sep 16 04:50:17.912725 containerd[1604]: time="2025-09-16T04:50:17.912671308Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 16 04:50:17.913896 containerd[1604]: time="2025-09-16T04:50:17.913859438Z" level=info msg="CreateContainer within sandbox \"28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 16 04:50:17.938432 containerd[1604]: time="2025-09-16T04:50:17.938340709Z" level=info msg="Container 53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:17.960155 containerd[1604]: time="2025-09-16T04:50:17.960088202Z" level=info msg="CreateContainer within sandbox \"28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96\"" Sep 16 04:50:17.962782 containerd[1604]: time="2025-09-16T04:50:17.962741751Z" level=info msg="StartContainer for \"53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96\"" Sep 16 04:50:17.965573 containerd[1604]: time="2025-09-16T04:50:17.964815532Z" level=info msg="connecting to shim 53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96" address="unix:///run/containerd/s/3f7e5dec723747c230d72b31987fe5ff9ba6ab1a59059e555d6f8b2de5ca1c32" protocol=ttrpc version=3 Sep 16 04:50:17.993892 systemd[1]: Started cri-containerd-53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96.scope - libcontainer container 53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96. Sep 16 04:50:18.047124 containerd[1604]: time="2025-09-16T04:50:18.047045212Z" level=info msg="StartContainer for \"53ca04b7d8444a5aac1abf504ddbdfb26326192d36ed00fcdc549ebfbe87ea96\" returns successfully" Sep 16 04:50:20.964803 containerd[1604]: time="2025-09-16T04:50:20.964729857Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:20.965730 containerd[1604]: time="2025-09-16T04:50:20.965662817Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Sep 16 04:50:20.967267 containerd[1604]: time="2025-09-16T04:50:20.967216042Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:20.970450 containerd[1604]: time="2025-09-16T04:50:20.970411918Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:20.971360 containerd[1604]: time="2025-09-16T04:50:20.971324521Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 3.058605403s" Sep 16 04:50:20.971360 containerd[1604]: time="2025-09-16T04:50:20.971357533Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 16 04:50:20.976534 containerd[1604]: time="2025-09-16T04:50:20.976456580Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 16 04:50:20.977997 containerd[1604]: time="2025-09-16T04:50:20.977962956Z" level=info msg="CreateContainer within sandbox \"1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 16 04:50:20.999531 containerd[1604]: time="2025-09-16T04:50:20.999450364Z" level=info msg="Container bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:21.009231 containerd[1604]: time="2025-09-16T04:50:21.009168437Z" level=info msg="CreateContainer within sandbox \"1a8e9164042c1baf2c55d507ad6efd166c9e56664e5251ac086ce66c009a0463\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f\"" Sep 16 04:50:21.009892 containerd[1604]: time="2025-09-16T04:50:21.009856789Z" level=info msg="StartContainer for \"bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f\"" Sep 16 04:50:21.011051 containerd[1604]: time="2025-09-16T04:50:21.011024110Z" level=info msg="connecting to shim bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f" address="unix:///run/containerd/s/6d3529eeb255681f3dc6302a300ce917b6fb139d16ad61a5af3ff3065463a0b5" protocol=ttrpc version=3 Sep 16 04:50:21.042720 systemd[1]: Started cri-containerd-bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f.scope - libcontainer container bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f. Sep 16 04:50:21.129254 containerd[1604]: time="2025-09-16T04:50:21.129196330Z" level=info msg="StartContainer for \"bffd45cbc38267a0acfe6eb4b9962439949c4cf17aa2a487db4fdfcb9a46ba0f\" returns successfully" Sep 16 04:50:21.874254 kubelet[2756]: I0916 04:50:21.874012 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6b89c956bf-92rwh" podStartSLOduration=31.493955172 podStartE2EDuration="37.873975632s" podCreationTimestamp="2025-09-16 04:49:44 +0000 UTC" firstStartedPulling="2025-09-16 04:50:14.596086704 +0000 UTC m=+47.068846263" lastFinishedPulling="2025-09-16 04:50:20.976107154 +0000 UTC m=+53.448866723" observedRunningTime="2025-09-16 04:50:21.873162126 +0000 UTC m=+54.345921715" watchObservedRunningTime="2025-09-16 04:50:21.873975632 +0000 UTC m=+54.346735201" Sep 16 04:50:22.267659 systemd[1]: Started sshd@9-10.0.0.34:22-10.0.0.1:38818.service - OpenSSH per-connection server daemon (10.0.0.1:38818). Sep 16 04:50:22.341593 sshd[4989]: Accepted publickey for core from 10.0.0.1 port 38818 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:22.344568 sshd-session[4989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:22.352799 systemd-logind[1587]: New session 10 of user core. Sep 16 04:50:22.362047 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 16 04:50:22.535933 sshd[4992]: Connection closed by 10.0.0.1 port 38818 Sep 16 04:50:22.536263 sshd-session[4989]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:22.540267 systemd[1]: sshd@9-10.0.0.34:22-10.0.0.1:38818.service: Deactivated successfully. Sep 16 04:50:22.543034 systemd[1]: session-10.scope: Deactivated successfully. Sep 16 04:50:22.545634 systemd-logind[1587]: Session 10 logged out. Waiting for processes to exit. Sep 16 04:50:22.547285 systemd-logind[1587]: Removed session 10. Sep 16 04:50:22.866420 kubelet[2756]: I0916 04:50:22.866273 2756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 16 04:50:23.889970 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount392632410.mount: Deactivated successfully. Sep 16 04:50:24.715616 containerd[1604]: time="2025-09-16T04:50:24.715539109Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:24.716625 containerd[1604]: time="2025-09-16T04:50:24.716564393Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Sep 16 04:50:24.718074 containerd[1604]: time="2025-09-16T04:50:24.717990849Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:24.721742 containerd[1604]: time="2025-09-16T04:50:24.721694568Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:24.722669 containerd[1604]: time="2025-09-16T04:50:24.722636243Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 3.745855886s" Sep 16 04:50:24.722669 containerd[1604]: time="2025-09-16T04:50:24.722668354Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Sep 16 04:50:24.724309 containerd[1604]: time="2025-09-16T04:50:24.724080042Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 16 04:50:24.726967 containerd[1604]: time="2025-09-16T04:50:24.726912917Z" level=info msg="CreateContainer within sandbox \"de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 16 04:50:24.737032 containerd[1604]: time="2025-09-16T04:50:24.736971046Z" level=info msg="Container b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:24.746043 containerd[1604]: time="2025-09-16T04:50:24.745990868Z" level=info msg="CreateContainer within sandbox \"de6dac0e5b01718514f975e0ea325ba2637fcf18b7605a07ba4b03b0901c0877\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\"" Sep 16 04:50:24.747204 containerd[1604]: time="2025-09-16T04:50:24.747151345Z" level=info msg="StartContainer for \"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\"" Sep 16 04:50:24.748459 containerd[1604]: time="2025-09-16T04:50:24.748431647Z" level=info msg="connecting to shim b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d" address="unix:///run/containerd/s/a110fe41a7a7d24cfbce6d3e9ad1e16744976a32bf9ce38b580c30bfbb969fb7" protocol=ttrpc version=3 Sep 16 04:50:24.822867 systemd[1]: Started cri-containerd-b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d.scope - libcontainer container b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d. Sep 16 04:50:24.880407 containerd[1604]: time="2025-09-16T04:50:24.880301238Z" level=info msg="StartContainer for \"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\" returns successfully" Sep 16 04:50:25.643970 containerd[1604]: time="2025-09-16T04:50:25.643919818Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-b8d5c47bf-nrnjv,Uid:735decdf-fb16-4265-96d7-4da1d18f1679,Namespace:calico-system,Attempt:0,}" Sep 16 04:50:25.760448 systemd-networkd[1495]: cali1e112429c13: Link UP Sep 16 04:50:25.760728 systemd-networkd[1495]: cali1e112429c13: Gained carrier Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.680 [INFO][5058] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0 calico-kube-controllers-b8d5c47bf- calico-system 735decdf-fb16-4265-96d7-4da1d18f1679 864 0 2025-09-16 04:49:48 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:b8d5c47bf projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-b8d5c47bf-nrnjv eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali1e112429c13 [] [] }} ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.680 [INFO][5058] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.711 [INFO][5073] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" HandleID="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Workload="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.711 [INFO][5073] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" HandleID="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Workload="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c72d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-b8d5c47bf-nrnjv", "timestamp":"2025-09-16 04:50:25.710981263 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.711 [INFO][5073] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.711 [INFO][5073] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.711 [INFO][5073] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.719 [INFO][5073] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.725 [INFO][5073] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.732 [INFO][5073] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.734 [INFO][5073] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.737 [INFO][5073] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.737 [INFO][5073] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.738 [INFO][5073] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.743 [INFO][5073] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.752 [INFO][5073] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.752 [INFO][5073] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" host="localhost" Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.752 [INFO][5073] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:25.776557 containerd[1604]: 2025-09-16 04:50:25.752 [INFO][5073] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" HandleID="k8s-pod-network.442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Workload="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.778040 containerd[1604]: 2025-09-16 04:50:25.756 [INFO][5058] cni-plugin/k8s.go 418: Populated endpoint ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0", GenerateName:"calico-kube-controllers-b8d5c47bf-", Namespace:"calico-system", SelfLink:"", UID:"735decdf-fb16-4265-96d7-4da1d18f1679", ResourceVersion:"864", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"b8d5c47bf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-b8d5c47bf-nrnjv", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali1e112429c13", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:25.778040 containerd[1604]: 2025-09-16 04:50:25.756 [INFO][5058] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.778040 containerd[1604]: 2025-09-16 04:50:25.756 [INFO][5058] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1e112429c13 ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.778040 containerd[1604]: 2025-09-16 04:50:25.759 [INFO][5058] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.778040 containerd[1604]: 2025-09-16 04:50:25.760 [INFO][5058] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0", GenerateName:"calico-kube-controllers-b8d5c47bf-", Namespace:"calico-system", SelfLink:"", UID:"735decdf-fb16-4265-96d7-4da1d18f1679", ResourceVersion:"864", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"b8d5c47bf", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe", Pod:"calico-kube-controllers-b8d5c47bf-nrnjv", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali1e112429c13", MAC:"4a:26:0d:d1:56:4d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:25.778040 containerd[1604]: 2025-09-16 04:50:25.771 [INFO][5058] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" Namespace="calico-system" Pod="calico-kube-controllers-b8d5c47bf-nrnjv" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--b8d5c47bf--nrnjv-eth0" Sep 16 04:50:25.803454 containerd[1604]: time="2025-09-16T04:50:25.803357387Z" level=info msg="connecting to shim 442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe" address="unix:///run/containerd/s/0751ff374c8174d092f1e44b5e826ac7e0d587c379cf7f5019c674809a28c014" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:25.833735 systemd[1]: Started cri-containerd-442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe.scope - libcontainer container 442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe. Sep 16 04:50:25.851300 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:25.887799 containerd[1604]: time="2025-09-16T04:50:25.887742942Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-b8d5c47bf-nrnjv,Uid:735decdf-fb16-4265-96d7-4da1d18f1679,Namespace:calico-system,Attempt:0,} returns sandbox id \"442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe\"" Sep 16 04:50:25.905640 kubelet[2756]: I0916 04:50:25.905440 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-mdn5g" podStartSLOduration=30.512491434 podStartE2EDuration="39.905414323s" podCreationTimestamp="2025-09-16 04:49:46 +0000 UTC" firstStartedPulling="2025-09-16 04:50:15.330968981 +0000 UTC m=+47.803728550" lastFinishedPulling="2025-09-16 04:50:24.72389187 +0000 UTC m=+57.196651439" observedRunningTime="2025-09-16 04:50:25.904372769 +0000 UTC m=+58.377132348" watchObservedRunningTime="2025-09-16 04:50:25.905414323 +0000 UTC m=+58.378173892" Sep 16 04:50:26.009100 containerd[1604]: time="2025-09-16T04:50:26.009019803Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\" id:\"7421e5e818bed6cd4e7b9ccf23eeececfa63e828328f56bbaaa3ce0297cfe881\" pid:5153 exit_status:1 exited_at:{seconds:1757998226 nanos:1535322}" Sep 16 04:50:26.643587 containerd[1604]: time="2025-09-16T04:50:26.643493395Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-6bfxs,Uid:3e572577-68f4-49a7-9725-3f0cb8aef839,Namespace:calico-apiserver,Attempt:0,}" Sep 16 04:50:26.984744 systemd-networkd[1495]: calieb8d37089e7: Link UP Sep 16 04:50:26.985718 systemd-networkd[1495]: calieb8d37089e7: Gained carrier Sep 16 04:50:27.029287 containerd[1604]: time="2025-09-16T04:50:27.029224899Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\" id:\"076e9d1f674f17d5855e88e698374949e7a3b743912b5593bc2f3460d9831b27\" pid:5201 exit_status:1 exited_at:{seconds:1757998227 nanos:28761279}" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.870 [INFO][5168] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0 calico-apiserver-6b89c956bf- calico-apiserver 3e572577-68f4-49a7-9725-3f0cb8aef839 869 0 2025-09-16 04:49:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6b89c956bf projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-6b89c956bf-6bfxs eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calieb8d37089e7 [] [] }} ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.870 [INFO][5168] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.909 [INFO][5182] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" HandleID="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Workload="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.909 [INFO][5182] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" HandleID="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Workload="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00051e660), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-6b89c956bf-6bfxs", "timestamp":"2025-09-16 04:50:26.909357571 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.909 [INFO][5182] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.910 [INFO][5182] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.910 [INFO][5182] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.920 [INFO][5182] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.930 [INFO][5182] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.939 [INFO][5182] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.943 [INFO][5182] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.947 [INFO][5182] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.947 [INFO][5182] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.949 [INFO][5182] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.957 [INFO][5182] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.967 [INFO][5182] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.968 [INFO][5182] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" host="localhost" Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.968 [INFO][5182] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 16 04:50:27.075688 containerd[1604]: 2025-09-16 04:50:26.968 [INFO][5182] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" HandleID="k8s-pod-network.03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Workload="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.076435 containerd[1604]: 2025-09-16 04:50:26.978 [INFO][5168] cni-plugin/k8s.go 418: Populated endpoint ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0", GenerateName:"calico-apiserver-6b89c956bf-", Namespace:"calico-apiserver", SelfLink:"", UID:"3e572577-68f4-49a7-9725-3f0cb8aef839", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b89c956bf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-6b89c956bf-6bfxs", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieb8d37089e7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:27.076435 containerd[1604]: 2025-09-16 04:50:26.979 [INFO][5168] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.076435 containerd[1604]: 2025-09-16 04:50:26.979 [INFO][5168] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieb8d37089e7 ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.076435 containerd[1604]: 2025-09-16 04:50:26.986 [INFO][5168] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.076435 containerd[1604]: 2025-09-16 04:50:26.987 [INFO][5168] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0", GenerateName:"calico-apiserver-6b89c956bf-", Namespace:"calico-apiserver", SelfLink:"", UID:"3e572577-68f4-49a7-9725-3f0cb8aef839", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.September, 16, 4, 49, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b89c956bf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee", Pod:"calico-apiserver-6b89c956bf-6bfxs", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieb8d37089e7", MAC:"82:b8:c7:72:85:23", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 16 04:50:27.076435 containerd[1604]: 2025-09-16 04:50:27.063 [INFO][5168] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" Namespace="calico-apiserver" Pod="calico-apiserver-6b89c956bf-6bfxs" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b89c956bf--6bfxs-eth0" Sep 16 04:50:27.116278 containerd[1604]: time="2025-09-16T04:50:27.116140821Z" level=info msg="connecting to shim 03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee" address="unix:///run/containerd/s/bb45090d945d92289fb27f68b36b055166b85ae349ac88211777fa76084fb748" namespace=k8s.io protocol=ttrpc version=3 Sep 16 04:50:27.158213 systemd[1]: Started cri-containerd-03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee.scope - libcontainer container 03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee. Sep 16 04:50:27.178263 systemd-resolved[1411]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 16 04:50:27.301497 containerd[1604]: time="2025-09-16T04:50:27.300137420Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b89c956bf-6bfxs,Uid:3e572577-68f4-49a7-9725-3f0cb8aef839,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee\"" Sep 16 04:50:27.307906 containerd[1604]: time="2025-09-16T04:50:27.307790266Z" level=info msg="CreateContainer within sandbox \"03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 16 04:50:27.324498 containerd[1604]: time="2025-09-16T04:50:27.324449396Z" level=info msg="Container dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:27.339591 containerd[1604]: time="2025-09-16T04:50:27.339498516Z" level=info msg="CreateContainer within sandbox \"03c45956df30946685a5b96b58dc560c33ac8a8104a4b2f2fc96c9d1d042e0ee\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c\"" Sep 16 04:50:27.342063 containerd[1604]: time="2025-09-16T04:50:27.342026047Z" level=info msg="StartContainer for \"dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c\"" Sep 16 04:50:27.344018 containerd[1604]: time="2025-09-16T04:50:27.343763836Z" level=info msg="connecting to shim dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c" address="unix:///run/containerd/s/bb45090d945d92289fb27f68b36b055166b85ae349ac88211777fa76084fb748" protocol=ttrpc version=3 Sep 16 04:50:27.380040 systemd[1]: Started cri-containerd-dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c.scope - libcontainer container dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c. Sep 16 04:50:27.445294 systemd-networkd[1495]: cali1e112429c13: Gained IPv6LL Sep 16 04:50:27.460129 containerd[1604]: time="2025-09-16T04:50:27.459211680Z" level=info msg="StartContainer for \"dc60e349835f0104df05d7f0ebd0566647c45b736e5487dd5fb2ba922e58398c\" returns successfully" Sep 16 04:50:27.556784 systemd[1]: Started sshd@10-10.0.0.34:22-10.0.0.1:38834.service - OpenSSH per-connection server daemon (10.0.0.1:38834). Sep 16 04:50:27.583176 containerd[1604]: time="2025-09-16T04:50:27.583104794Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:27.583962 containerd[1604]: time="2025-09-16T04:50:27.583915876Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Sep 16 04:50:27.585527 containerd[1604]: time="2025-09-16T04:50:27.585329497Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:27.588928 containerd[1604]: time="2025-09-16T04:50:27.588885438Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:27.590140 containerd[1604]: time="2025-09-16T04:50:27.590105907Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 2.865992793s" Sep 16 04:50:27.590198 containerd[1604]: time="2025-09-16T04:50:27.590158927Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Sep 16 04:50:27.593991 containerd[1604]: time="2025-09-16T04:50:27.593900866Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 16 04:50:27.596535 containerd[1604]: time="2025-09-16T04:50:27.596466067Z" level=info msg="CreateContainer within sandbox \"1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 16 04:50:27.609587 containerd[1604]: time="2025-09-16T04:50:27.609523011Z" level=info msg="Container 84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:27.634538 containerd[1604]: time="2025-09-16T04:50:27.632883100Z" level=info msg="CreateContainer within sandbox \"1e00b6ebd26d0915614e576a553a3d612fa5e86ca2401fad6861b1f23716e650\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462\"" Sep 16 04:50:27.635425 containerd[1604]: time="2025-09-16T04:50:27.635389281Z" level=info msg="StartContainer for \"84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462\"" Sep 16 04:50:27.637105 containerd[1604]: time="2025-09-16T04:50:27.637069925Z" level=info msg="connecting to shim 84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462" address="unix:///run/containerd/s/114ea8b83d913ec925f55f13dc091246f4c440cfac994aeccaddf406ade42d12" protocol=ttrpc version=3 Sep 16 04:50:27.640460 sshd[5307]: Accepted publickey for core from 10.0.0.1 port 38834 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:27.642405 sshd-session[5307]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:27.650014 systemd-logind[1587]: New session 11 of user core. Sep 16 04:50:27.656678 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 16 04:50:27.676944 systemd[1]: Started cri-containerd-84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462.scope - libcontainer container 84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462. Sep 16 04:50:27.858275 containerd[1604]: time="2025-09-16T04:50:27.858111084Z" level=info msg="StartContainer for \"84f75d7b61283aa19eae89fe4aac96a07adfca1ca61b7fcc0834810c0f753462\" returns successfully" Sep 16 04:50:27.886564 sshd[5326]: Connection closed by 10.0.0.1 port 38834 Sep 16 04:50:27.884079 sshd-session[5307]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:27.899019 systemd[1]: sshd@10-10.0.0.34:22-10.0.0.1:38834.service: Deactivated successfully. Sep 16 04:50:27.902549 systemd[1]: session-11.scope: Deactivated successfully. Sep 16 04:50:27.905891 systemd-logind[1587]: Session 11 logged out. Waiting for processes to exit. Sep 16 04:50:27.910790 systemd[1]: Started sshd@11-10.0.0.34:22-10.0.0.1:38846.service - OpenSSH per-connection server daemon (10.0.0.1:38846). Sep 16 04:50:27.912582 systemd-logind[1587]: Removed session 11. Sep 16 04:50:27.956542 kubelet[2756]: I0916 04:50:27.956073 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-79dc8ddbd6-2vn8p" podStartSLOduration=1.9829899370000001 podStartE2EDuration="15.956045484s" podCreationTimestamp="2025-09-16 04:50:12 +0000 UTC" firstStartedPulling="2025-09-16 04:50:13.619297385 +0000 UTC m=+46.092056954" lastFinishedPulling="2025-09-16 04:50:27.592352932 +0000 UTC m=+60.065112501" observedRunningTime="2025-09-16 04:50:27.931561536 +0000 UTC m=+60.404321105" watchObservedRunningTime="2025-09-16 04:50:27.956045484 +0000 UTC m=+60.428805053" Sep 16 04:50:27.957919 kubelet[2756]: I0916 04:50:27.957299 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6b89c956bf-6bfxs" podStartSLOduration=43.957289017 podStartE2EDuration="43.957289017s" podCreationTimestamp="2025-09-16 04:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-16 04:50:27.950743088 +0000 UTC m=+60.423502667" watchObservedRunningTime="2025-09-16 04:50:27.957289017 +0000 UTC m=+60.430048596" Sep 16 04:50:27.990531 sshd[5367]: Accepted publickey for core from 10.0.0.1 port 38846 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:27.995188 sshd-session[5367]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:28.002486 systemd-logind[1587]: New session 12 of user core. Sep 16 04:50:28.006761 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 16 04:50:28.061360 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount597168907.mount: Deactivated successfully. Sep 16 04:50:28.071118 containerd[1604]: time="2025-09-16T04:50:28.071064549Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\" id:\"50786fb224b2b958df4bd2af7f76953a5391fb877fc093eb2d3df9d3e11e5367\" pid:5385 exit_status:1 exited_at:{seconds:1757998228 nanos:70610608}" Sep 16 04:50:28.085000 systemd-networkd[1495]: calieb8d37089e7: Gained IPv6LL Sep 16 04:50:28.192875 sshd[5398]: Connection closed by 10.0.0.1 port 38846 Sep 16 04:50:28.195602 sshd-session[5367]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:28.208915 systemd[1]: sshd@11-10.0.0.34:22-10.0.0.1:38846.service: Deactivated successfully. Sep 16 04:50:28.217185 systemd[1]: session-12.scope: Deactivated successfully. Sep 16 04:50:28.220052 systemd-logind[1587]: Session 12 logged out. Waiting for processes to exit. Sep 16 04:50:28.225583 systemd[1]: Started sshd@12-10.0.0.34:22-10.0.0.1:38860.service - OpenSSH per-connection server daemon (10.0.0.1:38860). Sep 16 04:50:28.229299 systemd-logind[1587]: Removed session 12. Sep 16 04:50:28.292052 sshd[5410]: Accepted publickey for core from 10.0.0.1 port 38860 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:28.294076 sshd-session[5410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:28.299740 systemd-logind[1587]: New session 13 of user core. Sep 16 04:50:28.305683 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 16 04:50:28.438294 sshd[5413]: Connection closed by 10.0.0.1 port 38860 Sep 16 04:50:28.438712 sshd-session[5410]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:28.446428 systemd[1]: sshd@12-10.0.0.34:22-10.0.0.1:38860.service: Deactivated successfully. Sep 16 04:50:28.449983 systemd[1]: session-13.scope: Deactivated successfully. Sep 16 04:50:28.454488 systemd-logind[1587]: Session 13 logged out. Waiting for processes to exit. Sep 16 04:50:28.455848 systemd-logind[1587]: Removed session 13. Sep 16 04:50:29.578650 kubelet[2756]: I0916 04:50:29.578555 2756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 16 04:50:30.491215 containerd[1604]: time="2025-09-16T04:50:30.491139985Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:30.492146 containerd[1604]: time="2025-09-16T04:50:30.492118045Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Sep 16 04:50:30.493371 containerd[1604]: time="2025-09-16T04:50:30.493343351Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:30.497383 containerd[1604]: time="2025-09-16T04:50:30.497355682Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:30.497989 containerd[1604]: time="2025-09-16T04:50:30.497960211Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 2.903832919s" Sep 16 04:50:30.498045 containerd[1604]: time="2025-09-16T04:50:30.497991470Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Sep 16 04:50:30.499668 containerd[1604]: time="2025-09-16T04:50:30.499609235Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 16 04:50:30.500261 containerd[1604]: time="2025-09-16T04:50:30.500228742Z" level=info msg="CreateContainer within sandbox \"28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 16 04:50:30.514637 containerd[1604]: time="2025-09-16T04:50:30.514071272Z" level=info msg="Container 3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:30.542659 containerd[1604]: time="2025-09-16T04:50:30.542590684Z" level=info msg="CreateContainer within sandbox \"28a02a2a5d5dc1f2b3fdb4325baae7e2f1984c4701d2b9c27a583c6de6bcf1d4\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260\"" Sep 16 04:50:30.543577 containerd[1604]: time="2025-09-16T04:50:30.543473068Z" level=info msg="StartContainer for \"3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260\"" Sep 16 04:50:30.546254 containerd[1604]: time="2025-09-16T04:50:30.546186059Z" level=info msg="connecting to shim 3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260" address="unix:///run/containerd/s/3f7e5dec723747c230d72b31987fe5ff9ba6ab1a59059e555d6f8b2de5ca1c32" protocol=ttrpc version=3 Sep 16 04:50:30.573880 systemd[1]: Started cri-containerd-3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260.scope - libcontainer container 3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260. Sep 16 04:50:30.722437 kubelet[2756]: I0916 04:50:30.722374 2756 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 16 04:50:30.722437 kubelet[2756]: I0916 04:50:30.722427 2756 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 16 04:50:30.732343 containerd[1604]: time="2025-09-16T04:50:30.732268407Z" level=info msg="StartContainer for \"3abe0876654809307a9791af743381cece94451bcf90af2b2090295c2b693260\" returns successfully" Sep 16 04:50:33.455471 systemd[1]: Started sshd@13-10.0.0.34:22-10.0.0.1:55870.service - OpenSSH per-connection server daemon (10.0.0.1:55870). Sep 16 04:50:33.543601 sshd[5473]: Accepted publickey for core from 10.0.0.1 port 55870 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:33.546030 sshd-session[5473]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:33.551032 systemd-logind[1587]: New session 14 of user core. Sep 16 04:50:33.558651 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 16 04:50:33.724755 sshd[5476]: Connection closed by 10.0.0.1 port 55870 Sep 16 04:50:33.725224 sshd-session[5473]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:33.730424 systemd[1]: sshd@13-10.0.0.34:22-10.0.0.1:55870.service: Deactivated successfully. Sep 16 04:50:33.733770 systemd[1]: session-14.scope: Deactivated successfully. Sep 16 04:50:33.737931 systemd-logind[1587]: Session 14 logged out. Waiting for processes to exit. Sep 16 04:50:33.739283 systemd-logind[1587]: Removed session 14. Sep 16 04:50:34.685764 containerd[1604]: time="2025-09-16T04:50:34.685679071Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:34.686601 containerd[1604]: time="2025-09-16T04:50:34.686559998Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Sep 16 04:50:34.688014 containerd[1604]: time="2025-09-16T04:50:34.687926830Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:34.690650 containerd[1604]: time="2025-09-16T04:50:34.690595289Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 16 04:50:34.691369 containerd[1604]: time="2025-09-16T04:50:34.691329562Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 4.191677715s" Sep 16 04:50:34.691434 containerd[1604]: time="2025-09-16T04:50:34.691372164Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Sep 16 04:50:34.705555 containerd[1604]: time="2025-09-16T04:50:34.705428863Z" level=info msg="CreateContainer within sandbox \"442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 16 04:50:34.731801 containerd[1604]: time="2025-09-16T04:50:34.731706677Z" level=info msg="Container d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba: CDI devices from CRI Config.CDIDevices: []" Sep 16 04:50:34.746316 containerd[1604]: time="2025-09-16T04:50:34.746235495Z" level=info msg="CreateContainer within sandbox \"442603b3accd9dc827cd361ce2ff4d778ced287091e4b5e25dd9462a95a71fbe\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba\"" Sep 16 04:50:34.747559 containerd[1604]: time="2025-09-16T04:50:34.746969178Z" level=info msg="StartContainer for \"d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba\"" Sep 16 04:50:34.748573 containerd[1604]: time="2025-09-16T04:50:34.748530705Z" level=info msg="connecting to shim d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba" address="unix:///run/containerd/s/0751ff374c8174d092f1e44b5e826ac7e0d587c379cf7f5019c674809a28c014" protocol=ttrpc version=3 Sep 16 04:50:34.785838 systemd[1]: Started cri-containerd-d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba.scope - libcontainer container d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba. Sep 16 04:50:34.848729 containerd[1604]: time="2025-09-16T04:50:34.848657751Z" level=info msg="StartContainer for \"d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba\" returns successfully" Sep 16 04:50:34.953884 kubelet[2756]: I0916 04:50:34.953403 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-9872f" podStartSLOduration=30.673860282 podStartE2EDuration="46.953264041s" podCreationTimestamp="2025-09-16 04:49:48 +0000 UTC" firstStartedPulling="2025-09-16 04:50:14.219502528 +0000 UTC m=+46.692262097" lastFinishedPulling="2025-09-16 04:50:30.498906287 +0000 UTC m=+62.971665856" observedRunningTime="2025-09-16 04:50:31.028073647 +0000 UTC m=+63.500833216" watchObservedRunningTime="2025-09-16 04:50:34.953264041 +0000 UTC m=+67.426023611" Sep 16 04:50:34.989971 containerd[1604]: time="2025-09-16T04:50:34.989898949Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba\" id:\"a2946b782162f1834e8fb16ced52ddaff404a6ec5634715e81fe3d35df4b33d9\" pid:5551 exited_at:{seconds:1757998234 nanos:989657435}" Sep 16 04:50:35.002402 kubelet[2756]: I0916 04:50:35.002313 2756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-b8d5c47bf-nrnjv" podStartSLOduration=38.20116658 podStartE2EDuration="47.002288586s" podCreationTimestamp="2025-09-16 04:49:48 +0000 UTC" firstStartedPulling="2025-09-16 04:50:25.89119643 +0000 UTC m=+58.363955999" lastFinishedPulling="2025-09-16 04:50:34.692318436 +0000 UTC m=+67.165078005" observedRunningTime="2025-09-16 04:50:34.95475853 +0000 UTC m=+67.427518099" watchObservedRunningTime="2025-09-16 04:50:35.002288586 +0000 UTC m=+67.475048155" Sep 16 04:50:38.744177 systemd[1]: Started sshd@14-10.0.0.34:22-10.0.0.1:55874.service - OpenSSH per-connection server daemon (10.0.0.1:55874). Sep 16 04:50:38.821955 sshd[5572]: Accepted publickey for core from 10.0.0.1 port 55874 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:38.825251 sshd-session[5572]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:38.831969 systemd-logind[1587]: New session 15 of user core. Sep 16 04:50:38.842950 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 16 04:50:39.004477 sshd[5575]: Connection closed by 10.0.0.1 port 55874 Sep 16 04:50:39.004908 sshd-session[5572]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:39.013330 systemd[1]: sshd@14-10.0.0.34:22-10.0.0.1:55874.service: Deactivated successfully. Sep 16 04:50:39.016455 systemd[1]: session-15.scope: Deactivated successfully. Sep 16 04:50:39.017918 systemd-logind[1587]: Session 15 logged out. Waiting for processes to exit. Sep 16 04:50:39.020645 systemd-logind[1587]: Removed session 15. Sep 16 04:50:44.021972 systemd[1]: Started sshd@15-10.0.0.34:22-10.0.0.1:33902.service - OpenSSH per-connection server daemon (10.0.0.1:33902). Sep 16 04:50:44.095464 sshd[5589]: Accepted publickey for core from 10.0.0.1 port 33902 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:44.097243 sshd-session[5589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:44.102246 systemd-logind[1587]: New session 16 of user core. Sep 16 04:50:44.111654 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 16 04:50:44.230056 sshd[5592]: Connection closed by 10.0.0.1 port 33902 Sep 16 04:50:44.230443 sshd-session[5589]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:44.235082 systemd[1]: sshd@15-10.0.0.34:22-10.0.0.1:33902.service: Deactivated successfully. Sep 16 04:50:44.237108 systemd[1]: session-16.scope: Deactivated successfully. Sep 16 04:50:44.237933 systemd-logind[1587]: Session 16 logged out. Waiting for processes to exit. Sep 16 04:50:44.239180 systemd-logind[1587]: Removed session 16. Sep 16 04:50:44.642999 kubelet[2756]: E0916 04:50:44.642957 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:44.950061 containerd[1604]: time="2025-09-16T04:50:44.949909002Z" level=info msg="TaskExit event in podsandbox handler container_id:\"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\" id:\"1b17861f5a7f5b7f6d3fe45fb0f335691128b3a388e9c9a9bdaee2bcdd2cfa66\" pid:5617 exited_at:{seconds:1757998244 nanos:949294737}" Sep 16 04:50:45.643582 kubelet[2756]: E0916 04:50:45.643487 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:46.642994 kubelet[2756]: E0916 04:50:46.642917 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:50:49.249337 systemd[1]: Started sshd@16-10.0.0.34:22-10.0.0.1:33906.service - OpenSSH per-connection server daemon (10.0.0.1:33906). Sep 16 04:50:49.344736 sshd[5632]: Accepted publickey for core from 10.0.0.1 port 33906 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:49.347152 sshd-session[5632]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:49.355525 systemd-logind[1587]: New session 17 of user core. Sep 16 04:50:49.366833 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 16 04:50:49.552383 sshd[5635]: Connection closed by 10.0.0.1 port 33906 Sep 16 04:50:49.552705 sshd-session[5632]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:49.557475 systemd[1]: sshd@16-10.0.0.34:22-10.0.0.1:33906.service: Deactivated successfully. Sep 16 04:50:49.559888 systemd[1]: session-17.scope: Deactivated successfully. Sep 16 04:50:49.561014 systemd-logind[1587]: Session 17 logged out. Waiting for processes to exit. Sep 16 04:50:49.562485 systemd-logind[1587]: Removed session 17. Sep 16 04:50:53.465583 containerd[1604]: time="2025-09-16T04:50:53.465520323Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\" id:\"7f2de5d09ad8a718e8b68826073bf47fbb0a9bbd6e8d3576c1f05693d95d1f22\" pid:5659 exited_at:{seconds:1757998253 nanos:465124198}" Sep 16 04:50:54.571097 systemd[1]: Started sshd@17-10.0.0.34:22-10.0.0.1:35088.service - OpenSSH per-connection server daemon (10.0.0.1:35088). Sep 16 04:50:54.639848 sshd[5672]: Accepted publickey for core from 10.0.0.1 port 35088 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:54.641614 sshd-session[5672]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:54.646583 systemd-logind[1587]: New session 18 of user core. Sep 16 04:50:54.656652 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 16 04:50:54.785190 sshd[5675]: Connection closed by 10.0.0.1 port 35088 Sep 16 04:50:54.785686 sshd-session[5672]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:54.795571 systemd[1]: sshd@17-10.0.0.34:22-10.0.0.1:35088.service: Deactivated successfully. Sep 16 04:50:54.799115 systemd[1]: session-18.scope: Deactivated successfully. Sep 16 04:50:54.800253 systemd-logind[1587]: Session 18 logged out. Waiting for processes to exit. Sep 16 04:50:54.804658 systemd[1]: Started sshd@18-10.0.0.34:22-10.0.0.1:35104.service - OpenSSH per-connection server daemon (10.0.0.1:35104). Sep 16 04:50:54.805581 systemd-logind[1587]: Removed session 18. Sep 16 04:50:54.871775 sshd[5688]: Accepted publickey for core from 10.0.0.1 port 35104 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:54.873625 sshd-session[5688]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:54.878640 systemd-logind[1587]: New session 19 of user core. Sep 16 04:50:54.890833 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 16 04:50:55.564870 sshd[5691]: Connection closed by 10.0.0.1 port 35104 Sep 16 04:50:55.565499 sshd-session[5688]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:55.579161 systemd[1]: sshd@18-10.0.0.34:22-10.0.0.1:35104.service: Deactivated successfully. Sep 16 04:50:55.581842 systemd[1]: session-19.scope: Deactivated successfully. Sep 16 04:50:55.583142 systemd-logind[1587]: Session 19 logged out. Waiting for processes to exit. Sep 16 04:50:55.586531 systemd[1]: Started sshd@19-10.0.0.34:22-10.0.0.1:35114.service - OpenSSH per-connection server daemon (10.0.0.1:35114). Sep 16 04:50:55.587611 systemd-logind[1587]: Removed session 19. Sep 16 04:50:55.663536 sshd[5710]: Accepted publickey for core from 10.0.0.1 port 35114 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:55.665648 sshd-session[5710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:55.671643 systemd-logind[1587]: New session 20 of user core. Sep 16 04:50:55.676794 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 16 04:50:56.503544 sshd[5713]: Connection closed by 10.0.0.1 port 35114 Sep 16 04:50:56.504818 sshd-session[5710]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:56.517816 systemd[1]: sshd@19-10.0.0.34:22-10.0.0.1:35114.service: Deactivated successfully. Sep 16 04:50:56.520917 systemd[1]: session-20.scope: Deactivated successfully. Sep 16 04:50:56.524434 systemd-logind[1587]: Session 20 logged out. Waiting for processes to exit. Sep 16 04:50:56.533821 systemd[1]: Started sshd@20-10.0.0.34:22-10.0.0.1:35130.service - OpenSSH per-connection server daemon (10.0.0.1:35130). Sep 16 04:50:56.536138 systemd-logind[1587]: Removed session 20. Sep 16 04:50:56.604694 sshd[5736]: Accepted publickey for core from 10.0.0.1 port 35130 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:56.606648 sshd-session[5736]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:56.612026 systemd-logind[1587]: New session 21 of user core. Sep 16 04:50:56.622752 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 16 04:50:56.926303 sshd[5739]: Connection closed by 10.0.0.1 port 35130 Sep 16 04:50:56.926897 sshd-session[5736]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:56.942319 systemd[1]: sshd@20-10.0.0.34:22-10.0.0.1:35130.service: Deactivated successfully. Sep 16 04:50:56.945483 systemd[1]: session-21.scope: Deactivated successfully. Sep 16 04:50:56.947332 systemd-logind[1587]: Session 21 logged out. Waiting for processes to exit. Sep 16 04:50:56.951708 systemd[1]: Started sshd@21-10.0.0.34:22-10.0.0.1:35136.service - OpenSSH per-connection server daemon (10.0.0.1:35136). Sep 16 04:50:56.952798 systemd-logind[1587]: Removed session 21. Sep 16 04:50:57.016747 sshd[5751]: Accepted publickey for core from 10.0.0.1 port 35136 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:50:57.018630 sshd-session[5751]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:50:57.024191 systemd-logind[1587]: New session 22 of user core. Sep 16 04:50:57.032670 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 16 04:50:57.148457 sshd[5754]: Connection closed by 10.0.0.1 port 35136 Sep 16 04:50:57.148845 sshd-session[5751]: pam_unix(sshd:session): session closed for user core Sep 16 04:50:57.154254 systemd[1]: sshd@21-10.0.0.34:22-10.0.0.1:35136.service: Deactivated successfully. Sep 16 04:50:57.156389 systemd[1]: session-22.scope: Deactivated successfully. Sep 16 04:50:57.157347 systemd-logind[1587]: Session 22 logged out. Waiting for processes to exit. Sep 16 04:50:57.158789 systemd-logind[1587]: Removed session 22. Sep 16 04:50:58.024632 containerd[1604]: time="2025-09-16T04:50:58.024557332Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b70cd33559cd4a539c2189530e530fa9da900f89effef6271094e9817677421d\" id:\"36da4a50b7f4d62db6ed5b7695406c67f29113db740a7dc4f2d70e1ee7d82b18\" pid:5778 exited_at:{seconds:1757998258 nanos:24130932}" Sep 16 04:51:02.170391 systemd[1]: Started sshd@22-10.0.0.34:22-10.0.0.1:34816.service - OpenSSH per-connection server daemon (10.0.0.1:34816). Sep 16 04:51:02.236700 sshd[5793]: Accepted publickey for core from 10.0.0.1 port 34816 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:51:02.239313 sshd-session[5793]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:51:02.244438 systemd-logind[1587]: New session 23 of user core. Sep 16 04:51:02.254698 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 16 04:51:02.391526 sshd[5796]: Connection closed by 10.0.0.1 port 34816 Sep 16 04:51:02.391976 sshd-session[5793]: pam_unix(sshd:session): session closed for user core Sep 16 04:51:02.399844 systemd[1]: sshd@22-10.0.0.34:22-10.0.0.1:34816.service: Deactivated successfully. Sep 16 04:51:02.402869 systemd[1]: session-23.scope: Deactivated successfully. Sep 16 04:51:02.404904 systemd-logind[1587]: Session 23 logged out. Waiting for processes to exit. Sep 16 04:51:02.407381 systemd-logind[1587]: Removed session 23. Sep 16 04:51:05.002791 containerd[1604]: time="2025-09-16T04:51:05.002679767Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d2afab445d7f19bd6add937e41ff7f7d672e412100d5ee7236e40b30eace87ba\" id:\"3afd115ca0c310452e367838dd3720543bf9a1e5b31de9fb3c746cbe6045d000\" pid:5826 exited_at:{seconds:1757998265 nanos:2117149}" Sep 16 04:51:07.410426 systemd[1]: Started sshd@23-10.0.0.34:22-10.0.0.1:34824.service - OpenSSH per-connection server daemon (10.0.0.1:34824). Sep 16 04:51:07.497673 sshd[5839]: Accepted publickey for core from 10.0.0.1 port 34824 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:51:07.499835 sshd-session[5839]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:51:07.505883 systemd-logind[1587]: New session 24 of user core. Sep 16 04:51:07.514786 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 16 04:51:07.644790 kubelet[2756]: E0916 04:51:07.644558 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:51:07.768786 sshd[5842]: Connection closed by 10.0.0.1 port 34824 Sep 16 04:51:07.769139 sshd-session[5839]: pam_unix(sshd:session): session closed for user core Sep 16 04:51:07.775178 systemd-logind[1587]: Session 24 logged out. Waiting for processes to exit. Sep 16 04:51:07.775879 systemd[1]: sshd@23-10.0.0.34:22-10.0.0.1:34824.service: Deactivated successfully. Sep 16 04:51:07.778412 systemd[1]: session-24.scope: Deactivated successfully. Sep 16 04:51:07.780457 systemd-logind[1587]: Removed session 24. Sep 16 04:51:11.644164 kubelet[2756]: E0916 04:51:11.644092 2756 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 16 04:51:12.797375 systemd[1]: Started sshd@24-10.0.0.34:22-10.0.0.1:37670.service - OpenSSH per-connection server daemon (10.0.0.1:37670). Sep 16 04:51:12.862531 sshd[5855]: Accepted publickey for core from 10.0.0.1 port 37670 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:51:12.864238 sshd-session[5855]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:51:12.870467 systemd-logind[1587]: New session 25 of user core. Sep 16 04:51:12.880689 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 16 04:51:13.066070 sshd[5858]: Connection closed by 10.0.0.1 port 37670 Sep 16 04:51:13.068098 sshd-session[5855]: pam_unix(sshd:session): session closed for user core Sep 16 04:51:13.072854 systemd[1]: sshd@24-10.0.0.34:22-10.0.0.1:37670.service: Deactivated successfully. Sep 16 04:51:13.076810 systemd[1]: session-25.scope: Deactivated successfully. Sep 16 04:51:13.078910 systemd-logind[1587]: Session 25 logged out. Waiting for processes to exit. Sep 16 04:51:13.080894 systemd-logind[1587]: Removed session 25. Sep 16 04:51:14.921111 containerd[1604]: time="2025-09-16T04:51:14.921054394Z" level=info msg="TaskExit event in podsandbox handler container_id:\"85e4fe5dafc7534f6bfd509b73014b8e6c0039936981430ca6a2f412d3a30b95\" id:\"2ecb6fcb40f0516d01d3978dbc83ff1a68470d9a91995173efd13523092a9a69\" pid:5883 exited_at:{seconds:1757998274 nanos:920425032}" Sep 16 04:51:18.080078 systemd[1]: Started sshd@25-10.0.0.34:22-10.0.0.1:37680.service - OpenSSH per-connection server daemon (10.0.0.1:37680). Sep 16 04:51:18.150093 sshd[5896]: Accepted publickey for core from 10.0.0.1 port 37680 ssh2: RSA SHA256:mbQbrRoQoFei5kIXvdhlqPTOzK4bL8i/kdyxZ8Q4lDE Sep 16 04:51:18.152077 sshd-session[5896]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 16 04:51:18.157661 systemd-logind[1587]: New session 26 of user core. Sep 16 04:51:18.164747 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 16 04:51:18.311499 sshd[5899]: Connection closed by 10.0.0.1 port 37680 Sep 16 04:51:18.311892 sshd-session[5896]: pam_unix(sshd:session): session closed for user core Sep 16 04:51:18.317658 systemd[1]: sshd@25-10.0.0.34:22-10.0.0.1:37680.service: Deactivated successfully. Sep 16 04:51:18.319796 systemd[1]: session-26.scope: Deactivated successfully. Sep 16 04:51:18.320937 systemd-logind[1587]: Session 26 logged out. Waiting for processes to exit. Sep 16 04:51:18.322143 systemd-logind[1587]: Removed session 26.