Sep 4 00:04:29.071490 kernel: Linux version 6.12.44-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed Sep 3 22:05:39 -00 2025 Sep 4 00:04:29.071562 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c7fa427551c105672074cbcbe7e23c997f471a6e879d708e8d6cbfad2147666e Sep 4 00:04:29.071581 kernel: BIOS-provided physical RAM map: Sep 4 00:04:29.071589 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000002ffff] usable Sep 4 00:04:29.071597 kernel: BIOS-e820: [mem 0x0000000000030000-0x000000000004ffff] reserved Sep 4 00:04:29.071606 kernel: BIOS-e820: [mem 0x0000000000050000-0x000000000009efff] usable Sep 4 00:04:29.071616 kernel: BIOS-e820: [mem 0x000000000009f000-0x000000000009ffff] reserved Sep 4 00:04:29.071625 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009b8ecfff] usable Sep 4 00:04:29.071634 kernel: BIOS-e820: [mem 0x000000009b8ed000-0x000000009bb6cfff] reserved Sep 4 00:04:29.071643 kernel: BIOS-e820: [mem 0x000000009bb6d000-0x000000009bb7efff] ACPI data Sep 4 00:04:29.071652 kernel: BIOS-e820: [mem 0x000000009bb7f000-0x000000009bbfefff] ACPI NVS Sep 4 00:04:29.071674 kernel: BIOS-e820: [mem 0x000000009bbff000-0x000000009bfb0fff] usable Sep 4 00:04:29.071683 kernel: BIOS-e820: [mem 0x000000009bfb1000-0x000000009bfb4fff] reserved Sep 4 00:04:29.071692 kernel: BIOS-e820: [mem 0x000000009bfb5000-0x000000009bfb6fff] ACPI NVS Sep 4 00:04:29.071713 kernel: BIOS-e820: [mem 0x000000009bfb7000-0x000000009bffffff] usable Sep 4 00:04:29.071722 kernel: BIOS-e820: [mem 0x000000009c000000-0x000000009cffffff] reserved Sep 4 00:04:29.071740 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 4 00:04:29.071749 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 4 00:04:29.071758 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 4 00:04:29.071766 kernel: NX (Execute Disable) protection: active Sep 4 00:04:29.071775 kernel: APIC: Static calls initialized Sep 4 00:04:29.071790 kernel: e820: update [mem 0x9a13e018-0x9a147c57] usable ==> usable Sep 4 00:04:29.071800 kernel: e820: update [mem 0x9a101018-0x9a13de57] usable ==> usable Sep 4 00:04:29.071819 kernel: extended physical RAM map: Sep 4 00:04:29.071828 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000002ffff] usable Sep 4 00:04:29.071838 kernel: reserve setup_data: [mem 0x0000000000030000-0x000000000004ffff] reserved Sep 4 00:04:29.071847 kernel: reserve setup_data: [mem 0x0000000000050000-0x000000000009efff] usable Sep 4 00:04:29.071860 kernel: reserve setup_data: [mem 0x000000000009f000-0x000000000009ffff] reserved Sep 4 00:04:29.071869 kernel: reserve setup_data: [mem 0x0000000000100000-0x000000009a101017] usable Sep 4 00:04:29.071878 kernel: reserve setup_data: [mem 0x000000009a101018-0x000000009a13de57] usable Sep 4 00:04:29.071892 kernel: reserve setup_data: [mem 0x000000009a13de58-0x000000009a13e017] usable Sep 4 00:04:29.071903 kernel: reserve setup_data: [mem 0x000000009a13e018-0x000000009a147c57] usable Sep 4 00:04:29.071913 kernel: reserve setup_data: [mem 0x000000009a147c58-0x000000009b8ecfff] usable Sep 4 00:04:29.071922 kernel: reserve setup_data: [mem 0x000000009b8ed000-0x000000009bb6cfff] reserved Sep 4 00:04:29.071932 kernel: reserve setup_data: [mem 0x000000009bb6d000-0x000000009bb7efff] ACPI data Sep 4 00:04:29.071947 kernel: reserve setup_data: [mem 0x000000009bb7f000-0x000000009bbfefff] ACPI NVS Sep 4 00:04:29.071964 kernel: reserve setup_data: [mem 0x000000009bbff000-0x000000009bfb0fff] usable Sep 4 00:04:29.071987 kernel: reserve setup_data: [mem 0x000000009bfb1000-0x000000009bfb4fff] reserved Sep 4 00:04:29.072005 kernel: reserve setup_data: [mem 0x000000009bfb5000-0x000000009bfb6fff] ACPI NVS Sep 4 00:04:29.072014 kernel: reserve setup_data: [mem 0x000000009bfb7000-0x000000009bffffff] usable Sep 4 00:04:29.072029 kernel: reserve setup_data: [mem 0x000000009c000000-0x000000009cffffff] reserved Sep 4 00:04:29.072040 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Sep 4 00:04:29.072050 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 4 00:04:29.072061 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 4 00:04:29.072073 kernel: efi: EFI v2.7 by EDK II Sep 4 00:04:29.072083 kernel: efi: SMBIOS=0x9b9d5000 ACPI=0x9bb7e000 ACPI 2.0=0x9bb7e014 MEMATTR=0x9a1af018 RNG=0x9bb73018 Sep 4 00:04:29.072101 kernel: random: crng init done Sep 4 00:04:29.072119 kernel: Kernel is locked down from EFI Secure Boot; see man kernel_lockdown.7 Sep 4 00:04:29.072136 kernel: secureboot: Secure boot enabled Sep 4 00:04:29.072146 kernel: SMBIOS 2.8 present. Sep 4 00:04:29.072155 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Sep 4 00:04:29.072172 kernel: DMI: Memory slots populated: 1/1 Sep 4 00:04:29.072190 kernel: Hypervisor detected: KVM Sep 4 00:04:29.072201 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 4 00:04:29.072215 kernel: kvm-clock: using sched offset of 7570073273 cycles Sep 4 00:04:29.072228 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 4 00:04:29.072238 kernel: tsc: Detected 2794.748 MHz processor Sep 4 00:04:29.072249 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 4 00:04:29.072259 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 4 00:04:29.072269 kernel: last_pfn = 0x9c000 max_arch_pfn = 0x400000000 Sep 4 00:04:29.072279 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 4 00:04:29.072293 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 4 00:04:29.072303 kernel: Using GB pages for direct mapping Sep 4 00:04:29.072315 kernel: ACPI: Early table checksum verification disabled Sep 4 00:04:29.072330 kernel: ACPI: RSDP 0x000000009BB7E014 000024 (v02 BOCHS ) Sep 4 00:04:29.072340 kernel: ACPI: XSDT 0x000000009BB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Sep 4 00:04:29.072350 kernel: ACPI: FACP 0x000000009BB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 00:04:29.072360 kernel: ACPI: DSDT 0x000000009BB7A000 002237 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 00:04:29.072375 kernel: ACPI: FACS 0x000000009BBDD000 000040 Sep 4 00:04:29.072393 kernel: ACPI: APIC 0x000000009BB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 00:04:29.072414 kernel: ACPI: HPET 0x000000009BB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 00:04:29.072437 kernel: ACPI: MCFG 0x000000009BB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 00:04:29.072494 kernel: ACPI: WAET 0x000000009BB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 00:04:29.072518 kernel: ACPI: BGRT 0x000000009BB74000 000038 (v01 INTEL EDK2 00000002 01000013) Sep 4 00:04:29.072533 kernel: ACPI: Reserving FACP table memory at [mem 0x9bb79000-0x9bb790f3] Sep 4 00:04:29.072544 kernel: ACPI: Reserving DSDT table memory at [mem 0x9bb7a000-0x9bb7c236] Sep 4 00:04:29.072554 kernel: ACPI: Reserving FACS table memory at [mem 0x9bbdd000-0x9bbdd03f] Sep 4 00:04:29.072564 kernel: ACPI: Reserving APIC table memory at [mem 0x9bb78000-0x9bb7808f] Sep 4 00:04:29.072573 kernel: ACPI: Reserving HPET table memory at [mem 0x9bb77000-0x9bb77037] Sep 4 00:04:29.072593 kernel: ACPI: Reserving MCFG table memory at [mem 0x9bb76000-0x9bb7603b] Sep 4 00:04:29.072603 kernel: ACPI: Reserving WAET table memory at [mem 0x9bb75000-0x9bb75027] Sep 4 00:04:29.072616 kernel: ACPI: Reserving BGRT table memory at [mem 0x9bb74000-0x9bb74037] Sep 4 00:04:29.072633 kernel: No NUMA configuration found Sep 4 00:04:29.072650 kernel: Faking a node at [mem 0x0000000000000000-0x000000009bffffff] Sep 4 00:04:29.072668 kernel: NODE_DATA(0) allocated [mem 0x9bf57dc0-0x9bf5efff] Sep 4 00:04:29.072685 kernel: Zone ranges: Sep 4 00:04:29.072704 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 4 00:04:29.072728 kernel: DMA32 [mem 0x0000000001000000-0x000000009bffffff] Sep 4 00:04:29.072745 kernel: Normal empty Sep 4 00:04:29.072766 kernel: Device empty Sep 4 00:04:29.072789 kernel: Movable zone start for each node Sep 4 00:04:29.072814 kernel: Early memory node ranges Sep 4 00:04:29.072824 kernel: node 0: [mem 0x0000000000001000-0x000000000002ffff] Sep 4 00:04:29.072834 kernel: node 0: [mem 0x0000000000050000-0x000000000009efff] Sep 4 00:04:29.072856 kernel: node 0: [mem 0x0000000000100000-0x000000009b8ecfff] Sep 4 00:04:29.072878 kernel: node 0: [mem 0x000000009bbff000-0x000000009bfb0fff] Sep 4 00:04:29.072911 kernel: node 0: [mem 0x000000009bfb7000-0x000000009bffffff] Sep 4 00:04:29.072928 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009bffffff] Sep 4 00:04:29.072946 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 4 00:04:29.072971 kernel: On node 0, zone DMA: 32 pages in unavailable ranges Sep 4 00:04:29.072986 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 4 00:04:29.072996 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Sep 4 00:04:29.073006 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Sep 4 00:04:29.073014 kernel: On node 0, zone DMA32: 16384 pages in unavailable ranges Sep 4 00:04:29.073022 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 4 00:04:29.073030 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 4 00:04:29.073038 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 4 00:04:29.073045 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 4 00:04:29.073053 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 4 00:04:29.073076 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 4 00:04:29.073084 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 4 00:04:29.073100 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 4 00:04:29.073108 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 4 00:04:29.073116 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 4 00:04:29.073131 kernel: TSC deadline timer available Sep 4 00:04:29.073145 kernel: CPU topo: Max. logical packages: 1 Sep 4 00:04:29.073153 kernel: CPU topo: Max. logical dies: 1 Sep 4 00:04:29.073161 kernel: CPU topo: Max. dies per package: 1 Sep 4 00:04:29.073187 kernel: CPU topo: Max. threads per core: 1 Sep 4 00:04:29.073195 kernel: CPU topo: Num. cores per package: 4 Sep 4 00:04:29.073203 kernel: CPU topo: Num. threads per package: 4 Sep 4 00:04:29.073216 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Sep 4 00:04:29.073224 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 4 00:04:29.073232 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 4 00:04:29.073251 kernel: kvm-guest: setup PV sched yield Sep 4 00:04:29.073267 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Sep 4 00:04:29.073281 kernel: Booting paravirtualized kernel on KVM Sep 4 00:04:29.073305 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 4 00:04:29.073333 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 4 00:04:29.073370 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Sep 4 00:04:29.073394 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Sep 4 00:04:29.073420 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 4 00:04:29.073435 kernel: kvm-guest: PV spinlocks enabled Sep 4 00:04:29.073451 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 4 00:04:29.073496 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c7fa427551c105672074cbcbe7e23c997f471a6e879d708e8d6cbfad2147666e Sep 4 00:04:29.073527 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 4 00:04:29.073542 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 4 00:04:29.073555 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 4 00:04:29.073566 kernel: Fallback order for Node 0: 0 Sep 4 00:04:29.073579 kernel: Built 1 zonelists, mobility grouping on. Total pages: 638054 Sep 4 00:04:29.073591 kernel: Policy zone: DMA32 Sep 4 00:04:29.073605 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 4 00:04:29.073617 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 4 00:04:29.073635 kernel: ftrace: allocating 40099 entries in 157 pages Sep 4 00:04:29.073647 kernel: ftrace: allocated 157 pages with 5 groups Sep 4 00:04:29.073662 kernel: Dynamic Preempt: voluntary Sep 4 00:04:29.073671 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 4 00:04:29.073680 kernel: rcu: RCU event tracing is enabled. Sep 4 00:04:29.073695 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 4 00:04:29.073708 kernel: Trampoline variant of Tasks RCU enabled. Sep 4 00:04:29.073723 kernel: Rude variant of Tasks RCU enabled. Sep 4 00:04:29.073740 kernel: Tracing variant of Tasks RCU enabled. Sep 4 00:04:29.073752 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 4 00:04:29.073771 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 4 00:04:29.073783 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 4 00:04:29.073792 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 4 00:04:29.073831 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 4 00:04:29.073846 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 4 00:04:29.073862 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 4 00:04:29.073878 kernel: Console: colour dummy device 80x25 Sep 4 00:04:29.073893 kernel: printk: legacy console [ttyS0] enabled Sep 4 00:04:29.073914 kernel: ACPI: Core revision 20240827 Sep 4 00:04:29.073929 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 4 00:04:29.073942 kernel: APIC: Switch to symmetric I/O mode setup Sep 4 00:04:29.073959 kernel: x2apic enabled Sep 4 00:04:29.073975 kernel: APIC: Switched APIC routing to: physical x2apic Sep 4 00:04:29.073987 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 4 00:04:29.073995 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 4 00:04:29.074004 kernel: kvm-guest: setup PV IPIs Sep 4 00:04:29.074012 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 4 00:04:29.074027 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Sep 4 00:04:29.074036 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Sep 4 00:04:29.074047 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 4 00:04:29.074055 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 4 00:04:29.074063 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 4 00:04:29.074072 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 4 00:04:29.074080 kernel: Spectre V2 : Mitigation: Retpolines Sep 4 00:04:29.074088 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 4 00:04:29.074096 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 4 00:04:29.074110 kernel: active return thunk: retbleed_return_thunk Sep 4 00:04:29.074118 kernel: RETBleed: Mitigation: untrained return thunk Sep 4 00:04:29.074129 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 4 00:04:29.074137 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 4 00:04:29.074146 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 4 00:04:29.074159 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 4 00:04:29.074167 kernel: active return thunk: srso_return_thunk Sep 4 00:04:29.074176 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 4 00:04:29.074191 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 4 00:04:29.074203 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 4 00:04:29.074217 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 4 00:04:29.074232 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 4 00:04:29.074241 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 4 00:04:29.074255 kernel: Freeing SMP alternatives memory: 32K Sep 4 00:04:29.074264 kernel: pid_max: default: 32768 minimum: 301 Sep 4 00:04:29.074274 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 4 00:04:29.074283 kernel: landlock: Up and running. Sep 4 00:04:29.074294 kernel: SELinux: Initializing. Sep 4 00:04:29.074302 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 4 00:04:29.074311 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 4 00:04:29.074319 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 4 00:04:29.074327 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 4 00:04:29.074335 kernel: ... version: 0 Sep 4 00:04:29.074346 kernel: ... bit width: 48 Sep 4 00:04:29.074354 kernel: ... generic registers: 6 Sep 4 00:04:29.074363 kernel: ... value mask: 0000ffffffffffff Sep 4 00:04:29.074373 kernel: ... max period: 00007fffffffffff Sep 4 00:04:29.074381 kernel: ... fixed-purpose events: 0 Sep 4 00:04:29.074393 kernel: ... event mask: 000000000000003f Sep 4 00:04:29.074404 kernel: signal: max sigframe size: 1776 Sep 4 00:04:29.074412 kernel: rcu: Hierarchical SRCU implementation. Sep 4 00:04:29.074421 kernel: rcu: Max phase no-delay instances is 400. Sep 4 00:04:29.074429 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 4 00:04:29.074443 kernel: smp: Bringing up secondary CPUs ... Sep 4 00:04:29.074452 kernel: smpboot: x86: Booting SMP configuration: Sep 4 00:04:29.074460 kernel: .... node #0, CPUs: #1 #2 #3 Sep 4 00:04:29.074489 kernel: smp: Brought up 1 node, 4 CPUs Sep 4 00:04:29.074500 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Sep 4 00:04:29.074511 kernel: Memory: 2411268K/2552216K available (14336K kernel code, 2428K rwdata, 9956K rodata, 53832K init, 1088K bss, 135016K reserved, 0K cma-reserved) Sep 4 00:04:29.074522 kernel: devtmpfs: initialized Sep 4 00:04:29.074531 kernel: x86/mm: Memory block size: 128MB Sep 4 00:04:29.074540 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9bb7f000-0x9bbfefff] (524288 bytes) Sep 4 00:04:29.074548 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9bfb5000-0x9bfb6fff] (8192 bytes) Sep 4 00:04:29.074556 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 4 00:04:29.074568 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 4 00:04:29.074576 kernel: pinctrl core: initialized pinctrl subsystem Sep 4 00:04:29.074584 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 4 00:04:29.074595 kernel: audit: initializing netlink subsys (disabled) Sep 4 00:04:29.074603 kernel: audit: type=2000 audit(1756944266.168:1): state=initialized audit_enabled=0 res=1 Sep 4 00:04:29.074612 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 4 00:04:29.074620 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 4 00:04:29.074628 kernel: cpuidle: using governor menu Sep 4 00:04:29.074636 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 4 00:04:29.074647 kernel: dca service started, version 1.12.1 Sep 4 00:04:29.074655 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Sep 4 00:04:29.074663 kernel: PCI: Using configuration type 1 for base access Sep 4 00:04:29.074672 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 4 00:04:29.074680 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 4 00:04:29.074690 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 4 00:04:29.074699 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 4 00:04:29.074707 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 4 00:04:29.074715 kernel: ACPI: Added _OSI(Module Device) Sep 4 00:04:29.074727 kernel: ACPI: Added _OSI(Processor Device) Sep 4 00:04:29.074736 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 4 00:04:29.074744 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 4 00:04:29.074752 kernel: ACPI: Interpreter enabled Sep 4 00:04:29.074760 kernel: ACPI: PM: (supports S0 S5) Sep 4 00:04:29.074768 kernel: ACPI: Using IOAPIC for interrupt routing Sep 4 00:04:29.074777 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 4 00:04:29.074785 kernel: PCI: Using E820 reservations for host bridge windows Sep 4 00:04:29.074793 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 4 00:04:29.074814 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 4 00:04:29.075062 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 4 00:04:29.075953 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 4 00:04:29.076146 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 4 00:04:29.076162 kernel: PCI host bridge to bus 0000:00 Sep 4 00:04:29.076307 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 4 00:04:29.076447 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 4 00:04:29.076622 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 4 00:04:29.076737 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Sep 4 00:04:29.076871 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Sep 4 00:04:29.077023 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Sep 4 00:04:29.077179 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 4 00:04:29.077380 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Sep 4 00:04:29.077590 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Sep 4 00:04:29.077747 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Sep 4 00:04:29.077935 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Sep 4 00:04:29.078106 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Sep 4 00:04:29.078269 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 4 00:04:29.078482 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 4 00:04:29.078641 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Sep 4 00:04:29.078796 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Sep 4 00:04:29.078959 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Sep 4 00:04:29.079144 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Sep 4 00:04:29.079314 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Sep 4 00:04:29.079515 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Sep 4 00:04:29.079670 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Sep 4 00:04:29.079865 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Sep 4 00:04:29.080025 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Sep 4 00:04:29.080177 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Sep 4 00:04:29.080327 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Sep 4 00:04:29.080500 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Sep 4 00:04:29.080687 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Sep 4 00:04:29.080859 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 4 00:04:29.081048 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Sep 4 00:04:29.081213 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Sep 4 00:04:29.081400 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Sep 4 00:04:29.081593 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Sep 4 00:04:29.081746 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Sep 4 00:04:29.081768 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 4 00:04:29.081780 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 4 00:04:29.081791 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 4 00:04:29.081822 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 4 00:04:29.081833 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 4 00:04:29.081852 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 4 00:04:29.081868 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 4 00:04:29.081886 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 4 00:04:29.081905 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 4 00:04:29.081916 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 4 00:04:29.081928 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 4 00:04:29.081939 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 4 00:04:29.081958 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 4 00:04:29.081969 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 4 00:04:29.081980 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 4 00:04:29.081993 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 4 00:04:29.082005 kernel: iommu: Default domain type: Translated Sep 4 00:04:29.082018 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 4 00:04:29.082030 kernel: efivars: Registered efivars operations Sep 4 00:04:29.082041 kernel: PCI: Using ACPI for IRQ routing Sep 4 00:04:29.082052 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 4 00:04:29.082066 kernel: e820: reserve RAM buffer [mem 0x0009f000-0x0009ffff] Sep 4 00:04:29.082077 kernel: e820: reserve RAM buffer [mem 0x9a101018-0x9bffffff] Sep 4 00:04:29.082088 kernel: e820: reserve RAM buffer [mem 0x9a13e018-0x9bffffff] Sep 4 00:04:29.082099 kernel: e820: reserve RAM buffer [mem 0x9b8ed000-0x9bffffff] Sep 4 00:04:29.082110 kernel: e820: reserve RAM buffer [mem 0x9bfb1000-0x9bffffff] Sep 4 00:04:29.082288 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 4 00:04:29.082562 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 4 00:04:29.082791 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 4 00:04:29.082826 kernel: vgaarb: loaded Sep 4 00:04:29.082838 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 4 00:04:29.082850 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 4 00:04:29.082862 kernel: clocksource: Switched to clocksource kvm-clock Sep 4 00:04:29.082873 kernel: VFS: Disk quotas dquot_6.6.0 Sep 4 00:04:29.082885 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 4 00:04:29.082896 kernel: pnp: PnP ACPI init Sep 4 00:04:29.083078 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Sep 4 00:04:29.083100 kernel: pnp: PnP ACPI: found 6 devices Sep 4 00:04:29.083112 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 4 00:04:29.083124 kernel: NET: Registered PF_INET protocol family Sep 4 00:04:29.083136 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 4 00:04:29.083148 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 4 00:04:29.083159 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 4 00:04:29.083170 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 4 00:04:29.083182 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 4 00:04:29.083193 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 4 00:04:29.083208 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 4 00:04:29.083219 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 4 00:04:29.083231 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 4 00:04:29.083243 kernel: NET: Registered PF_XDP protocol family Sep 4 00:04:29.083421 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Sep 4 00:04:29.083610 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Sep 4 00:04:29.083768 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 4 00:04:29.083918 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 4 00:04:29.084065 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 4 00:04:29.084204 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Sep 4 00:04:29.084366 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Sep 4 00:04:29.084527 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Sep 4 00:04:29.084543 kernel: PCI: CLS 0 bytes, default 64 Sep 4 00:04:29.084555 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Sep 4 00:04:29.084567 kernel: Initialise system trusted keyrings Sep 4 00:04:29.084578 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 4 00:04:29.084588 kernel: Key type asymmetric registered Sep 4 00:04:29.084604 kernel: Asymmetric key parser 'x509' registered Sep 4 00:04:29.084884 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 4 00:04:29.084901 kernel: io scheduler mq-deadline registered Sep 4 00:04:29.084913 kernel: io scheduler kyber registered Sep 4 00:04:29.084925 kernel: io scheduler bfq registered Sep 4 00:04:29.084936 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 4 00:04:29.084954 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 4 00:04:29.084971 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 4 00:04:29.084986 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 4 00:04:29.085002 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 4 00:04:29.085014 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 4 00:04:29.085026 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 4 00:04:29.085040 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 4 00:04:29.085051 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 4 00:04:29.085381 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 4 00:04:29.085706 kernel: rtc_cmos 00:04: registered as rtc0 Sep 4 00:04:29.085728 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 4 00:04:29.085907 kernel: rtc_cmos 00:04: setting system clock to 2025-09-04T00:04:28 UTC (1756944268) Sep 4 00:04:29.086328 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Sep 4 00:04:29.086362 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 4 00:04:29.086382 kernel: efifb: probing for efifb Sep 4 00:04:29.086398 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Sep 4 00:04:29.086414 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Sep 4 00:04:29.086428 kernel: efifb: scrolling: redraw Sep 4 00:04:29.086442 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 4 00:04:29.086458 kernel: Console: switching to colour frame buffer device 160x50 Sep 4 00:04:29.086492 kernel: fb0: EFI VGA frame buffer device Sep 4 00:04:29.086507 kernel: pstore: Using crash dump compression: deflate Sep 4 00:04:29.086519 kernel: pstore: Registered efi_pstore as persistent store backend Sep 4 00:04:29.086533 kernel: NET: Registered PF_INET6 protocol family Sep 4 00:04:29.086546 kernel: Segment Routing with IPv6 Sep 4 00:04:29.086559 kernel: In-situ OAM (IOAM) with IPv6 Sep 4 00:04:29.086571 kernel: NET: Registered PF_PACKET protocol family Sep 4 00:04:29.086583 kernel: Key type dns_resolver registered Sep 4 00:04:29.086595 kernel: IPI shorthand broadcast: enabled Sep 4 00:04:29.086606 kernel: sched_clock: Marking stable (3517007571, 149612150)->(3705376527, -38756806) Sep 4 00:04:29.086618 kernel: registered taskstats version 1 Sep 4 00:04:29.086642 kernel: Loading compiled-in X.509 certificates Sep 4 00:04:29.086662 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.44-flatcar: 247a8159a15e16f8eb89737aa66cd9cf9bbb3c10' Sep 4 00:04:29.086690 kernel: Demotion targets for Node 0: null Sep 4 00:04:29.086717 kernel: Key type .fscrypt registered Sep 4 00:04:29.086737 kernel: Key type fscrypt-provisioning registered Sep 4 00:04:29.086750 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 4 00:04:29.086769 kernel: ima: Allocated hash algorithm: sha1 Sep 4 00:04:29.086781 kernel: ima: No architecture policies found Sep 4 00:04:29.086792 kernel: clk: Disabling unused clocks Sep 4 00:04:29.086818 kernel: Warning: unable to open an initial console. Sep 4 00:04:29.086830 kernel: Freeing unused kernel image (initmem) memory: 53832K Sep 4 00:04:29.086841 kernel: Write protecting the kernel read-only data: 24576k Sep 4 00:04:29.086858 kernel: Freeing unused kernel image (rodata/data gap) memory: 284K Sep 4 00:04:29.086870 kernel: Run /init as init process Sep 4 00:04:29.086881 kernel: with arguments: Sep 4 00:04:29.086893 kernel: /init Sep 4 00:04:29.086904 kernel: with environment: Sep 4 00:04:29.086915 kernel: HOME=/ Sep 4 00:04:29.086926 kernel: TERM=linux Sep 4 00:04:29.086950 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 4 00:04:29.086973 systemd[1]: Successfully made /usr/ read-only. Sep 4 00:04:29.086999 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 4 00:04:29.087013 systemd[1]: Detected virtualization kvm. Sep 4 00:04:29.087026 systemd[1]: Detected architecture x86-64. Sep 4 00:04:29.087038 systemd[1]: Running in initrd. Sep 4 00:04:29.087049 systemd[1]: No hostname configured, using default hostname. Sep 4 00:04:29.087063 systemd[1]: Hostname set to . Sep 4 00:04:29.087075 systemd[1]: Initializing machine ID from VM UUID. Sep 4 00:04:29.087091 systemd[1]: Queued start job for default target initrd.target. Sep 4 00:04:29.087103 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 00:04:29.087116 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 00:04:29.087129 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 4 00:04:29.087151 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 00:04:29.087178 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 4 00:04:29.087193 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 4 00:04:29.087215 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 4 00:04:29.087227 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 4 00:04:29.087240 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 00:04:29.087253 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 00:04:29.087265 systemd[1]: Reached target paths.target - Path Units. Sep 4 00:04:29.087278 systemd[1]: Reached target slices.target - Slice Units. Sep 4 00:04:29.087290 systemd[1]: Reached target swap.target - Swaps. Sep 4 00:04:29.087302 systemd[1]: Reached target timers.target - Timer Units. Sep 4 00:04:29.087319 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 00:04:29.087332 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 00:04:29.087345 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 4 00:04:29.087369 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 4 00:04:29.087390 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 00:04:29.087403 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 00:04:29.087416 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 00:04:29.087428 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 00:04:29.087444 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 4 00:04:29.087464 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 00:04:29.087510 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 4 00:04:29.087524 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 4 00:04:29.087536 systemd[1]: Starting systemd-fsck-usr.service... Sep 4 00:04:29.087549 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 00:04:29.087562 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 00:04:29.087574 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 00:04:29.087586 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 4 00:04:29.087612 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 00:04:29.087624 systemd[1]: Finished systemd-fsck-usr.service. Sep 4 00:04:29.087645 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 00:04:29.087706 systemd-journald[220]: Collecting audit messages is disabled. Sep 4 00:04:29.087763 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 00:04:29.087776 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 00:04:29.087797 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 00:04:29.087823 systemd-journald[220]: Journal started Sep 4 00:04:29.087858 systemd-journald[220]: Runtime Journal (/run/log/journal/f99aefdb731b4c16a34c05e00bee0899) is 6M, max 48.2M, 42.2M free. Sep 4 00:04:29.073408 systemd-modules-load[221]: Inserted module 'overlay' Sep 4 00:04:29.094646 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 00:04:29.094693 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 00:04:29.098513 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 00:04:29.113883 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 4 00:04:29.114446 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 00:04:29.119638 kernel: Bridge firewalling registered Sep 4 00:04:29.119119 systemd-modules-load[221]: Inserted module 'br_netfilter' Sep 4 00:04:29.119599 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 4 00:04:29.121144 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 00:04:29.127151 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 00:04:29.127232 systemd-tmpfiles[247]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 4 00:04:29.132979 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 00:04:29.136158 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 00:04:29.150634 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 00:04:29.153257 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c7fa427551c105672074cbcbe7e23c997f471a6e879d708e8d6cbfad2147666e Sep 4 00:04:29.154700 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 00:04:29.221265 systemd-resolved[273]: Positive Trust Anchors: Sep 4 00:04:29.221291 systemd-resolved[273]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 00:04:29.221337 systemd-resolved[273]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 00:04:29.224712 systemd-resolved[273]: Defaulting to hostname 'linux'. Sep 4 00:04:29.226150 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 00:04:29.236315 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 00:04:29.329525 kernel: SCSI subsystem initialized Sep 4 00:04:29.339517 kernel: Loading iSCSI transport class v2.0-870. Sep 4 00:04:29.351527 kernel: iscsi: registered transport (tcp) Sep 4 00:04:29.381647 kernel: iscsi: registered transport (qla4xxx) Sep 4 00:04:29.381743 kernel: QLogic iSCSI HBA Driver Sep 4 00:04:29.408838 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 4 00:04:29.427478 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 4 00:04:29.428162 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 4 00:04:29.501341 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 4 00:04:29.503534 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 4 00:04:29.573553 kernel: raid6: avx2x4 gen() 19542 MB/s Sep 4 00:04:29.590556 kernel: raid6: avx2x2 gen() 24493 MB/s Sep 4 00:04:29.607910 kernel: raid6: avx2x1 gen() 21609 MB/s Sep 4 00:04:29.607997 kernel: raid6: using algorithm avx2x2 gen() 24493 MB/s Sep 4 00:04:29.625681 kernel: raid6: .... xor() 17188 MB/s, rmw enabled Sep 4 00:04:29.625759 kernel: raid6: using avx2x2 recovery algorithm Sep 4 00:04:29.647524 kernel: xor: automatically using best checksumming function avx Sep 4 00:04:29.822526 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 4 00:04:29.831376 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 4 00:04:29.833768 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 00:04:29.869311 systemd-udevd[471]: Using default interface naming scheme 'v255'. Sep 4 00:04:29.875482 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 00:04:29.879113 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 4 00:04:29.908392 dracut-pre-trigger[480]: rd.md=0: removing MD RAID activation Sep 4 00:04:29.949685 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 00:04:29.951508 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 00:04:30.046312 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 00:04:30.049157 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 4 00:04:30.142512 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 4 00:04:30.145204 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 4 00:04:30.154520 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 4 00:04:30.154608 kernel: GPT:9289727 != 19775487 Sep 4 00:04:30.154619 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 4 00:04:30.154630 kernel: GPT:9289727 != 19775487 Sep 4 00:04:30.154639 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 4 00:04:30.154649 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 00:04:30.162518 kernel: cryptd: max_cpu_qlen set to 1000 Sep 4 00:04:30.169557 kernel: libata version 3.00 loaded. Sep 4 00:04:30.181563 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 00:04:30.181797 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 00:04:30.184536 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 00:04:30.189780 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 00:04:30.194525 kernel: AES CTR mode by8 optimization enabled Sep 4 00:04:30.194560 kernel: ahci 0000:00:1f.2: version 3.0 Sep 4 00:04:30.194427 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 4 00:04:30.219514 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 4 00:04:30.275542 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 00:04:30.275748 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 00:04:30.279047 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 4 00:04:30.283885 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 00:04:30.294638 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Sep 4 00:04:30.299584 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Sep 4 00:04:30.299811 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 4 00:04:30.300005 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Sep 4 00:04:30.312504 kernel: scsi host0: ahci Sep 4 00:04:30.318498 kernel: scsi host1: ahci Sep 4 00:04:30.319146 kernel: scsi host2: ahci Sep 4 00:04:30.319351 kernel: scsi host3: ahci Sep 4 00:04:30.320726 kernel: scsi host4: ahci Sep 4 00:04:30.320946 kernel: scsi host5: ahci Sep 4 00:04:30.321253 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 4 00:04:30.331114 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 1 Sep 4 00:04:30.331166 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 1 Sep 4 00:04:30.331180 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 1 Sep 4 00:04:30.331205 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 1 Sep 4 00:04:30.331217 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 1 Sep 4 00:04:30.331229 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 1 Sep 4 00:04:30.340978 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 4 00:04:30.341791 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 4 00:04:30.356321 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 4 00:04:30.378012 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 4 00:04:30.378613 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 00:04:30.393176 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 4 00:04:30.426904 disk-uuid[637]: Primary Header is updated. Sep 4 00:04:30.426904 disk-uuid[637]: Secondary Entries is updated. Sep 4 00:04:30.426904 disk-uuid[637]: Secondary Header is updated. Sep 4 00:04:30.431494 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 00:04:30.438517 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 00:04:30.638839 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 4 00:04:30.638927 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 4 00:04:30.639511 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 4 00:04:30.640537 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 4 00:04:30.641508 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 4 00:04:30.642506 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 4 00:04:30.643511 kernel: ata3.00: LPM support broken, forcing max_power Sep 4 00:04:30.644602 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 4 00:04:30.644642 kernel: ata3.00: applying bridge limits Sep 4 00:04:30.647494 kernel: ata3.00: LPM support broken, forcing max_power Sep 4 00:04:30.647578 kernel: ata3.00: configured for UDMA/100 Sep 4 00:04:30.649514 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 4 00:04:30.707640 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 4 00:04:30.708026 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 4 00:04:30.722527 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 4 00:04:31.084506 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 4 00:04:31.085364 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 00:04:31.090089 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 00:04:31.092883 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 00:04:31.096594 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 4 00:04:31.132351 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 4 00:04:31.438486 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 00:04:31.439009 disk-uuid[638]: The operation has completed successfully. Sep 4 00:04:31.478265 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 4 00:04:31.478398 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 4 00:04:31.508725 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 4 00:04:31.534412 sh[667]: Success Sep 4 00:04:31.557078 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 4 00:04:31.557153 kernel: device-mapper: uevent: version 1.0.3 Sep 4 00:04:31.558414 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 4 00:04:31.569565 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Sep 4 00:04:31.611337 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 4 00:04:31.615217 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 4 00:04:31.629722 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 4 00:04:31.636495 kernel: BTRFS: device fsid 8a9c2e34-3d3c-49a9-acce-59bf90003071 devid 1 transid 37 /dev/mapper/usr (253:0) scanned by mount (679) Sep 4 00:04:31.636547 kernel: BTRFS info (device dm-0): first mount of filesystem 8a9c2e34-3d3c-49a9-acce-59bf90003071 Sep 4 00:04:31.636563 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 4 00:04:31.642731 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 4 00:04:31.642791 kernel: BTRFS info (device dm-0): enabling free space tree Sep 4 00:04:31.644491 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 4 00:04:31.646398 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 4 00:04:31.648063 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 4 00:04:31.649173 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 4 00:04:31.651206 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 4 00:04:31.684190 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (710) Sep 4 00:04:31.684261 kernel: BTRFS info (device vda6): first mount of filesystem 75efd3be-3398-4525-8f67-b36cc847539d Sep 4 00:04:31.684279 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 00:04:31.689516 kernel: BTRFS info (device vda6): turning on async discard Sep 4 00:04:31.689564 kernel: BTRFS info (device vda6): enabling free space tree Sep 4 00:04:31.695538 kernel: BTRFS info (device vda6): last unmount of filesystem 75efd3be-3398-4525-8f67-b36cc847539d Sep 4 00:04:31.696912 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 4 00:04:31.698381 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 4 00:04:31.802865 ignition[753]: Ignition 2.21.0 Sep 4 00:04:31.802879 ignition[753]: Stage: fetch-offline Sep 4 00:04:31.802918 ignition[753]: no configs at "/usr/lib/ignition/base.d" Sep 4 00:04:31.802929 ignition[753]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 00:04:31.803019 ignition[753]: parsed url from cmdline: "" Sep 4 00:04:31.803022 ignition[753]: no config URL provided Sep 4 00:04:31.803027 ignition[753]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 00:04:31.803036 ignition[753]: no config at "/usr/lib/ignition/user.ign" Sep 4 00:04:31.803061 ignition[753]: op(1): [started] loading QEMU firmware config module Sep 4 00:04:31.803066 ignition[753]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 4 00:04:31.828811 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 00:04:31.841879 ignition[753]: op(1): [finished] loading QEMU firmware config module Sep 4 00:04:31.844686 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 00:04:31.886319 ignition[753]: parsing config with SHA512: e520fac5ff95d4f4fbd01ae0b32813a56ed3f37f0dd3af0f4acb5f030138d7949f5425ad034b0336f65ea1c29510b31267f0fe2fff40fa1fe1f201e4cc7a141a Sep 4 00:04:31.892212 unknown[753]: fetched base config from "system" Sep 4 00:04:31.892400 unknown[753]: fetched user config from "qemu" Sep 4 00:04:31.893162 ignition[753]: fetch-offline: fetch-offline passed Sep 4 00:04:31.893225 ignition[753]: Ignition finished successfully Sep 4 00:04:31.896643 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 00:04:31.902318 systemd-networkd[857]: lo: Link UP Sep 4 00:04:31.902332 systemd-networkd[857]: lo: Gained carrier Sep 4 00:04:31.903925 systemd-networkd[857]: Enumeration completed Sep 4 00:04:31.904015 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 00:04:31.904302 systemd-networkd[857]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 00:04:31.904307 systemd-networkd[857]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 00:04:31.905693 systemd-networkd[857]: eth0: Link UP Sep 4 00:04:31.905837 systemd-networkd[857]: eth0: Gained carrier Sep 4 00:04:31.905846 systemd-networkd[857]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 00:04:31.907360 systemd[1]: Reached target network.target - Network. Sep 4 00:04:31.908812 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 4 00:04:31.909947 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 4 00:04:31.929532 systemd-networkd[857]: eth0: DHCPv4 address 10.0.0.91/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 4 00:04:31.949465 ignition[861]: Ignition 2.21.0 Sep 4 00:04:31.949498 ignition[861]: Stage: kargs Sep 4 00:04:31.949624 ignition[861]: no configs at "/usr/lib/ignition/base.d" Sep 4 00:04:31.949634 ignition[861]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 00:04:31.951031 ignition[861]: kargs: kargs passed Sep 4 00:04:31.951145 ignition[861]: Ignition finished successfully Sep 4 00:04:31.956551 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 4 00:04:31.958914 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 4 00:04:31.999102 ignition[870]: Ignition 2.21.0 Sep 4 00:04:31.999118 ignition[870]: Stage: disks Sep 4 00:04:31.999249 ignition[870]: no configs at "/usr/lib/ignition/base.d" Sep 4 00:04:31.999260 ignition[870]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 00:04:32.003617 ignition[870]: disks: disks passed Sep 4 00:04:32.003712 ignition[870]: Ignition finished successfully Sep 4 00:04:32.007137 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 4 00:04:32.008430 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 4 00:04:32.010389 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 4 00:04:32.011461 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 00:04:32.011983 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 00:04:32.012296 systemd[1]: Reached target basic.target - Basic System. Sep 4 00:04:32.013883 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 4 00:04:32.053037 systemd-fsck[880]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 4 00:04:32.518172 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 4 00:04:32.520528 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 4 00:04:32.640533 kernel: EXT4-fs (vda9): mounted filesystem c3518c93-f823-4477-a620-ff9666a59be5 r/w with ordered data mode. Quota mode: none. Sep 4 00:04:32.641270 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 4 00:04:32.642735 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 4 00:04:32.645622 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 00:04:32.646590 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 4 00:04:32.648354 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 4 00:04:32.648392 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 4 00:04:32.648414 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 00:04:32.666282 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 4 00:04:32.667803 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 4 00:04:32.673554 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (889) Sep 4 00:04:32.673585 kernel: BTRFS info (device vda6): first mount of filesystem 75efd3be-3398-4525-8f67-b36cc847539d Sep 4 00:04:32.673600 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 00:04:32.678013 kernel: BTRFS info (device vda6): turning on async discard Sep 4 00:04:32.678068 kernel: BTRFS info (device vda6): enabling free space tree Sep 4 00:04:32.679497 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 00:04:32.706823 initrd-setup-root[913]: cut: /sysroot/etc/passwd: No such file or directory Sep 4 00:04:32.712257 initrd-setup-root[920]: cut: /sysroot/etc/group: No such file or directory Sep 4 00:04:32.717194 initrd-setup-root[927]: cut: /sysroot/etc/shadow: No such file or directory Sep 4 00:04:32.721975 initrd-setup-root[934]: cut: /sysroot/etc/gshadow: No such file or directory Sep 4 00:04:32.851227 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 4 00:04:32.854387 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 4 00:04:32.856310 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 4 00:04:32.869818 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 4 00:04:32.882916 kernel: BTRFS info (device vda6): last unmount of filesystem 75efd3be-3398-4525-8f67-b36cc847539d Sep 4 00:04:32.895684 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 4 00:04:32.915039 ignition[1004]: INFO : Ignition 2.21.0 Sep 4 00:04:32.915039 ignition[1004]: INFO : Stage: mount Sep 4 00:04:32.917616 ignition[1004]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 00:04:32.917616 ignition[1004]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 00:04:32.919981 ignition[1004]: INFO : mount: mount passed Sep 4 00:04:32.919981 ignition[1004]: INFO : Ignition finished successfully Sep 4 00:04:32.921724 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 4 00:04:32.924161 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 4 00:04:33.643161 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 00:04:33.670494 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1016) Sep 4 00:04:33.672772 kernel: BTRFS info (device vda6): first mount of filesystem 75efd3be-3398-4525-8f67-b36cc847539d Sep 4 00:04:33.672810 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 00:04:33.676673 kernel: BTRFS info (device vda6): turning on async discard Sep 4 00:04:33.676765 kernel: BTRFS info (device vda6): enabling free space tree Sep 4 00:04:33.678634 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 00:04:33.711231 ignition[1033]: INFO : Ignition 2.21.0 Sep 4 00:04:33.711231 ignition[1033]: INFO : Stage: files Sep 4 00:04:33.713164 ignition[1033]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 00:04:33.713164 ignition[1033]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 00:04:33.715422 ignition[1033]: DEBUG : files: compiled without relabeling support, skipping Sep 4 00:04:33.716576 ignition[1033]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 4 00:04:33.716576 ignition[1033]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 4 00:04:33.719577 ignition[1033]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 4 00:04:33.719577 ignition[1033]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 4 00:04:33.719577 ignition[1033]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 4 00:04:33.718654 unknown[1033]: wrote ssh authorized keys file for user: core Sep 4 00:04:33.724885 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 4 00:04:33.724885 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Sep 4 00:04:33.826393 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 4 00:04:33.935637 systemd-networkd[857]: eth0: Gained IPv6LL Sep 4 00:04:34.857707 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 4 00:04:34.857707 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 4 00:04:34.862009 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 4 00:04:34.944168 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 4 00:04:35.632381 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 4 00:04:35.632381 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 4 00:04:35.636849 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 4 00:04:35.638656 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 4 00:04:35.640773 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 4 00:04:35.642804 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 00:04:35.644999 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 00:04:35.646925 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 00:04:35.649131 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 00:04:35.656214 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 00:04:35.658754 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 00:04:35.658754 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 4 00:04:35.663876 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 4 00:04:35.663876 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 4 00:04:35.663876 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Sep 4 00:04:35.908898 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 4 00:04:36.740894 ignition[1033]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 4 00:04:36.740894 ignition[1033]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 4 00:04:36.744454 ignition[1033]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 00:04:36.961898 ignition[1033]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 00:04:36.961898 ignition[1033]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 4 00:04:36.961898 ignition[1033]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Sep 4 00:04:36.961898 ignition[1033]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 4 00:04:36.969481 ignition[1033]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 4 00:04:36.969481 ignition[1033]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Sep 4 00:04:36.969481 ignition[1033]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Sep 4 00:04:37.034483 ignition[1033]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 4 00:04:37.043146 ignition[1033]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 4 00:04:37.044768 ignition[1033]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Sep 4 00:04:37.044768 ignition[1033]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Sep 4 00:04:37.044768 ignition[1033]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Sep 4 00:04:37.044768 ignition[1033]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 4 00:04:37.044768 ignition[1033]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 4 00:04:37.044768 ignition[1033]: INFO : files: files passed Sep 4 00:04:37.044768 ignition[1033]: INFO : Ignition finished successfully Sep 4 00:04:37.056463 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 4 00:04:37.058912 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 4 00:04:37.059672 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 4 00:04:37.077179 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 4 00:04:37.077305 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 4 00:04:37.080367 initrd-setup-root-after-ignition[1062]: grep: /sysroot/oem/oem-release: No such file or directory Sep 4 00:04:37.084968 initrd-setup-root-after-ignition[1064]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 00:04:37.086714 initrd-setup-root-after-ignition[1064]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 4 00:04:37.088296 initrd-setup-root-after-ignition[1068]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 00:04:37.089406 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 00:04:37.091327 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 4 00:04:37.094327 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 4 00:04:37.169185 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 4 00:04:37.169343 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 4 00:04:37.172226 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 4 00:04:37.174684 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 4 00:04:37.176948 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 4 00:04:37.179678 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 4 00:04:37.213950 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 00:04:37.216500 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 4 00:04:37.247257 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 4 00:04:37.249628 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 00:04:37.249791 systemd[1]: Stopped target timers.target - Timer Units. Sep 4 00:04:37.252023 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 4 00:04:37.252165 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 00:04:37.256828 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 4 00:04:37.257921 systemd[1]: Stopped target basic.target - Basic System. Sep 4 00:04:37.259825 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 4 00:04:37.260733 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 00:04:37.261062 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 4 00:04:37.261393 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 4 00:04:37.261924 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 4 00:04:37.262263 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 00:04:37.262788 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 4 00:04:37.263110 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 4 00:04:37.263436 systemd[1]: Stopped target swap.target - Swaps. Sep 4 00:04:37.263923 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 4 00:04:37.264048 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 4 00:04:37.280259 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 4 00:04:37.280460 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 00:04:37.284539 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 4 00:04:37.284716 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 00:04:37.287986 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 4 00:04:37.288165 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 4 00:04:37.291286 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 4 00:04:37.291459 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 00:04:37.294885 systemd[1]: Stopped target paths.target - Path Units. Sep 4 00:04:37.296613 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 4 00:04:37.300574 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 00:04:37.303421 systemd[1]: Stopped target slices.target - Slice Units. Sep 4 00:04:37.305565 systemd[1]: Stopped target sockets.target - Socket Units. Sep 4 00:04:37.305779 systemd[1]: iscsid.socket: Deactivated successfully. Sep 4 00:04:37.305925 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 00:04:37.309397 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 4 00:04:37.309508 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 00:04:37.310340 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 4 00:04:37.310488 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 00:04:37.312151 systemd[1]: ignition-files.service: Deactivated successfully. Sep 4 00:04:37.312280 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 4 00:04:37.317572 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 4 00:04:37.320090 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 4 00:04:37.321231 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 4 00:04:37.321402 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 00:04:37.323548 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 4 00:04:37.323672 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 00:04:37.330172 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 4 00:04:37.338711 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 4 00:04:37.355646 ignition[1089]: INFO : Ignition 2.21.0 Sep 4 00:04:37.355646 ignition[1089]: INFO : Stage: umount Sep 4 00:04:37.357836 ignition[1089]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 00:04:37.357836 ignition[1089]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 00:04:37.360486 ignition[1089]: INFO : umount: umount passed Sep 4 00:04:37.360486 ignition[1089]: INFO : Ignition finished successfully Sep 4 00:04:37.361070 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 4 00:04:37.361713 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 4 00:04:37.361830 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 4 00:04:37.363944 systemd[1]: Stopped target network.target - Network. Sep 4 00:04:37.364391 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 4 00:04:37.364448 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 4 00:04:37.366100 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 4 00:04:37.366148 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 4 00:04:37.366416 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 4 00:04:37.366477 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 4 00:04:37.366763 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 4 00:04:37.366802 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 4 00:04:37.371247 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 4 00:04:37.373804 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 4 00:04:37.379350 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 4 00:04:37.379545 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 4 00:04:37.384116 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 4 00:04:37.384503 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 4 00:04:37.384677 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 4 00:04:37.386002 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 4 00:04:37.386119 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 4 00:04:37.389228 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 4 00:04:37.389293 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 00:04:37.394918 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 4 00:04:37.400692 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 4 00:04:37.400858 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 4 00:04:37.404873 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 4 00:04:37.405101 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 4 00:04:37.418944 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 4 00:04:37.418998 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 4 00:04:37.422207 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 4 00:04:37.423273 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 4 00:04:37.423339 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 00:04:37.425122 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 00:04:37.425181 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 00:04:37.431770 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 4 00:04:37.431867 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 4 00:04:37.436837 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 00:04:37.441085 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 4 00:04:37.458323 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 4 00:04:37.461647 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 00:04:37.470526 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 4 00:04:37.470586 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 4 00:04:37.472864 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 4 00:04:37.472901 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 00:04:37.475023 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 4 00:04:37.475074 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 4 00:04:37.477589 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 4 00:04:37.477638 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 4 00:04:37.478389 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 00:04:37.478437 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 00:04:37.486351 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 4 00:04:37.488747 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 4 00:04:37.488802 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 4 00:04:37.492161 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 4 00:04:37.492219 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 00:04:37.495709 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 4 00:04:37.495755 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 00:04:37.499132 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 4 00:04:37.499177 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 00:04:37.500720 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 00:04:37.500767 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 00:04:37.505624 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 4 00:04:37.505731 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 4 00:04:37.509275 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 4 00:04:37.509388 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 4 00:04:37.511329 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 4 00:04:37.513016 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 4 00:04:37.545901 systemd[1]: Switching root. Sep 4 00:04:37.591368 systemd-journald[220]: Journal stopped Sep 4 00:04:39.515162 systemd-journald[220]: Received SIGTERM from PID 1 (systemd). Sep 4 00:04:39.515248 kernel: SELinux: policy capability network_peer_controls=1 Sep 4 00:04:39.515268 kernel: SELinux: policy capability open_perms=1 Sep 4 00:04:39.515293 kernel: SELinux: policy capability extended_socket_class=1 Sep 4 00:04:39.515315 kernel: SELinux: policy capability always_check_network=0 Sep 4 00:04:39.515330 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 4 00:04:39.515345 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 4 00:04:39.515361 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 4 00:04:39.515376 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 4 00:04:39.515391 kernel: SELinux: policy capability userspace_initial_context=0 Sep 4 00:04:39.515407 kernel: audit: type=1403 audit(1756944278.534:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 4 00:04:39.515429 systemd[1]: Successfully loaded SELinux policy in 53.298ms. Sep 4 00:04:39.515460 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 17.295ms. Sep 4 00:04:39.515494 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 4 00:04:39.515512 systemd[1]: Detected virtualization kvm. Sep 4 00:04:39.515535 systemd[1]: Detected architecture x86-64. Sep 4 00:04:39.515551 systemd[1]: Detected first boot. Sep 4 00:04:39.515568 systemd[1]: Initializing machine ID from VM UUID. Sep 4 00:04:39.515595 zram_generator::config[1136]: No configuration found. Sep 4 00:04:39.515613 kernel: Guest personality initialized and is inactive Sep 4 00:04:39.515633 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 4 00:04:39.515648 kernel: Initialized host personality Sep 4 00:04:39.515663 kernel: NET: Registered PF_VSOCK protocol family Sep 4 00:04:39.515679 systemd[1]: Populated /etc with preset unit settings. Sep 4 00:04:39.515703 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 4 00:04:39.515720 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 4 00:04:39.515737 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 4 00:04:39.515754 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 4 00:04:39.515771 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 4 00:04:39.515790 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 4 00:04:39.515807 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 4 00:04:39.515823 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 4 00:04:39.515839 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 4 00:04:39.515856 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 4 00:04:39.515884 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 4 00:04:39.515901 systemd[1]: Created slice user.slice - User and Session Slice. Sep 4 00:04:39.515921 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 00:04:39.515945 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 00:04:39.515966 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 4 00:04:39.515997 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 4 00:04:39.516018 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 4 00:04:39.516039 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 00:04:39.516058 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 4 00:04:39.516075 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 00:04:39.516092 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 00:04:39.516112 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 4 00:04:39.516128 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 4 00:04:39.516144 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 4 00:04:39.516160 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 4 00:04:39.516177 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 00:04:39.516193 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 00:04:39.516209 systemd[1]: Reached target slices.target - Slice Units. Sep 4 00:04:39.516225 systemd[1]: Reached target swap.target - Swaps. Sep 4 00:04:39.516241 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 4 00:04:39.516261 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 4 00:04:39.516277 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 4 00:04:39.516293 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 00:04:39.516310 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 00:04:39.516326 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 00:04:39.516342 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 4 00:04:39.516368 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 4 00:04:39.516384 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 4 00:04:39.516401 systemd[1]: Mounting media.mount - External Media Directory... Sep 4 00:04:39.516420 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:39.516437 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 4 00:04:39.516453 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 4 00:04:39.516485 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 4 00:04:39.516502 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 4 00:04:39.516528 systemd[1]: Reached target machines.target - Containers. Sep 4 00:04:39.516545 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 4 00:04:39.516562 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 00:04:39.516585 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 00:04:39.516602 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 4 00:04:39.516618 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 00:04:39.516635 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 00:04:39.516651 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 00:04:39.516668 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 4 00:04:39.516684 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 00:04:39.516701 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 4 00:04:39.516718 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 4 00:04:39.516737 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 4 00:04:39.516763 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 4 00:04:39.516779 systemd[1]: Stopped systemd-fsck-usr.service. Sep 4 00:04:39.516796 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 4 00:04:39.516813 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 00:04:39.516829 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 00:04:39.516848 kernel: loop: module loaded Sep 4 00:04:39.516863 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 4 00:04:39.516905 systemd-journald[1200]: Collecting audit messages is disabled. Sep 4 00:04:39.516934 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 4 00:04:39.516952 systemd-journald[1200]: Journal started Sep 4 00:04:39.516985 systemd-journald[1200]: Runtime Journal (/run/log/journal/f99aefdb731b4c16a34c05e00bee0899) is 6M, max 48.2M, 42.2M free. Sep 4 00:04:39.203831 systemd[1]: Queued start job for default target multi-user.target. Sep 4 00:04:39.229854 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 4 00:04:39.230401 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 4 00:04:39.524514 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 4 00:04:39.528752 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 00:04:39.528794 kernel: fuse: init (API version 7.41) Sep 4 00:04:39.529964 systemd[1]: verity-setup.service: Deactivated successfully. Sep 4 00:04:39.530634 systemd[1]: Stopped verity-setup.service. Sep 4 00:04:39.535504 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:39.540022 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 00:04:39.540910 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 4 00:04:39.542079 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 4 00:04:39.543458 systemd[1]: Mounted media.mount - External Media Directory. Sep 4 00:04:39.544596 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 4 00:04:39.545918 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 4 00:04:39.547258 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 4 00:04:39.548628 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 00:04:39.550149 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 4 00:04:39.550422 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 4 00:04:39.551933 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 00:04:39.552186 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 00:04:39.553783 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 00:04:39.553993 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 00:04:39.555528 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 4 00:04:39.555825 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 4 00:04:39.557308 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 00:04:39.557609 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 00:04:39.559093 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 00:04:39.560646 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 4 00:04:39.562360 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 4 00:04:39.564453 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 4 00:04:39.566511 kernel: ACPI: bus type drm_connector registered Sep 4 00:04:39.568171 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 00:04:39.568670 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 00:04:39.581878 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 4 00:04:39.584993 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 4 00:04:39.593903 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 4 00:04:39.595429 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 4 00:04:39.595694 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 00:04:39.598508 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 4 00:04:39.604181 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 4 00:04:39.626998 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 00:04:39.629238 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 4 00:04:39.631864 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 4 00:04:39.633404 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 00:04:39.634667 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 4 00:04:39.666596 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 00:04:39.672631 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 00:04:39.701638 systemd-journald[1200]: Time spent on flushing to /var/log/journal/f99aefdb731b4c16a34c05e00bee0899 is 13.377ms for 1044 entries. Sep 4 00:04:39.701638 systemd-journald[1200]: System Journal (/var/log/journal/f99aefdb731b4c16a34c05e00bee0899) is 8M, max 195.6M, 187.6M free. Sep 4 00:04:41.181840 kernel: loop0: detected capacity change from 0 to 113872 Sep 4 00:04:41.181992 systemd-journald[1200]: Received client request to flush runtime journal. Sep 4 00:04:41.182066 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 4 00:04:41.182101 kernel: loop1: detected capacity change from 0 to 146240 Sep 4 00:04:41.182139 kernel: loop2: detected capacity change from 0 to 224512 Sep 4 00:04:41.182162 kernel: loop3: detected capacity change from 0 to 113872 Sep 4 00:04:41.182194 kernel: loop4: detected capacity change from 0 to 146240 Sep 4 00:04:41.182232 kernel: loop5: detected capacity change from 0 to 224512 Sep 4 00:04:41.182268 zram_generator::config[1293]: No configuration found. Sep 4 00:04:41.182303 zram_generator::config[1359]: No configuration found. Sep 4 00:04:39.677909 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 4 00:04:41.184397 ldconfig[1235]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 4 00:04:39.680711 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 00:04:39.686827 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 00:04:39.692051 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 4 00:04:39.694084 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 4 00:04:39.803086 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 00:04:39.805980 systemd-tmpfiles[1241]: ACLs are not supported, ignoring. Sep 4 00:04:39.805996 systemd-tmpfiles[1241]: ACLs are not supported, ignoring. Sep 4 00:04:39.813132 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 00:04:39.922552 (sd-merge)[1259]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 4 00:04:39.923147 (sd-merge)[1259]: Merged extensions into '/usr'. Sep 4 00:04:39.928347 systemd[1]: Reload requested from client PID 1240 ('systemd-sysext') (unit systemd-sysext.service)... Sep 4 00:04:39.928359 systemd[1]: Reloading... Sep 4 00:04:40.404826 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 00:04:40.493672 systemd[1]: Reloading finished in 563 ms. Sep 4 00:04:40.524346 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 4 00:04:40.526149 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 4 00:04:40.530289 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 4 00:04:40.550927 systemd[1]: Starting ensure-sysext.service... Sep 4 00:04:40.566832 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 4 00:04:40.703415 systemd[1]: Reload requested from client PID 1331 ('systemctl') (unit ensure-sysext.service)... Sep 4 00:04:40.703428 systemd[1]: Reloading... Sep 4 00:04:40.994676 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 00:04:41.076711 systemd[1]: Reloading finished in 372 ms. Sep 4 00:04:41.128093 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:41.128272 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 00:04:41.129718 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 00:04:41.132167 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 00:04:41.149148 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 00:04:41.150365 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 00:04:41.150585 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 4 00:04:41.150824 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:41.152274 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 00:04:41.152628 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 00:04:41.154301 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 00:04:41.154580 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 00:04:41.156925 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 00:04:41.158745 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 00:04:41.158953 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 00:04:41.160408 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:41.160638 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 00:04:41.161976 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 00:04:41.164242 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 00:04:41.165608 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 00:04:41.165717 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 4 00:04:41.165834 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:41.169296 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:41.169548 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 00:04:41.187800 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 00:04:41.190434 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 00:04:41.191855 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 00:04:41.191995 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 4 00:04:41.192163 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 00:04:41.203799 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 4 00:04:41.206517 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 4 00:04:41.208428 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 4 00:04:41.261721 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 00:04:41.261976 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 00:04:41.263673 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 00:04:41.263896 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 00:04:41.265655 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 00:04:41.265940 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 00:04:41.267691 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 00:04:41.267896 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 00:04:41.274336 systemd[1]: Finished ensure-sysext.service. Sep 4 00:04:41.284121 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 00:04:41.284203 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 00:04:41.285561 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 4 00:04:41.703044 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 4 00:04:41.710873 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 00:04:41.719686 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 00:04:41.771794 systemd-tmpfiles[1421]: ACLs are not supported, ignoring. Sep 4 00:04:41.773515 systemd-tmpfiles[1421]: ACLs are not supported, ignoring. Sep 4 00:04:41.776444 systemd-tmpfiles[1422]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 4 00:04:41.778550 systemd-tmpfiles[1422]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 4 00:04:41.778868 systemd-tmpfiles[1422]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 4 00:04:41.779106 systemd-tmpfiles[1422]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 4 00:04:41.780178 systemd-tmpfiles[1422]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 4 00:04:41.782620 systemd-tmpfiles[1422]: ACLs are not supported, ignoring. Sep 4 00:04:41.782713 systemd-tmpfiles[1422]: ACLs are not supported, ignoring. Sep 4 00:04:41.795153 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 00:04:41.930233 systemd-tmpfiles[1422]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 00:04:41.934542 systemd-tmpfiles[1422]: Skipping /boot Sep 4 00:04:41.967480 systemd-tmpfiles[1422]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 00:04:41.968018 systemd-tmpfiles[1422]: Skipping /boot Sep 4 00:04:42.129915 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 4 00:04:42.134236 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 4 00:04:42.136463 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 4 00:04:42.153708 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 00:04:42.159624 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 00:04:42.188733 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 4 00:04:42.213614 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 4 00:04:42.220705 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 4 00:04:42.232403 systemd-udevd[1428]: Using default interface naming scheme 'v255'. Sep 4 00:04:42.239282 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 00:04:42.243888 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 4 00:04:42.248392 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 4 00:04:42.256367 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 4 00:04:42.290064 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 4 00:04:42.311624 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 4 00:04:42.317214 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 4 00:04:42.331271 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 00:04:42.346832 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 00:04:42.363952 augenrules[1459]: No rules Sep 4 00:04:42.396073 systemd[1]: audit-rules.service: Deactivated successfully. Sep 4 00:04:42.396754 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 4 00:04:42.406164 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 4 00:04:42.423947 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 4 00:04:42.511769 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 4 00:04:42.517060 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 4 00:04:42.564636 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 4 00:04:42.686504 kernel: mousedev: PS/2 mouse device common for all mice Sep 4 00:04:42.704502 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 4 00:04:42.708753 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 4 00:04:42.740321 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 4 00:04:42.750564 kernel: ACPI: button: Power Button [PWRF] Sep 4 00:04:42.754405 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 4 00:04:42.758713 systemd-resolved[1434]: Positive Trust Anchors: Sep 4 00:04:42.759230 systemd-resolved[1434]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 00:04:42.759351 systemd-resolved[1434]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 00:04:42.767917 systemd-resolved[1434]: Defaulting to hostname 'linux'. Sep 4 00:04:42.772441 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 00:04:42.774177 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 4 00:04:42.775854 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 00:04:42.777622 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 00:04:42.778897 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 4 00:04:42.779612 systemd-networkd[1461]: lo: Link UP Sep 4 00:04:42.779617 systemd-networkd[1461]: lo: Gained carrier Sep 4 00:04:42.781602 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 4 00:04:42.784458 systemd-networkd[1461]: Enumeration completed Sep 4 00:04:42.785257 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Sep 4 00:04:42.786555 systemd-networkd[1461]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 00:04:42.786562 systemd-networkd[1461]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 00:04:42.787540 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 4 00:04:42.789256 systemd-networkd[1461]: eth0: Link UP Sep 4 00:04:42.789536 systemd-networkd[1461]: eth0: Gained carrier Sep 4 00:04:42.789570 systemd-networkd[1461]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 00:04:42.797977 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 4 00:04:42.798041 systemd[1]: Reached target paths.target - Path Units. Sep 4 00:04:42.799245 systemd[1]: Reached target time-set.target - System Time Set. Sep 4 00:04:42.800732 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 4 00:04:42.802156 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 4 00:04:42.803965 systemd[1]: Reached target timers.target - Timer Units. Sep 4 00:04:42.808989 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 4 00:04:42.815336 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 4 00:04:42.830852 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 4 00:04:42.832759 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 4 00:04:42.834258 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 4 00:04:42.839762 systemd-networkd[1461]: eth0: DHCPv4 address 10.0.0.91/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 4 00:04:42.840962 systemd-timesyncd[1437]: Network configuration changed, trying to establish connection. Sep 4 00:04:43.899090 systemd-timesyncd[1437]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 4 00:04:43.899141 systemd-timesyncd[1437]: Initial clock synchronization to Thu 2025-09-04 00:04:43.898933 UTC. Sep 4 00:04:43.899187 systemd-resolved[1434]: Clock change detected. Flushing caches. Sep 4 00:04:44.059800 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 4 00:04:44.066048 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 4 00:04:44.075567 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 00:04:44.077407 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 4 00:04:44.091071 systemd[1]: Reached target network.target - Network. Sep 4 00:04:44.092457 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 00:04:44.093880 systemd[1]: Reached target basic.target - Basic System. Sep 4 00:04:44.095497 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 4 00:04:44.095636 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 4 00:04:44.099086 systemd[1]: Starting containerd.service - containerd container runtime... Sep 4 00:04:44.108106 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 4 00:04:44.120954 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 4 00:04:44.125855 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 4 00:04:44.130624 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 4 00:04:44.132009 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 4 00:04:44.146940 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Sep 4 00:04:44.152526 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 4 00:04:44.164810 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 4 00:04:44.182892 jq[1538]: false Sep 4 00:04:44.189074 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 4 00:04:44.190673 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Refreshing passwd entry cache Sep 4 00:04:44.190684 oslogin_cache_refresh[1540]: Refreshing passwd entry cache Sep 4 00:04:44.200759 extend-filesystems[1539]: Found /dev/vda6 Sep 4 00:04:44.208261 extend-filesystems[1539]: Found /dev/vda9 Sep 4 00:04:44.208261 extend-filesystems[1539]: Checking size of /dev/vda9 Sep 4 00:04:44.202822 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 4 00:04:44.201724 oslogin_cache_refresh[1540]: Failure getting users, quitting Sep 4 00:04:44.214731 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Failure getting users, quitting Sep 4 00:04:44.214731 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 4 00:04:44.214731 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Refreshing group entry cache Sep 4 00:04:44.214731 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Failure getting groups, quitting Sep 4 00:04:44.214731 google_oslogin_nss_cache[1540]: oslogin_cache_refresh[1540]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 4 00:04:44.201748 oslogin_cache_refresh[1540]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 4 00:04:44.201819 oslogin_cache_refresh[1540]: Refreshing group entry cache Sep 4 00:04:44.209669 oslogin_cache_refresh[1540]: Failure getting groups, quitting Sep 4 00:04:44.209686 oslogin_cache_refresh[1540]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 4 00:04:44.238628 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Sep 4 00:04:44.239146 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 4 00:04:44.243211 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 4 00:04:44.262672 extend-filesystems[1539]: Resized partition /dev/vda9 Sep 4 00:04:44.286265 extend-filesystems[1559]: resize2fs 1.47.2 (1-Jan-2025) Sep 4 00:04:44.289013 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 4 00:04:44.307577 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 4 00:04:44.342973 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 4 00:04:44.350898 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 4 00:04:44.353645 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 4 00:04:44.354442 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 4 00:04:44.356892 systemd[1]: Starting update-engine.service - Update Engine... Sep 4 00:04:44.360787 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 4 00:04:44.373717 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 4 00:04:44.375831 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 4 00:04:44.376162 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 4 00:04:44.377211 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Sep 4 00:04:44.377574 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Sep 4 00:04:44.378244 jq[1569]: true Sep 4 00:04:44.379716 systemd[1]: motdgen.service: Deactivated successfully. Sep 4 00:04:44.380233 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 4 00:04:44.386726 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 4 00:04:44.387088 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 4 00:04:44.418213 jq[1572]: true Sep 4 00:04:44.429758 update_engine[1568]: I20250904 00:04:44.429626 1568 main.cc:92] Flatcar Update Engine starting Sep 4 00:04:44.435872 (ntainerd)[1582]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 4 00:04:44.443112 tar[1571]: linux-amd64/LICENSE Sep 4 00:04:44.447704 tar[1571]: linux-amd64/helm Sep 4 00:04:44.466665 systemd-logind[1561]: Watching system buttons on /dev/input/event2 (Power Button) Sep 4 00:04:44.466697 systemd-logind[1561]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 4 00:04:44.467227 systemd-logind[1561]: New seat seat0. Sep 4 00:04:44.473932 systemd[1]: Started systemd-logind.service - User Login Management. Sep 4 00:04:44.476665 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 4 00:04:44.499714 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 00:04:44.500586 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 4 00:04:44.506146 kernel: kvm_amd: TSC scaling supported Sep 4 00:04:44.506220 kernel: kvm_amd: Nested Virtualization enabled Sep 4 00:04:44.506238 kernel: kvm_amd: Nested Paging enabled Sep 4 00:04:44.508246 kernel: kvm_amd: LBR virtualization supported Sep 4 00:04:44.508279 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 4 00:04:44.545276 kernel: kvm_amd: Virtual GIF supported Sep 4 00:04:44.548457 extend-filesystems[1559]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 4 00:04:44.548457 extend-filesystems[1559]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 4 00:04:44.548457 extend-filesystems[1559]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 4 00:04:44.553871 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 4 00:04:44.650682 sshd_keygen[1564]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 4 00:04:44.650880 extend-filesystems[1539]: Resized filesystem in /dev/vda9 Sep 4 00:04:44.658278 dbus-daemon[1533]: [system] SELinux support is enabled Sep 4 00:04:44.663082 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 4 00:04:44.665022 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 4 00:04:44.667345 update_engine[1568]: I20250904 00:04:44.667158 1568 update_check_scheduler.cc:74] Next update check in 11m39s Sep 4 00:04:44.684124 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 4 00:04:44.851125 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 4 00:04:44.854337 dbus-daemon[1533]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 4 00:04:44.865476 bash[1601]: Updated "/home/core/.ssh/authorized_keys" Sep 4 00:04:44.865944 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 4 00:04:44.872896 systemd[1]: Started sshd@0-10.0.0.91:22-10.0.0.1:38366.service - OpenSSH per-connection server daemon (10.0.0.1:38366). Sep 4 00:04:44.879015 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 4 00:04:44.879068 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 4 00:04:44.882881 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 4 00:04:44.882914 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 4 00:04:45.002980 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 4 00:04:45.006579 systemd[1]: Started update-engine.service - Update Engine. Sep 4 00:04:45.011288 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 4 00:04:45.013942 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 4 00:04:45.017983 systemd[1]: issuegen.service: Deactivated successfully. Sep 4 00:04:45.018383 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 4 00:04:45.028842 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 4 00:04:45.126230 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 4 00:04:45.145996 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 4 00:04:45.156825 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 4 00:04:45.157174 systemd[1]: Reached target getty.target - Login Prompts. Sep 4 00:04:45.165576 kernel: EDAC MC: Ver: 3.0.0 Sep 4 00:04:45.173251 locksmithd[1621]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 4 00:04:45.208666 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 00:04:45.290242 sshd[1616]: Accepted publickey for core from 10.0.0.1 port 38366 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:45.295090 sshd-session[1616]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:45.312090 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 4 00:04:45.315020 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 4 00:04:45.331442 systemd-logind[1561]: New session 1 of user core. Sep 4 00:04:45.364530 containerd[1582]: time="2025-09-04T00:04:45Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 4 00:04:45.365588 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 4 00:04:45.367566 containerd[1582]: time="2025-09-04T00:04:45.366505205Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 Sep 4 00:04:45.376751 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 4 00:04:45.390907 containerd[1582]: time="2025-09-04T00:04:45.390849317Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.632µs" Sep 4 00:04:45.391077 containerd[1582]: time="2025-09-04T00:04:45.391054542Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 4 00:04:45.391247 containerd[1582]: time="2025-09-04T00:04:45.391225884Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.391537719Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.391612289Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.391647775Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.391741922Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.391759294Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.392149206Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.392169193Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.392192046Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.392205561Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.392338300Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 4 00:04:45.393462 containerd[1582]: time="2025-09-04T00:04:45.392701271Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 4 00:04:45.397507 containerd[1582]: time="2025-09-04T00:04:45.392740885Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 4 00:04:45.397507 containerd[1582]: time="2025-09-04T00:04:45.392755323Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 4 00:04:45.397507 containerd[1582]: time="2025-09-04T00:04:45.392801910Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 4 00:04:45.397507 containerd[1582]: time="2025-09-04T00:04:45.393140655Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 4 00:04:45.397507 containerd[1582]: time="2025-09-04T00:04:45.393233339Z" level=info msg="metadata content store policy set" policy=shared Sep 4 00:04:45.420195 (systemd)[1644]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.420217834Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421138451Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421164801Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421191811Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421210055Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421223851Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421243969Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421258987Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421275388Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421290096Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421302068Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 4 00:04:45.421447 containerd[1582]: time="2025-09-04T00:04:45.421319100Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421512573Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421578276Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421599847Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421614865Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421629232Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421642657Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421656272Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 4 00:04:45.421747 containerd[1582]: time="2025-09-04T00:04:45.421668065Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 4 00:04:45.421927 containerd[1582]: time="2025-09-04T00:04:45.421915849Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 4 00:04:45.421958 containerd[1582]: time="2025-09-04T00:04:45.421934454Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 4 00:04:45.421958 containerd[1582]: time="2025-09-04T00:04:45.421948601Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 4 00:04:45.423366 containerd[1582]: time="2025-09-04T00:04:45.422038790Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 4 00:04:45.423366 containerd[1582]: time="2025-09-04T00:04:45.422064538Z" level=info msg="Start snapshots syncer" Sep 4 00:04:45.423366 containerd[1582]: time="2025-09-04T00:04:45.422097981Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 4 00:04:45.424572 containerd[1582]: time="2025-09-04T00:04:45.422430274Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 4 00:04:45.424739 containerd[1582]: time="2025-09-04T00:04:45.424579636Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 4 00:04:45.424955 containerd[1582]: time="2025-09-04T00:04:45.424914744Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 4 00:04:45.425140 containerd[1582]: time="2025-09-04T00:04:45.425105272Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 4 00:04:45.425181 containerd[1582]: time="2025-09-04T00:04:45.425140548Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 4 00:04:45.425181 containerd[1582]: time="2025-09-04T00:04:45.425170153Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 4 00:04:45.425245 containerd[1582]: time="2025-09-04T00:04:45.425186294Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 4 00:04:45.425245 containerd[1582]: time="2025-09-04T00:04:45.425202725Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 4 00:04:45.425245 containerd[1582]: time="2025-09-04T00:04:45.425214968Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 4 00:04:45.425245 containerd[1582]: time="2025-09-04T00:04:45.425229835Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 4 00:04:45.425338 containerd[1582]: time="2025-09-04T00:04:45.425265873Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 4 00:04:45.425338 containerd[1582]: time="2025-09-04T00:04:45.425280991Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 4 00:04:45.425338 containerd[1582]: time="2025-09-04T00:04:45.425301430Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 4 00:04:45.425418 containerd[1582]: time="2025-09-04T00:04:45.425356192Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 4 00:04:45.425418 containerd[1582]: time="2025-09-04T00:04:45.425376601Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 4 00:04:45.425750 containerd[1582]: time="2025-09-04T00:04:45.425388613Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 4 00:04:45.425808 containerd[1582]: time="2025-09-04T00:04:45.425743769Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 4 00:04:45.425808 containerd[1582]: time="2025-09-04T00:04:45.425760571Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 4 00:04:45.425808 containerd[1582]: time="2025-09-04T00:04:45.425775289Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 4 00:04:45.425808 containerd[1582]: time="2025-09-04T00:04:45.425790086Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 4 00:04:45.425907 containerd[1582]: time="2025-09-04T00:04:45.425815825Z" level=info msg="runtime interface created" Sep 4 00:04:45.425907 containerd[1582]: time="2025-09-04T00:04:45.425824150Z" level=info msg="created NRI interface" Sep 4 00:04:45.425907 containerd[1582]: time="2025-09-04T00:04:45.425836032Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 4 00:04:45.425907 containerd[1582]: time="2025-09-04T00:04:45.425851812Z" level=info msg="Connect containerd service" Sep 4 00:04:45.425907 containerd[1582]: time="2025-09-04T00:04:45.425881428Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 4 00:04:45.427622 containerd[1582]: time="2025-09-04T00:04:45.426996309Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 00:04:45.427133 systemd-logind[1561]: New session c1 of user core. Sep 4 00:04:45.661662 containerd[1582]: time="2025-09-04T00:04:45.659450677Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 4 00:04:45.661662 containerd[1582]: time="2025-09-04T00:04:45.659566083Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 4 00:04:45.661914 containerd[1582]: time="2025-09-04T00:04:45.659448753Z" level=info msg="Start subscribing containerd event" Sep 4 00:04:45.662104 containerd[1582]: time="2025-09-04T00:04:45.662043640Z" level=info msg="Start recovering state" Sep 4 00:04:45.662904 containerd[1582]: time="2025-09-04T00:04:45.662852257Z" level=info msg="Start event monitor" Sep 4 00:04:45.662904 containerd[1582]: time="2025-09-04T00:04:45.662881813Z" level=info msg="Start cni network conf syncer for default" Sep 4 00:04:45.662904 containerd[1582]: time="2025-09-04T00:04:45.662898033Z" level=info msg="Start streaming server" Sep 4 00:04:45.663067 containerd[1582]: time="2025-09-04T00:04:45.662913432Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 4 00:04:45.663067 containerd[1582]: time="2025-09-04T00:04:45.662922519Z" level=info msg="runtime interface starting up..." Sep 4 00:04:45.663067 containerd[1582]: time="2025-09-04T00:04:45.662930243Z" level=info msg="starting plugins..." Sep 4 00:04:45.663067 containerd[1582]: time="2025-09-04T00:04:45.662950111Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 4 00:04:45.663525 containerd[1582]: time="2025-09-04T00:04:45.663117374Z" level=info msg="containerd successfully booted in 0.299782s" Sep 4 00:04:45.663265 systemd[1]: Started containerd.service - containerd container runtime. Sep 4 00:04:45.734802 tar[1571]: linux-amd64/README.md Sep 4 00:04:45.759476 systemd[1644]: Queued start job for default target default.target. Sep 4 00:04:45.763049 systemd[1644]: Created slice app.slice - User Application Slice. Sep 4 00:04:45.763084 systemd[1644]: Reached target paths.target - Paths. Sep 4 00:04:45.763150 systemd[1644]: Reached target timers.target - Timers. Sep 4 00:04:45.771500 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 4 00:04:45.778348 systemd[1644]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 4 00:04:45.791872 systemd[1644]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 4 00:04:45.792052 systemd[1644]: Reached target sockets.target - Sockets. Sep 4 00:04:45.792117 systemd[1644]: Reached target basic.target - Basic System. Sep 4 00:04:45.792172 systemd[1644]: Reached target default.target - Main User Target. Sep 4 00:04:45.792217 systemd[1644]: Startup finished in 355ms. Sep 4 00:04:45.793108 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 4 00:04:45.809950 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 4 00:04:45.872856 systemd-networkd[1461]: eth0: Gained IPv6LL Sep 4 00:04:45.907611 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 4 00:04:45.927350 systemd[1]: Reached target network-online.target - Network is Online. Sep 4 00:04:45.933815 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 4 00:04:45.939217 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:04:45.964758 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 4 00:04:45.973937 systemd[1]: Started sshd@1-10.0.0.91:22-10.0.0.1:38372.service - OpenSSH per-connection server daemon (10.0.0.1:38372). Sep 4 00:04:46.015958 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 4 00:04:46.016348 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 4 00:04:46.022019 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 4 00:04:46.029509 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 4 00:04:46.049291 sshd[1678]: Accepted publickey for core from 10.0.0.1 port 38372 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:46.051659 sshd-session[1678]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:46.058838 systemd-logind[1561]: New session 2 of user core. Sep 4 00:04:46.072852 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 4 00:04:46.144656 sshd[1694]: Connection closed by 10.0.0.1 port 38372 Sep 4 00:04:46.145644 sshd-session[1678]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:46.167913 systemd[1]: sshd@1-10.0.0.91:22-10.0.0.1:38372.service: Deactivated successfully. Sep 4 00:04:46.173525 systemd[1]: session-2.scope: Deactivated successfully. Sep 4 00:04:46.175213 systemd-logind[1561]: Session 2 logged out. Waiting for processes to exit. Sep 4 00:04:46.181830 systemd[1]: Started sshd@2-10.0.0.91:22-10.0.0.1:38378.service - OpenSSH per-connection server daemon (10.0.0.1:38378). Sep 4 00:04:46.187850 systemd-logind[1561]: Removed session 2. Sep 4 00:04:46.258435 sshd[1700]: Accepted publickey for core from 10.0.0.1 port 38378 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:46.260966 sshd-session[1700]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:46.276764 systemd-logind[1561]: New session 3 of user core. Sep 4 00:04:46.300480 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 4 00:04:46.379950 sshd[1702]: Connection closed by 10.0.0.1 port 38378 Sep 4 00:04:46.380891 sshd-session[1700]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:46.395952 systemd[1]: sshd@2-10.0.0.91:22-10.0.0.1:38378.service: Deactivated successfully. Sep 4 00:04:46.400271 systemd[1]: session-3.scope: Deactivated successfully. Sep 4 00:04:46.404506 systemd-logind[1561]: Session 3 logged out. Waiting for processes to exit. Sep 4 00:04:46.409973 systemd-logind[1561]: Removed session 3. Sep 4 00:04:47.676711 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:04:47.679341 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 4 00:04:47.682922 systemd[1]: Startup finished in 3.653s (kernel) + 9.812s (initrd) + 8.142s (userspace) = 21.608s. Sep 4 00:04:47.701931 (kubelet)[1712]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 00:04:48.700942 kubelet[1712]: E0904 00:04:48.700291 1712 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 00:04:48.708844 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 00:04:48.709905 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 00:04:48.712931 systemd[1]: kubelet.service: Consumed 1.418s CPU time, 266.7M memory peak. Sep 4 00:04:56.417681 systemd[1]: Started sshd@3-10.0.0.91:22-10.0.0.1:42630.service - OpenSSH per-connection server daemon (10.0.0.1:42630). Sep 4 00:04:56.532630 sshd[1726]: Accepted publickey for core from 10.0.0.1 port 42630 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:56.534854 sshd-session[1726]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:56.583224 systemd-logind[1561]: New session 4 of user core. Sep 4 00:04:56.601891 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 4 00:04:56.677720 sshd[1728]: Connection closed by 10.0.0.1 port 42630 Sep 4 00:04:56.681453 sshd-session[1726]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:56.704094 systemd[1]: sshd@3-10.0.0.91:22-10.0.0.1:42630.service: Deactivated successfully. Sep 4 00:04:56.712393 systemd[1]: session-4.scope: Deactivated successfully. Sep 4 00:04:56.718572 systemd-logind[1561]: Session 4 logged out. Waiting for processes to exit. Sep 4 00:04:56.727860 systemd-logind[1561]: Removed session 4. Sep 4 00:04:56.730462 systemd[1]: Started sshd@4-10.0.0.91:22-10.0.0.1:42640.service - OpenSSH per-connection server daemon (10.0.0.1:42640). Sep 4 00:04:56.815686 sshd[1734]: Accepted publickey for core from 10.0.0.1 port 42640 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:56.820849 sshd-session[1734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:56.836930 systemd-logind[1561]: New session 5 of user core. Sep 4 00:04:56.846946 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 4 00:04:56.918218 sshd[1736]: Connection closed by 10.0.0.1 port 42640 Sep 4 00:04:56.919007 sshd-session[1734]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:56.957265 systemd[1]: sshd@4-10.0.0.91:22-10.0.0.1:42640.service: Deactivated successfully. Sep 4 00:04:56.964072 systemd[1]: session-5.scope: Deactivated successfully. Sep 4 00:04:56.969960 systemd-logind[1561]: Session 5 logged out. Waiting for processes to exit. Sep 4 00:04:56.980528 systemd[1]: Started sshd@5-10.0.0.91:22-10.0.0.1:42642.service - OpenSSH per-connection server daemon (10.0.0.1:42642). Sep 4 00:04:56.982698 systemd-logind[1561]: Removed session 5. Sep 4 00:04:57.105340 sshd[1742]: Accepted publickey for core from 10.0.0.1 port 42642 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:57.108076 sshd-session[1742]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:57.136239 systemd-logind[1561]: New session 6 of user core. Sep 4 00:04:57.155337 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 4 00:04:57.239538 sshd[1744]: Connection closed by 10.0.0.1 port 42642 Sep 4 00:04:57.239159 sshd-session[1742]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:57.268939 systemd[1]: sshd@5-10.0.0.91:22-10.0.0.1:42642.service: Deactivated successfully. Sep 4 00:04:57.279676 systemd[1]: session-6.scope: Deactivated successfully. Sep 4 00:04:57.283798 systemd-logind[1561]: Session 6 logged out. Waiting for processes to exit. Sep 4 00:04:57.286010 systemd[1]: Started sshd@6-10.0.0.91:22-10.0.0.1:42646.service - OpenSSH per-connection server daemon (10.0.0.1:42646). Sep 4 00:04:57.293023 systemd-logind[1561]: Removed session 6. Sep 4 00:04:57.383925 sshd[1750]: Accepted publickey for core from 10.0.0.1 port 42646 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:57.392482 sshd-session[1750]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:57.432218 systemd-logind[1561]: New session 7 of user core. Sep 4 00:04:57.441852 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 4 00:04:57.564527 sudo[1753]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 4 00:04:57.565286 sudo[1753]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 00:04:57.611148 sudo[1753]: pam_unix(sudo:session): session closed for user root Sep 4 00:04:57.618980 sshd[1752]: Connection closed by 10.0.0.1 port 42646 Sep 4 00:04:57.620090 sshd-session[1750]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:57.650784 systemd[1]: sshd@6-10.0.0.91:22-10.0.0.1:42646.service: Deactivated successfully. Sep 4 00:04:57.653848 systemd[1]: session-7.scope: Deactivated successfully. Sep 4 00:04:57.657634 systemd-logind[1561]: Session 7 logged out. Waiting for processes to exit. Sep 4 00:04:57.681720 systemd[1]: Started sshd@7-10.0.0.91:22-10.0.0.1:42658.service - OpenSSH per-connection server daemon (10.0.0.1:42658). Sep 4 00:04:57.683222 systemd-logind[1561]: Removed session 7. Sep 4 00:04:57.772027 sshd[1759]: Accepted publickey for core from 10.0.0.1 port 42658 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:57.778986 sshd-session[1759]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:57.802889 systemd-logind[1561]: New session 8 of user core. Sep 4 00:04:57.813015 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 4 00:04:57.884570 sudo[1763]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 4 00:04:57.886632 sudo[1763]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 00:04:57.997096 sudo[1763]: pam_unix(sudo:session): session closed for user root Sep 4 00:04:58.010530 sudo[1762]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 4 00:04:58.013204 sudo[1762]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 00:04:58.048737 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 4 00:04:58.135483 augenrules[1785]: No rules Sep 4 00:04:58.139685 systemd[1]: audit-rules.service: Deactivated successfully. Sep 4 00:04:58.140093 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 4 00:04:58.143515 sudo[1762]: pam_unix(sudo:session): session closed for user root Sep 4 00:04:58.145714 sshd[1761]: Connection closed by 10.0.0.1 port 42658 Sep 4 00:04:58.146008 sshd-session[1759]: pam_unix(sshd:session): session closed for user core Sep 4 00:04:58.158945 systemd[1]: sshd@7-10.0.0.91:22-10.0.0.1:42658.service: Deactivated successfully. Sep 4 00:04:58.165500 systemd[1]: session-8.scope: Deactivated successfully. Sep 4 00:04:58.167040 systemd-logind[1561]: Session 8 logged out. Waiting for processes to exit. Sep 4 00:04:58.183524 systemd-logind[1561]: Removed session 8. Sep 4 00:04:58.189913 systemd[1]: Started sshd@8-10.0.0.91:22-10.0.0.1:42660.service - OpenSSH per-connection server daemon (10.0.0.1:42660). Sep 4 00:04:58.315173 sshd[1794]: Accepted publickey for core from 10.0.0.1 port 42660 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:04:58.317355 sshd-session[1794]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:04:58.355264 systemd-logind[1561]: New session 9 of user core. Sep 4 00:04:58.366500 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 4 00:04:58.446172 sudo[1797]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 4 00:04:58.450584 sudo[1797]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 00:04:58.958777 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 4 00:04:58.962654 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:04:59.336924 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 4 00:04:59.343570 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:04:59.360217 (kubelet)[1825]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 00:04:59.361451 (dockerd)[1823]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 4 00:04:59.499074 kubelet[1825]: E0904 00:04:59.496649 1825 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 00:04:59.514764 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 00:04:59.515197 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 00:04:59.515840 systemd[1]: kubelet.service: Consumed 353ms CPU time, 112.7M memory peak. Sep 4 00:04:59.965225 dockerd[1823]: time="2025-09-04T00:04:59.965125654Z" level=info msg="Starting up" Sep 4 00:04:59.968098 dockerd[1823]: time="2025-09-04T00:04:59.967225934Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 4 00:05:00.452731 dockerd[1823]: time="2025-09-04T00:05:00.451890706Z" level=info msg="Loading containers: start." Sep 4 00:05:00.485757 kernel: Initializing XFRM netlink socket Sep 4 00:05:01.262388 systemd-networkd[1461]: docker0: Link UP Sep 4 00:05:01.280489 dockerd[1823]: time="2025-09-04T00:05:01.280391948Z" level=info msg="Loading containers: done." Sep 4 00:05:01.301614 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2522632264-merged.mount: Deactivated successfully. Sep 4 00:05:01.315380 dockerd[1823]: time="2025-09-04T00:05:01.314622605Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 4 00:05:01.315380 dockerd[1823]: time="2025-09-04T00:05:01.314741598Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 Sep 4 00:05:01.315380 dockerd[1823]: time="2025-09-04T00:05:01.314983972Z" level=info msg="Initializing buildkit" Sep 4 00:05:01.398216 dockerd[1823]: time="2025-09-04T00:05:01.398081219Z" level=info msg="Completed buildkit initialization" Sep 4 00:05:01.411855 dockerd[1823]: time="2025-09-04T00:05:01.410915102Z" level=info msg="Daemon has completed initialization" Sep 4 00:05:01.411855 dockerd[1823]: time="2025-09-04T00:05:01.411567285Z" level=info msg="API listen on /run/docker.sock" Sep 4 00:05:01.411249 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 4 00:05:02.673976 containerd[1582]: time="2025-09-04T00:05:02.670659602Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.8\"" Sep 4 00:05:03.624216 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2337908811.mount: Deactivated successfully. Sep 4 00:05:04.864707 containerd[1582]: time="2025-09-04T00:05:04.864634051Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:04.865759 containerd[1582]: time="2025-09-04T00:05:04.865707765Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.8: active requests=0, bytes read=28800687" Sep 4 00:05:04.867691 containerd[1582]: time="2025-09-04T00:05:04.867626675Z" level=info msg="ImageCreate event name:\"sha256:0d4edaa48e2f940c934e0f7cfd5209fc85e65ab5e842b980f41263d1764661f1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:04.873978 containerd[1582]: time="2025-09-04T00:05:04.873782151Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6e1a2f9b24f69ee77d0c0edaf32b31fdbb5e1a613f4476272197e6e1e239050b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:04.874892 containerd[1582]: time="2025-09-04T00:05:04.874526377Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.8\" with image id \"sha256:0d4edaa48e2f940c934e0f7cfd5209fc85e65ab5e842b980f41263d1764661f1\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6e1a2f9b24f69ee77d0c0edaf32b31fdbb5e1a613f4476272197e6e1e239050b\", size \"28797487\" in 2.2013164s" Sep 4 00:05:04.874892 containerd[1582]: time="2025-09-04T00:05:04.874586650Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.8\" returns image reference \"sha256:0d4edaa48e2f940c934e0f7cfd5209fc85e65ab5e842b980f41263d1764661f1\"" Sep 4 00:05:04.875592 containerd[1582]: time="2025-09-04T00:05:04.875340524Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.8\"" Sep 4 00:05:06.105474 containerd[1582]: time="2025-09-04T00:05:06.105393782Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:06.106094 containerd[1582]: time="2025-09-04T00:05:06.106060573Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.8: active requests=0, bytes read=24784128" Sep 4 00:05:06.107267 containerd[1582]: time="2025-09-04T00:05:06.107225999Z" level=info msg="ImageCreate event name:\"sha256:b248d0b0c74ad8230e0bae0cbed477560e8a1e8c7ef5f29b7e75c1f273c8a091\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:06.110435 containerd[1582]: time="2025-09-04T00:05:06.110399852Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:8788ccd28ceed9e2e5f8fc31375ef5771df8ea6e518b362c9a06f3cc709cd6c7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:06.111370 containerd[1582]: time="2025-09-04T00:05:06.111323735Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.8\" with image id \"sha256:b248d0b0c74ad8230e0bae0cbed477560e8a1e8c7ef5f29b7e75c1f273c8a091\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:8788ccd28ceed9e2e5f8fc31375ef5771df8ea6e518b362c9a06f3cc709cd6c7\", size \"26387322\" in 1.235954086s" Sep 4 00:05:06.111370 containerd[1582]: time="2025-09-04T00:05:06.111362608Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.8\" returns image reference \"sha256:b248d0b0c74ad8230e0bae0cbed477560e8a1e8c7ef5f29b7e75c1f273c8a091\"" Sep 4 00:05:06.112005 containerd[1582]: time="2025-09-04T00:05:06.111813764Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.8\"" Sep 4 00:05:07.433129 containerd[1582]: time="2025-09-04T00:05:07.433039456Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:07.433929 containerd[1582]: time="2025-09-04T00:05:07.433849345Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.8: active requests=0, bytes read=19175036" Sep 4 00:05:07.435176 containerd[1582]: time="2025-09-04T00:05:07.435133113Z" level=info msg="ImageCreate event name:\"sha256:2ac266f06c9a5a3d0d20ae482dbccb54d3be454d5ca49f48b528bdf5bae3e908\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:07.438562 containerd[1582]: time="2025-09-04T00:05:07.438510818Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:43c58bcbd1c7812dd19f8bfa5ae11093ebefd28699453ce86fc710869e155cd4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:07.439858 containerd[1582]: time="2025-09-04T00:05:07.439788916Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.8\" with image id \"sha256:2ac266f06c9a5a3d0d20ae482dbccb54d3be454d5ca49f48b528bdf5bae3e908\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:43c58bcbd1c7812dd19f8bfa5ae11093ebefd28699453ce86fc710869e155cd4\", size \"20778248\" in 1.327930819s" Sep 4 00:05:07.439858 containerd[1582]: time="2025-09-04T00:05:07.439849419Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.8\" returns image reference \"sha256:2ac266f06c9a5a3d0d20ae482dbccb54d3be454d5ca49f48b528bdf5bae3e908\"" Sep 4 00:05:07.440507 containerd[1582]: time="2025-09-04T00:05:07.440472328Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.8\"" Sep 4 00:05:08.726614 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2044612886.mount: Deactivated successfully. Sep 4 00:05:09.673119 containerd[1582]: time="2025-09-04T00:05:09.673058742Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:09.674040 containerd[1582]: time="2025-09-04T00:05:09.673997382Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.8: active requests=0, bytes read=30897170" Sep 4 00:05:09.676416 containerd[1582]: time="2025-09-04T00:05:09.676362208Z" level=info msg="ImageCreate event name:\"sha256:d7b94972d43c5d6ce8088a8bcd08614a5ecf2bf04166232c688adcd0b8ed4b12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:09.678674 containerd[1582]: time="2025-09-04T00:05:09.678635091Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:adc1335b480ddd833aac3b0bd20f68ff0f3c3cf7a0bd337933b006d9f5cec40a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:09.679362 containerd[1582]: time="2025-09-04T00:05:09.679313203Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.8\" with image id \"sha256:d7b94972d43c5d6ce8088a8bcd08614a5ecf2bf04166232c688adcd0b8ed4b12\", repo tag \"registry.k8s.io/kube-proxy:v1.32.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:adc1335b480ddd833aac3b0bd20f68ff0f3c3cf7a0bd337933b006d9f5cec40a\", size \"30896189\" in 2.238804126s" Sep 4 00:05:09.679403 containerd[1582]: time="2025-09-04T00:05:09.679366213Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.8\" returns image reference \"sha256:d7b94972d43c5d6ce8088a8bcd08614a5ecf2bf04166232c688adcd0b8ed4b12\"" Sep 4 00:05:09.680034 containerd[1582]: time="2025-09-04T00:05:09.679987067Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 4 00:05:09.765706 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 4 00:05:09.767900 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:05:09.982785 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:05:09.987640 (kubelet)[2121]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 00:05:10.313092 kubelet[2121]: E0904 00:05:10.312948 2121 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 00:05:10.317768 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 00:05:10.318080 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 00:05:10.319284 systemd[1]: kubelet.service: Consumed 225ms CPU time, 110.9M memory peak. Sep 4 00:05:10.748048 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3356080415.mount: Deactivated successfully. Sep 4 00:05:11.591948 containerd[1582]: time="2025-09-04T00:05:11.591887065Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:11.592813 containerd[1582]: time="2025-09-04T00:05:11.592784849Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Sep 4 00:05:11.594021 containerd[1582]: time="2025-09-04T00:05:11.593998736Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:11.596615 containerd[1582]: time="2025-09-04T00:05:11.596581571Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:11.597613 containerd[1582]: time="2025-09-04T00:05:11.597575966Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.917552872s" Sep 4 00:05:11.597613 containerd[1582]: time="2025-09-04T00:05:11.597609930Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 4 00:05:11.598101 containerd[1582]: time="2025-09-04T00:05:11.598060515Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 4 00:05:12.176177 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2810194494.mount: Deactivated successfully. Sep 4 00:05:12.186266 containerd[1582]: time="2025-09-04T00:05:12.186222128Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 00:05:12.187139 containerd[1582]: time="2025-09-04T00:05:12.187116696Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 4 00:05:12.189165 containerd[1582]: time="2025-09-04T00:05:12.189113742Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 00:05:12.191189 containerd[1582]: time="2025-09-04T00:05:12.191144230Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 00:05:12.191964 containerd[1582]: time="2025-09-04T00:05:12.191923993Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 593.8215ms" Sep 4 00:05:12.191964 containerd[1582]: time="2025-09-04T00:05:12.191957927Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 4 00:05:12.192482 containerd[1582]: time="2025-09-04T00:05:12.192457333Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Sep 4 00:05:12.801882 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3231959595.mount: Deactivated successfully. Sep 4 00:05:15.338054 containerd[1582]: time="2025-09-04T00:05:15.337995834Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:15.339190 containerd[1582]: time="2025-09-04T00:05:15.339159176Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Sep 4 00:05:15.340595 containerd[1582]: time="2025-09-04T00:05:15.340563290Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:15.343205 containerd[1582]: time="2025-09-04T00:05:15.343157216Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:15.344389 containerd[1582]: time="2025-09-04T00:05:15.344342880Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 3.151856843s" Sep 4 00:05:15.344460 containerd[1582]: time="2025-09-04T00:05:15.344388966Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Sep 4 00:05:17.193824 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:05:17.194051 systemd[1]: kubelet.service: Consumed 225ms CPU time, 110.9M memory peak. Sep 4 00:05:17.196423 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:05:17.222971 systemd[1]: Reload requested from client PID 2272 ('systemctl') (unit session-9.scope)... Sep 4 00:05:17.222987 systemd[1]: Reloading... Sep 4 00:05:17.381567 zram_generator::config[2323]: No configuration found. Sep 4 00:05:17.776205 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 00:05:17.894712 systemd[1]: Reloading finished in 671 ms. Sep 4 00:05:17.970564 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 4 00:05:17.970696 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 4 00:05:17.971001 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:05:17.971045 systemd[1]: kubelet.service: Consumed 170ms CPU time, 98.3M memory peak. Sep 4 00:05:17.972610 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:05:18.146612 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:05:18.159912 (kubelet)[2362]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 00:05:18.203111 kubelet[2362]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 00:05:18.203111 kubelet[2362]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 4 00:05:18.203111 kubelet[2362]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 00:05:18.203500 kubelet[2362]: I0904 00:05:18.203164 2362 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 00:05:18.447758 kubelet[2362]: I0904 00:05:18.447657 2362 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 4 00:05:18.447758 kubelet[2362]: I0904 00:05:18.447685 2362 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 00:05:18.447980 kubelet[2362]: I0904 00:05:18.447954 2362 server.go:954] "Client rotation is on, will bootstrap in background" Sep 4 00:05:18.474126 kubelet[2362]: I0904 00:05:18.474067 2362 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 00:05:18.474269 kubelet[2362]: E0904 00:05:18.474230 2362 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.91:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:18.481111 kubelet[2362]: I0904 00:05:18.481080 2362 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 4 00:05:18.486073 kubelet[2362]: I0904 00:05:18.486041 2362 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 00:05:18.486378 kubelet[2362]: I0904 00:05:18.486332 2362 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 00:05:18.486520 kubelet[2362]: I0904 00:05:18.486358 2362 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 4 00:05:18.486520 kubelet[2362]: I0904 00:05:18.486521 2362 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 00:05:18.486717 kubelet[2362]: I0904 00:05:18.486528 2362 container_manager_linux.go:304] "Creating device plugin manager" Sep 4 00:05:18.486717 kubelet[2362]: I0904 00:05:18.486706 2362 state_mem.go:36] "Initialized new in-memory state store" Sep 4 00:05:18.489307 kubelet[2362]: I0904 00:05:18.489265 2362 kubelet.go:446] "Attempting to sync node with API server" Sep 4 00:05:18.489307 kubelet[2362]: I0904 00:05:18.489296 2362 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 00:05:18.489389 kubelet[2362]: I0904 00:05:18.489323 2362 kubelet.go:352] "Adding apiserver pod source" Sep 4 00:05:18.489389 kubelet[2362]: I0904 00:05:18.489334 2362 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 00:05:18.494274 kubelet[2362]: I0904 00:05:18.493739 2362 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Sep 4 00:05:18.494274 kubelet[2362]: I0904 00:05:18.494148 2362 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 4 00:05:18.494274 kubelet[2362]: W0904 00:05:18.494202 2362 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 4 00:05:18.495593 kubelet[2362]: W0904 00:05:18.495445 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:18.495593 kubelet[2362]: W0904 00:05:18.495446 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:18.495593 kubelet[2362]: E0904 00:05:18.495521 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:18.495593 kubelet[2362]: E0904 00:05:18.495525 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:18.496617 kubelet[2362]: I0904 00:05:18.496223 2362 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 4 00:05:18.496617 kubelet[2362]: I0904 00:05:18.496264 2362 server.go:1287] "Started kubelet" Sep 4 00:05:18.497858 kubelet[2362]: I0904 00:05:18.497819 2362 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 00:05:18.498954 kubelet[2362]: I0904 00:05:18.498897 2362 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 4 00:05:18.499243 kubelet[2362]: I0904 00:05:18.499214 2362 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 00:05:18.499306 kubelet[2362]: I0904 00:05:18.499284 2362 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 00:05:18.500278 kubelet[2362]: I0904 00:05:18.500245 2362 server.go:479] "Adding debug handlers to kubelet server" Sep 4 00:05:18.503483 kubelet[2362]: I0904 00:05:18.502982 2362 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 4 00:05:18.503885 kubelet[2362]: I0904 00:05:18.503859 2362 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 4 00:05:18.504090 kubelet[2362]: E0904 00:05:18.504068 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:18.504150 kubelet[2362]: I0904 00:05:18.504119 2362 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 4 00:05:18.504229 kubelet[2362]: I0904 00:05:18.504212 2362 reconciler.go:26] "Reconciler: start to sync state" Sep 4 00:05:18.505744 kubelet[2362]: W0904 00:05:18.504923 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:18.505744 kubelet[2362]: E0904 00:05:18.504976 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:18.505744 kubelet[2362]: E0904 00:05:18.505043 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.91:6443: connect: connection refused" interval="200ms" Sep 4 00:05:18.505977 kubelet[2362]: E0904 00:05:18.504481 2362 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.91:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.91:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1861eb8c8a1bebad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-04 00:05:18.496238509 +0000 UTC m=+0.331600055,LastTimestamp:2025-09-04 00:05:18.496238509 +0000 UTC m=+0.331600055,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 4 00:05:18.506328 kubelet[2362]: I0904 00:05:18.506308 2362 factory.go:221] Registration of the containerd container factory successfully Sep 4 00:05:18.506420 kubelet[2362]: I0904 00:05:18.506392 2362 factory.go:221] Registration of the systemd container factory successfully Sep 4 00:05:18.506539 kubelet[2362]: E0904 00:05:18.506359 2362 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 00:05:18.506690 kubelet[2362]: I0904 00:05:18.506672 2362 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 4 00:05:18.521726 kubelet[2362]: I0904 00:05:18.521683 2362 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 4 00:05:18.521726 kubelet[2362]: I0904 00:05:18.521708 2362 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 4 00:05:18.521726 kubelet[2362]: I0904 00:05:18.521725 2362 state_mem.go:36] "Initialized new in-memory state store" Sep 4 00:05:18.522685 kubelet[2362]: I0904 00:05:18.522649 2362 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 00:05:18.524111 kubelet[2362]: I0904 00:05:18.524087 2362 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 00:05:18.524111 kubelet[2362]: I0904 00:05:18.524111 2362 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 4 00:05:18.524204 kubelet[2362]: I0904 00:05:18.524131 2362 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 4 00:05:18.524204 kubelet[2362]: I0904 00:05:18.524138 2362 kubelet.go:2382] "Starting kubelet main sync loop" Sep 4 00:05:18.524204 kubelet[2362]: E0904 00:05:18.524187 2362 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 00:05:18.605073 kubelet[2362]: E0904 00:05:18.605012 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:18.624292 kubelet[2362]: E0904 00:05:18.624231 2362 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 4 00:05:18.705385 kubelet[2362]: E0904 00:05:18.705203 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:18.705722 kubelet[2362]: E0904 00:05:18.705680 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.91:6443: connect: connection refused" interval="400ms" Sep 4 00:05:18.806009 kubelet[2362]: E0904 00:05:18.805907 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:18.825211 kubelet[2362]: E0904 00:05:18.825152 2362 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 4 00:05:18.906718 kubelet[2362]: E0904 00:05:18.906628 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:19.007590 kubelet[2362]: E0904 00:05:19.007379 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:19.082381 kubelet[2362]: W0904 00:05:19.082208 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:19.082381 kubelet[2362]: E0904 00:05:19.082316 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:19.106340 kubelet[2362]: E0904 00:05:19.106265 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.91:6443: connect: connection refused" interval="800ms" Sep 4 00:05:19.108504 kubelet[2362]: E0904 00:05:19.108459 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:19.156517 kubelet[2362]: I0904 00:05:19.156435 2362 policy_none.go:49] "None policy: Start" Sep 4 00:05:19.156517 kubelet[2362]: I0904 00:05:19.156479 2362 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 4 00:05:19.156517 kubelet[2362]: I0904 00:05:19.156494 2362 state_mem.go:35] "Initializing new in-memory state store" Sep 4 00:05:19.209280 kubelet[2362]: E0904 00:05:19.209219 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:19.225476 kubelet[2362]: E0904 00:05:19.225433 2362 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 4 00:05:19.273061 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 4 00:05:19.287933 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 4 00:05:19.291027 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 4 00:05:19.310283 kubelet[2362]: E0904 00:05:19.310249 2362 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:19.313712 kubelet[2362]: I0904 00:05:19.313620 2362 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 00:05:19.313903 kubelet[2362]: I0904 00:05:19.313889 2362 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 4 00:05:19.313932 kubelet[2362]: I0904 00:05:19.313904 2362 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 4 00:05:19.314130 kubelet[2362]: I0904 00:05:19.314110 2362 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 00:05:19.314973 kubelet[2362]: E0904 00:05:19.314951 2362 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 4 00:05:19.315032 kubelet[2362]: E0904 00:05:19.314992 2362 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 4 00:05:19.415482 kubelet[2362]: I0904 00:05:19.415449 2362 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 4 00:05:19.415845 kubelet[2362]: E0904 00:05:19.415807 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.91:6443/api/v1/nodes\": dial tcp 10.0.0.91:6443: connect: connection refused" node="localhost" Sep 4 00:05:19.427220 kubelet[2362]: W0904 00:05:19.427173 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:19.427283 kubelet[2362]: E0904 00:05:19.427232 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:19.526729 kubelet[2362]: W0904 00:05:19.526574 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:19.526729 kubelet[2362]: E0904 00:05:19.526626 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:19.617338 kubelet[2362]: I0904 00:05:19.617295 2362 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 4 00:05:19.617709 kubelet[2362]: E0904 00:05:19.617674 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.91:6443/api/v1/nodes\": dial tcp 10.0.0.91:6443: connect: connection refused" node="localhost" Sep 4 00:05:19.907560 kubelet[2362]: E0904 00:05:19.907407 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.91:6443: connect: connection refused" interval="1.6s" Sep 4 00:05:19.981308 kubelet[2362]: W0904 00:05:19.981236 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:19.981308 kubelet[2362]: E0904 00:05:19.981299 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.91:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:20.019284 kubelet[2362]: I0904 00:05:20.019261 2362 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 4 00:05:20.019567 kubelet[2362]: E0904 00:05:20.019518 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.91:6443/api/v1/nodes\": dial tcp 10.0.0.91:6443: connect: connection refused" node="localhost" Sep 4 00:05:20.032806 systemd[1]: Created slice kubepods-burstable-pod6061e845d3b93496d55d826b9c4df46e.slice - libcontainer container kubepods-burstable-pod6061e845d3b93496d55d826b9c4df46e.slice. Sep 4 00:05:20.047439 kubelet[2362]: E0904 00:05:20.047401 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:20.050232 systemd[1]: Created slice kubepods-burstable-poda88c9297c136b0f15880bf567e89a977.slice - libcontainer container kubepods-burstable-poda88c9297c136b0f15880bf567e89a977.slice. Sep 4 00:05:20.052097 kubelet[2362]: E0904 00:05:20.052079 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:20.054628 systemd[1]: Created slice kubepods-burstable-poda9176403b596d0b29ae8ad12d635226d.slice - libcontainer container kubepods-burstable-poda9176403b596d0b29ae8ad12d635226d.slice. Sep 4 00:05:20.056361 kubelet[2362]: E0904 00:05:20.056325 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:20.115792 kubelet[2362]: I0904 00:05:20.115751 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6061e845d3b93496d55d826b9c4df46e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"6061e845d3b93496d55d826b9c4df46e\") " pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:20.115792 kubelet[2362]: I0904 00:05:20.115795 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6061e845d3b93496d55d826b9c4df46e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"6061e845d3b93496d55d826b9c4df46e\") " pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:20.115926 kubelet[2362]: I0904 00:05:20.115810 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:20.115926 kubelet[2362]: I0904 00:05:20.115828 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:20.115926 kubelet[2362]: I0904 00:05:20.115849 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:20.115926 kubelet[2362]: I0904 00:05:20.115868 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6061e845d3b93496d55d826b9c4df46e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"6061e845d3b93496d55d826b9c4df46e\") " pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:20.115926 kubelet[2362]: I0904 00:05:20.115886 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:20.116034 kubelet[2362]: I0904 00:05:20.115932 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:20.116034 kubelet[2362]: I0904 00:05:20.115967 2362 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a9176403b596d0b29ae8ad12d635226d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a9176403b596d0b29ae8ad12d635226d\") " pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:20.348619 kubelet[2362]: E0904 00:05:20.348536 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:20.349295 containerd[1582]: time="2025-09-04T00:05:20.349254542Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:6061e845d3b93496d55d826b9c4df46e,Namespace:kube-system,Attempt:0,}" Sep 4 00:05:20.353404 kubelet[2362]: E0904 00:05:20.353380 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:20.353695 containerd[1582]: time="2025-09-04T00:05:20.353672588Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:a88c9297c136b0f15880bf567e89a977,Namespace:kube-system,Attempt:0,}" Sep 4 00:05:20.356936 kubelet[2362]: E0904 00:05:20.356916 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:20.357160 containerd[1582]: time="2025-09-04T00:05:20.357139753Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a9176403b596d0b29ae8ad12d635226d,Namespace:kube-system,Attempt:0,}" Sep 4 00:05:20.441348 kubelet[2362]: W0904 00:05:20.441293 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:20.441348 kubelet[2362]: E0904 00:05:20.441337 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.91:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:20.624717 kubelet[2362]: E0904 00:05:20.624539 2362 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.91:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:20.821792 kubelet[2362]: I0904 00:05:20.821737 2362 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 4 00:05:20.822261 kubelet[2362]: E0904 00:05:20.822208 2362 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.91:6443/api/v1/nodes\": dial tcp 10.0.0.91:6443: connect: connection refused" node="localhost" Sep 4 00:05:21.158494 containerd[1582]: time="2025-09-04T00:05:21.157745944Z" level=info msg="connecting to shim 05427b4c6b064f5a6f9449c5deea12be8ddaf675b56b1c197648acc508092556" address="unix:///run/containerd/s/43c41a263310782a1db75127e19d55df9fb07784013c7521f9e873068dde3ab8" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:05:21.170828 containerd[1582]: time="2025-09-04T00:05:21.170748739Z" level=info msg="connecting to shim fc0f927bc5c91355757c0a952283b2c57c18fb1f38cf4e12f9b6894bd2942d2d" address="unix:///run/containerd/s/cb3d84660feb8a3431c63d2cdcb7cb2b7a8a2c6507ae7d5a8ad92de8c80b61c6" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:05:21.188288 containerd[1582]: time="2025-09-04T00:05:21.188215707Z" level=info msg="connecting to shim 1f833b836ae4c89201322fc5432c2635bd75fbaa8b18166fea9c57b5d8d62934" address="unix:///run/containerd/s/6ee5ac9e294ebb6c637555a01115f6a2b55928dc944e2c36f71d116b42b92011" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:05:21.196013 kubelet[2362]: W0904 00:05:21.195965 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:21.196157 kubelet[2362]: E0904 00:05:21.196040 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.91:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:21.420831 systemd[1]: Started cri-containerd-05427b4c6b064f5a6f9449c5deea12be8ddaf675b56b1c197648acc508092556.scope - libcontainer container 05427b4c6b064f5a6f9449c5deea12be8ddaf675b56b1c197648acc508092556. Sep 4 00:05:21.437753 systemd[1]: Started cri-containerd-fc0f927bc5c91355757c0a952283b2c57c18fb1f38cf4e12f9b6894bd2942d2d.scope - libcontainer container fc0f927bc5c91355757c0a952283b2c57c18fb1f38cf4e12f9b6894bd2942d2d. Sep 4 00:05:21.446779 systemd[1]: Started cri-containerd-1f833b836ae4c89201322fc5432c2635bd75fbaa8b18166fea9c57b5d8d62934.scope - libcontainer container 1f833b836ae4c89201322fc5432c2635bd75fbaa8b18166fea9c57b5d8d62934. Sep 4 00:05:21.713372 kubelet[2362]: E0904 00:05:21.713229 2362 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.91:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.91:6443: connect: connection refused" interval="3.2s" Sep 4 00:05:21.753640 containerd[1582]: time="2025-09-04T00:05:21.752393781Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:6061e845d3b93496d55d826b9c4df46e,Namespace:kube-system,Attempt:0,} returns sandbox id \"fc0f927bc5c91355757c0a952283b2c57c18fb1f38cf4e12f9b6894bd2942d2d\"" Sep 4 00:05:21.755656 kubelet[2362]: E0904 00:05:21.755635 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:21.759072 containerd[1582]: time="2025-09-04T00:05:21.759043953Z" level=info msg="CreateContainer within sandbox \"fc0f927bc5c91355757c0a952283b2c57c18fb1f38cf4e12f9b6894bd2942d2d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 4 00:05:21.765197 containerd[1582]: time="2025-09-04T00:05:21.765128514Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a9176403b596d0b29ae8ad12d635226d,Namespace:kube-system,Attempt:0,} returns sandbox id \"05427b4c6b064f5a6f9449c5deea12be8ddaf675b56b1c197648acc508092556\"" Sep 4 00:05:21.766390 kubelet[2362]: E0904 00:05:21.765860 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:21.768627 containerd[1582]: time="2025-09-04T00:05:21.768223580Z" level=info msg="CreateContainer within sandbox \"05427b4c6b064f5a6f9449c5deea12be8ddaf675b56b1c197648acc508092556\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 4 00:05:21.787150 containerd[1582]: time="2025-09-04T00:05:21.787026682Z" level=info msg="Container c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:21.789796 containerd[1582]: time="2025-09-04T00:05:21.789693260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:a88c9297c136b0f15880bf567e89a977,Namespace:kube-system,Attempt:0,} returns sandbox id \"1f833b836ae4c89201322fc5432c2635bd75fbaa8b18166fea9c57b5d8d62934\"" Sep 4 00:05:21.790734 kubelet[2362]: E0904 00:05:21.790688 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:21.792868 containerd[1582]: time="2025-09-04T00:05:21.792813394Z" level=info msg="CreateContainer within sandbox \"1f833b836ae4c89201322fc5432c2635bd75fbaa8b18166fea9c57b5d8d62934\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 4 00:05:21.800473 containerd[1582]: time="2025-09-04T00:05:21.800407790Z" level=info msg="Container c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:21.800841 containerd[1582]: time="2025-09-04T00:05:21.800793107Z" level=info msg="CreateContainer within sandbox \"fc0f927bc5c91355757c0a952283b2c57c18fb1f38cf4e12f9b6894bd2942d2d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193\"" Sep 4 00:05:21.801526 containerd[1582]: time="2025-09-04T00:05:21.801497273Z" level=info msg="StartContainer for \"c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193\"" Sep 4 00:05:21.803042 containerd[1582]: time="2025-09-04T00:05:21.802997962Z" level=info msg="connecting to shim c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193" address="unix:///run/containerd/s/cb3d84660feb8a3431c63d2cdcb7cb2b7a8a2c6507ae7d5a8ad92de8c80b61c6" protocol=ttrpc version=3 Sep 4 00:05:21.814459 containerd[1582]: time="2025-09-04T00:05:21.814229441Z" level=info msg="Container 618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:21.820503 containerd[1582]: time="2025-09-04T00:05:21.820453027Z" level=info msg="CreateContainer within sandbox \"05427b4c6b064f5a6f9449c5deea12be8ddaf675b56b1c197648acc508092556\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9\"" Sep 4 00:05:21.821399 containerd[1582]: time="2025-09-04T00:05:21.821353128Z" level=info msg="StartContainer for \"c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9\"" Sep 4 00:05:21.823776 containerd[1582]: time="2025-09-04T00:05:21.823726074Z" level=info msg="connecting to shim c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9" address="unix:///run/containerd/s/43c41a263310782a1db75127e19d55df9fb07784013c7521f9e873068dde3ab8" protocol=ttrpc version=3 Sep 4 00:05:21.827954 containerd[1582]: time="2025-09-04T00:05:21.827917075Z" level=info msg="CreateContainer within sandbox \"1f833b836ae4c89201322fc5432c2635bd75fbaa8b18166fea9c57b5d8d62934\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439\"" Sep 4 00:05:21.828479 containerd[1582]: time="2025-09-04T00:05:21.828459431Z" level=info msg="StartContainer for \"618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439\"" Sep 4 00:05:21.829875 containerd[1582]: time="2025-09-04T00:05:21.829851592Z" level=info msg="connecting to shim 618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439" address="unix:///run/containerd/s/6ee5ac9e294ebb6c637555a01115f6a2b55928dc944e2c36f71d116b42b92011" protocol=ttrpc version=3 Sep 4 00:05:21.977208 systemd[1]: Started cri-containerd-c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193.scope - libcontainer container c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193. Sep 4 00:05:22.005957 systemd[1]: Started cri-containerd-618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439.scope - libcontainer container 618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439. Sep 4 00:05:22.031155 systemd[1]: Started cri-containerd-c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9.scope - libcontainer container c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9. Sep 4 00:05:22.067614 kubelet[2362]: W0904 00:05:22.067316 2362 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.91:6443: connect: connection refused Sep 4 00:05:22.067614 kubelet[2362]: E0904 00:05:22.067420 2362 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.91:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.91:6443: connect: connection refused" logger="UnhandledError" Sep 4 00:05:22.151357 containerd[1582]: time="2025-09-04T00:05:22.151275492Z" level=info msg="StartContainer for \"c4372f564f74bf18be97c041c8263c657f0c380a3a223af1765d5aa916d42193\" returns successfully" Sep 4 00:05:22.187585 containerd[1582]: time="2025-09-04T00:05:22.187507198Z" level=info msg="StartContainer for \"618b4d9a347b0a2409b5acd7190899a27a18c3f198897ee4eea2daf4fe564439\" returns successfully" Sep 4 00:05:22.194374 containerd[1582]: time="2025-09-04T00:05:22.194253529Z" level=info msg="StartContainer for \"c182fbfa95327f6bc765698a38419a431df2a68992f4eb87b172b69b1493b8d9\" returns successfully" Sep 4 00:05:22.424235 kubelet[2362]: I0904 00:05:22.424118 2362 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 4 00:05:22.721575 kubelet[2362]: E0904 00:05:22.721473 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:22.722732 kubelet[2362]: E0904 00:05:22.722095 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:22.725377 kubelet[2362]: E0904 00:05:22.725196 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:22.725377 kubelet[2362]: E0904 00:05:22.725325 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:22.728319 kubelet[2362]: E0904 00:05:22.728238 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:22.728416 kubelet[2362]: E0904 00:05:22.728400 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:23.732009 kubelet[2362]: E0904 00:05:23.731959 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:23.732581 kubelet[2362]: E0904 00:05:23.732102 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:23.732581 kubelet[2362]: E0904 00:05:23.732319 2362 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 4 00:05:23.732581 kubelet[2362]: E0904 00:05:23.732401 2362 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:24.582224 kubelet[2362]: I0904 00:05:24.582147 2362 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 4 00:05:24.605799 kubelet[2362]: I0904 00:05:24.605743 2362 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:24.613372 kubelet[2362]: E0904 00:05:24.613324 2362 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:24.613372 kubelet[2362]: I0904 00:05:24.613352 2362 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:24.615998 kubelet[2362]: E0904 00:05:24.615767 2362 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:24.615998 kubelet[2362]: I0904 00:05:24.615811 2362 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:24.618427 kubelet[2362]: E0904 00:05:24.618387 2362 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:24.712141 kubelet[2362]: I0904 00:05:24.712092 2362 apiserver.go:52] "Watching apiserver" Sep 4 00:05:24.805290 kubelet[2362]: I0904 00:05:24.805224 2362 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 4 00:05:26.592802 systemd[1]: Reload requested from client PID 2642 ('systemctl') (unit session-9.scope)... Sep 4 00:05:26.592818 systemd[1]: Reloading... Sep 4 00:05:26.681592 zram_generator::config[2685]: No configuration found. Sep 4 00:05:26.784292 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 00:05:26.931042 systemd[1]: Reloading finished in 337 ms. Sep 4 00:05:26.956798 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:05:26.979188 systemd[1]: kubelet.service: Deactivated successfully. Sep 4 00:05:26.979469 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:05:26.979525 systemd[1]: kubelet.service: Consumed 1.078s CPU time, 130.9M memory peak. Sep 4 00:05:26.982747 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 00:05:27.249048 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 00:05:27.266314 (kubelet)[2730]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 00:05:27.651957 kubelet[2730]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 00:05:27.651957 kubelet[2730]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 4 00:05:27.651957 kubelet[2730]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 00:05:27.651957 kubelet[2730]: I0904 00:05:27.651912 2730 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 00:05:27.657828 kubelet[2730]: I0904 00:05:27.657785 2730 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 4 00:05:27.657828 kubelet[2730]: I0904 00:05:27.657812 2730 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 00:05:27.658095 kubelet[2730]: I0904 00:05:27.658070 2730 server.go:954] "Client rotation is on, will bootstrap in background" Sep 4 00:05:27.659228 kubelet[2730]: I0904 00:05:27.659201 2730 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 4 00:05:27.661147 kubelet[2730]: I0904 00:05:27.661108 2730 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 00:05:27.667567 kubelet[2730]: I0904 00:05:27.666488 2730 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 4 00:05:27.671379 kubelet[2730]: I0904 00:05:27.671344 2730 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 00:05:27.671630 kubelet[2730]: I0904 00:05:27.671585 2730 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 00:05:27.671850 kubelet[2730]: I0904 00:05:27.671622 2730 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 4 00:05:27.671957 kubelet[2730]: I0904 00:05:27.671853 2730 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 00:05:27.671957 kubelet[2730]: I0904 00:05:27.671866 2730 container_manager_linux.go:304] "Creating device plugin manager" Sep 4 00:05:27.671957 kubelet[2730]: I0904 00:05:27.671922 2730 state_mem.go:36] "Initialized new in-memory state store" Sep 4 00:05:27.672122 kubelet[2730]: I0904 00:05:27.672099 2730 kubelet.go:446] "Attempting to sync node with API server" Sep 4 00:05:27.672179 kubelet[2730]: I0904 00:05:27.672133 2730 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 00:05:27.672179 kubelet[2730]: I0904 00:05:27.672162 2730 kubelet.go:352] "Adding apiserver pod source" Sep 4 00:05:27.672179 kubelet[2730]: I0904 00:05:27.672174 2730 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 00:05:27.673180 kubelet[2730]: I0904 00:05:27.673159 2730 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Sep 4 00:05:27.673474 kubelet[2730]: I0904 00:05:27.673460 2730 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 4 00:05:27.673880 kubelet[2730]: I0904 00:05:27.673852 2730 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 4 00:05:27.673880 kubelet[2730]: I0904 00:05:27.673876 2730 server.go:1287] "Started kubelet" Sep 4 00:05:27.677448 kubelet[2730]: I0904 00:05:27.675399 2730 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 00:05:27.677448 kubelet[2730]: I0904 00:05:27.676779 2730 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 00:05:27.678438 kubelet[2730]: I0904 00:05:27.678383 2730 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 4 00:05:27.678993 kubelet[2730]: I0904 00:05:27.678969 2730 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 00:05:27.680579 kubelet[2730]: I0904 00:05:27.679936 2730 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 4 00:05:27.684974 kubelet[2730]: E0904 00:05:27.684943 2730 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 00:05:27.685995 kubelet[2730]: E0904 00:05:27.685953 2730 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 00:05:27.686084 kubelet[2730]: I0904 00:05:27.686028 2730 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 4 00:05:27.686205 kubelet[2730]: I0904 00:05:27.686187 2730 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 4 00:05:27.687584 kubelet[2730]: I0904 00:05:27.687190 2730 server.go:479] "Adding debug handlers to kubelet server" Sep 4 00:05:27.688040 kubelet[2730]: I0904 00:05:27.688007 2730 reconciler.go:26] "Reconciler: start to sync state" Sep 4 00:05:27.691355 kubelet[2730]: I0904 00:05:27.691284 2730 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 00:05:27.691594 kubelet[2730]: I0904 00:05:27.691436 2730 factory.go:221] Registration of the systemd container factory successfully Sep 4 00:05:27.691781 kubelet[2730]: I0904 00:05:27.691754 2730 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 4 00:05:27.693701 kubelet[2730]: I0904 00:05:27.693670 2730 factory.go:221] Registration of the containerd container factory successfully Sep 4 00:05:27.693973 kubelet[2730]: I0904 00:05:27.693943 2730 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 00:05:27.693973 kubelet[2730]: I0904 00:05:27.693964 2730 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 4 00:05:27.694149 kubelet[2730]: I0904 00:05:27.693994 2730 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 4 00:05:27.694149 kubelet[2730]: I0904 00:05:27.694004 2730 kubelet.go:2382] "Starting kubelet main sync loop" Sep 4 00:05:27.694149 kubelet[2730]: E0904 00:05:27.694056 2730 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 00:05:27.725020 kubelet[2730]: I0904 00:05:27.724988 2730 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 4 00:05:27.725020 kubelet[2730]: I0904 00:05:27.725005 2730 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 4 00:05:27.725020 kubelet[2730]: I0904 00:05:27.725022 2730 state_mem.go:36] "Initialized new in-memory state store" Sep 4 00:05:27.725205 kubelet[2730]: I0904 00:05:27.725165 2730 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 4 00:05:27.725205 kubelet[2730]: I0904 00:05:27.725175 2730 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 4 00:05:27.725205 kubelet[2730]: I0904 00:05:27.725193 2730 policy_none.go:49] "None policy: Start" Sep 4 00:05:27.725205 kubelet[2730]: I0904 00:05:27.725201 2730 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 4 00:05:27.725301 kubelet[2730]: I0904 00:05:27.725210 2730 state_mem.go:35] "Initializing new in-memory state store" Sep 4 00:05:27.726561 kubelet[2730]: I0904 00:05:27.725484 2730 state_mem.go:75] "Updated machine memory state" Sep 4 00:05:27.730827 kubelet[2730]: I0904 00:05:27.730788 2730 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 00:05:27.731003 kubelet[2730]: I0904 00:05:27.730985 2730 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 4 00:05:27.731063 kubelet[2730]: I0904 00:05:27.730998 2730 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 4 00:05:27.731262 kubelet[2730]: I0904 00:05:27.731242 2730 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 00:05:27.735955 kubelet[2730]: E0904 00:05:27.735383 2730 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 4 00:05:27.795046 kubelet[2730]: I0904 00:05:27.794990 2730 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:27.795046 kubelet[2730]: I0904 00:05:27.795034 2730 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:27.795230 kubelet[2730]: I0904 00:05:27.794990 2730 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:27.835646 kubelet[2730]: I0904 00:05:27.835612 2730 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 4 00:05:27.889173 kubelet[2730]: I0904 00:05:27.889116 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a9176403b596d0b29ae8ad12d635226d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a9176403b596d0b29ae8ad12d635226d\") " pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:27.889173 kubelet[2730]: I0904 00:05:27.889170 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6061e845d3b93496d55d826b9c4df46e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"6061e845d3b93496d55d826b9c4df46e\") " pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:27.889346 kubelet[2730]: I0904 00:05:27.889213 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:27.889346 kubelet[2730]: I0904 00:05:27.889262 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:27.889346 kubelet[2730]: I0904 00:05:27.889308 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6061e845d3b93496d55d826b9c4df46e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"6061e845d3b93496d55d826b9c4df46e\") " pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:27.889346 kubelet[2730]: I0904 00:05:27.889336 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6061e845d3b93496d55d826b9c4df46e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"6061e845d3b93496d55d826b9c4df46e\") " pod="kube-system/kube-apiserver-localhost" Sep 4 00:05:27.889513 kubelet[2730]: I0904 00:05:27.889358 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:27.889513 kubelet[2730]: I0904 00:05:27.889379 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:27.889513 kubelet[2730]: I0904 00:05:27.889399 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 00:05:28.043583 kubelet[2730]: E0904 00:05:28.043521 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:28.043975 kubelet[2730]: E0904 00:05:28.043907 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:28.044211 kubelet[2730]: E0904 00:05:28.044098 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:28.096469 kubelet[2730]: I0904 00:05:28.096428 2730 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 4 00:05:28.096659 kubelet[2730]: I0904 00:05:28.096534 2730 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 4 00:05:28.672704 kubelet[2730]: I0904 00:05:28.672620 2730 apiserver.go:52] "Watching apiserver" Sep 4 00:05:28.686690 kubelet[2730]: I0904 00:05:28.686623 2730 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 4 00:05:28.707488 kubelet[2730]: E0904 00:05:28.707400 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:28.707931 kubelet[2730]: I0904 00:05:28.707914 2730 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:28.708178 kubelet[2730]: E0904 00:05:28.708111 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:28.788318 kubelet[2730]: E0904 00:05:28.788240 2730 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 4 00:05:28.788704 kubelet[2730]: E0904 00:05:28.788628 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:28.880803 kubelet[2730]: I0904 00:05:28.880688 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.880662934 podStartE2EDuration="1.880662934s" podCreationTimestamp="2025-09-04 00:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:05:28.880467883 +0000 UTC m=+1.608758766" watchObservedRunningTime="2025-09-04 00:05:28.880662934 +0000 UTC m=+1.608953817" Sep 4 00:05:28.890698 sudo[2764]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 4 00:05:28.891103 sudo[2764]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 4 00:05:28.899301 kubelet[2730]: I0904 00:05:28.899232 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.8992118690000002 podStartE2EDuration="1.899211869s" podCreationTimestamp="2025-09-04 00:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:05:28.889474923 +0000 UTC m=+1.617765806" watchObservedRunningTime="2025-09-04 00:05:28.899211869 +0000 UTC m=+1.627502752" Sep 4 00:05:28.899456 kubelet[2730]: I0904 00:05:28.899333 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.899326427 podStartE2EDuration="1.899326427s" podCreationTimestamp="2025-09-04 00:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:05:28.899116578 +0000 UTC m=+1.627407471" watchObservedRunningTime="2025-09-04 00:05:28.899326427 +0000 UTC m=+1.627617310" Sep 4 00:05:29.403868 sudo[2764]: pam_unix(sudo:session): session closed for user root Sep 4 00:05:29.709399 kubelet[2730]: E0904 00:05:29.709362 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:29.709967 kubelet[2730]: E0904 00:05:29.709539 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:30.075983 update_engine[1568]: I20250904 00:05:30.075785 1568 update_attempter.cc:509] Updating boot flags... Sep 4 00:05:30.968022 sudo[1797]: pam_unix(sudo:session): session closed for user root Sep 4 00:05:30.969984 sshd[1796]: Connection closed by 10.0.0.1 port 42660 Sep 4 00:05:30.970389 sshd-session[1794]: pam_unix(sshd:session): session closed for user core Sep 4 00:05:30.975200 systemd[1]: sshd@8-10.0.0.91:22-10.0.0.1:42660.service: Deactivated successfully. Sep 4 00:05:30.977336 systemd[1]: session-9.scope: Deactivated successfully. Sep 4 00:05:30.977598 systemd[1]: session-9.scope: Consumed 4.180s CPU time, 255.8M memory peak. Sep 4 00:05:30.979027 systemd-logind[1561]: Session 9 logged out. Waiting for processes to exit. Sep 4 00:05:30.981314 systemd-logind[1561]: Removed session 9. Sep 4 00:05:32.601324 kubelet[2730]: I0904 00:05:32.601280 2730 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 4 00:05:32.601806 kubelet[2730]: I0904 00:05:32.601737 2730 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 4 00:05:32.601845 containerd[1582]: time="2025-09-04T00:05:32.601583554Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 4 00:05:33.313798 systemd[1]: Created slice kubepods-besteffort-podcdf64e75_c0e8_4979_92ef_883d54dd53c4.slice - libcontainer container kubepods-besteffort-podcdf64e75_c0e8_4979_92ef_883d54dd53c4.slice. Sep 4 00:05:33.324276 kubelet[2730]: I0904 00:05:33.324221 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/cdf64e75-c0e8-4979-92ef-883d54dd53c4-kube-proxy\") pod \"kube-proxy-vkpf4\" (UID: \"cdf64e75-c0e8-4979-92ef-883d54dd53c4\") " pod="kube-system/kube-proxy-vkpf4" Sep 4 00:05:33.324276 kubelet[2730]: I0904 00:05:33.324269 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cdf64e75-c0e8-4979-92ef-883d54dd53c4-xtables-lock\") pod \"kube-proxy-vkpf4\" (UID: \"cdf64e75-c0e8-4979-92ef-883d54dd53c4\") " pod="kube-system/kube-proxy-vkpf4" Sep 4 00:05:33.324436 kubelet[2730]: I0904 00:05:33.324297 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cdf64e75-c0e8-4979-92ef-883d54dd53c4-lib-modules\") pod \"kube-proxy-vkpf4\" (UID: \"cdf64e75-c0e8-4979-92ef-883d54dd53c4\") " pod="kube-system/kube-proxy-vkpf4" Sep 4 00:05:33.324436 kubelet[2730]: I0904 00:05:33.324318 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz8wj\" (UniqueName: \"kubernetes.io/projected/cdf64e75-c0e8-4979-92ef-883d54dd53c4-kube-api-access-lz8wj\") pod \"kube-proxy-vkpf4\" (UID: \"cdf64e75-c0e8-4979-92ef-883d54dd53c4\") " pod="kube-system/kube-proxy-vkpf4" Sep 4 00:05:33.329607 systemd[1]: Created slice kubepods-burstable-poda61e66f7_5c59_4052_9364_69f00bf363d7.slice - libcontainer container kubepods-burstable-poda61e66f7_5c59_4052_9364_69f00bf363d7.slice. Sep 4 00:05:33.424759 kubelet[2730]: I0904 00:05:33.424701 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/a61e66f7-5c59-4052-9364-69f00bf363d7-clustermesh-secrets\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.424759 kubelet[2730]: I0904 00:05:33.424741 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-config-path\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.424759 kubelet[2730]: I0904 00:05:33.424762 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-kernel\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425010 kubelet[2730]: I0904 00:05:33.424782 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ln5n\" (UniqueName: \"kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-kube-api-access-6ln5n\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425408 kubelet[2730]: I0904 00:05:33.425376 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-bpf-maps\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425533 kubelet[2730]: I0904 00:05:33.425419 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-cgroup\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425533 kubelet[2730]: I0904 00:05:33.425437 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-etc-cni-netd\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425533 kubelet[2730]: I0904 00:05:33.425456 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-net\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425533 kubelet[2730]: I0904 00:05:33.425475 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-xtables-lock\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425533 kubelet[2730]: I0904 00:05:33.425504 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-hostproc\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425533 kubelet[2730]: I0904 00:05:33.425533 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-lib-modules\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425766 kubelet[2730]: I0904 00:05:33.425570 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-run\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425766 kubelet[2730]: I0904 00:05:33.425591 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cni-path\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.425766 kubelet[2730]: I0904 00:05:33.425613 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-hubble-tls\") pod \"cilium-fq794\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " pod="kube-system/cilium-fq794" Sep 4 00:05:33.430161 kubelet[2730]: E0904 00:05:33.430122 2730 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 4 00:05:33.430161 kubelet[2730]: E0904 00:05:33.430149 2730 projected.go:194] Error preparing data for projected volume kube-api-access-lz8wj for pod kube-system/kube-proxy-vkpf4: configmap "kube-root-ca.crt" not found Sep 4 00:05:33.430268 kubelet[2730]: E0904 00:05:33.430216 2730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cdf64e75-c0e8-4979-92ef-883d54dd53c4-kube-api-access-lz8wj podName:cdf64e75-c0e8-4979-92ef-883d54dd53c4 nodeName:}" failed. No retries permitted until 2025-09-04 00:05:33.930190006 +0000 UTC m=+6.658480949 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-lz8wj" (UniqueName: "kubernetes.io/projected/cdf64e75-c0e8-4979-92ef-883d54dd53c4-kube-api-access-lz8wj") pod "kube-proxy-vkpf4" (UID: "cdf64e75-c0e8-4979-92ef-883d54dd53c4") : configmap "kube-root-ca.crt" not found Sep 4 00:05:33.533900 kubelet[2730]: E0904 00:05:33.533850 2730 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 4 00:05:33.533900 kubelet[2730]: E0904 00:05:33.533883 2730 projected.go:194] Error preparing data for projected volume kube-api-access-6ln5n for pod kube-system/cilium-fq794: configmap "kube-root-ca.crt" not found Sep 4 00:05:33.534057 kubelet[2730]: E0904 00:05:33.533933 2730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-kube-api-access-6ln5n podName:a61e66f7-5c59-4052-9364-69f00bf363d7 nodeName:}" failed. No retries permitted until 2025-09-04 00:05:34.033911119 +0000 UTC m=+6.762202002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6ln5n" (UniqueName: "kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-kube-api-access-6ln5n") pod "cilium-fq794" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7") : configmap "kube-root-ca.crt" not found Sep 4 00:05:33.691113 systemd[1]: Created slice kubepods-besteffort-pod84438b01_62a9_44cf_9db8_0c9d2cca272c.slice - libcontainer container kubepods-besteffort-pod84438b01_62a9_44cf_9db8_0c9d2cca272c.slice. Sep 4 00:05:33.726762 kubelet[2730]: I0904 00:05:33.726696 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsnds\" (UniqueName: \"kubernetes.io/projected/84438b01-62a9-44cf-9db8-0c9d2cca272c-kube-api-access-zsnds\") pod \"cilium-operator-6c4d7847fc-vdjlm\" (UID: \"84438b01-62a9-44cf-9db8-0c9d2cca272c\") " pod="kube-system/cilium-operator-6c4d7847fc-vdjlm" Sep 4 00:05:33.726762 kubelet[2730]: I0904 00:05:33.726763 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/84438b01-62a9-44cf-9db8-0c9d2cca272c-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-vdjlm\" (UID: \"84438b01-62a9-44cf-9db8-0c9d2cca272c\") " pod="kube-system/cilium-operator-6c4d7847fc-vdjlm" Sep 4 00:05:33.996623 kubelet[2730]: E0904 00:05:33.996435 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:33.997218 containerd[1582]: time="2025-09-04T00:05:33.997175735Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-vdjlm,Uid:84438b01-62a9-44cf-9db8-0c9d2cca272c,Namespace:kube-system,Attempt:0,}" Sep 4 00:05:34.042909 containerd[1582]: time="2025-09-04T00:05:34.042856821Z" level=info msg="connecting to shim 3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6" address="unix:///run/containerd/s/6f0400d6c08286a17b8ab9fdc5e8519109e8e607cd422fc308bf8d0b39784ff9" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:05:34.109791 systemd[1]: Started cri-containerd-3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6.scope - libcontainer container 3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6. Sep 4 00:05:34.155846 containerd[1582]: time="2025-09-04T00:05:34.155783876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-vdjlm,Uid:84438b01-62a9-44cf-9db8-0c9d2cca272c,Namespace:kube-system,Attempt:0,} returns sandbox id \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\"" Sep 4 00:05:34.156728 kubelet[2730]: E0904 00:05:34.156699 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:34.157848 containerd[1582]: time="2025-09-04T00:05:34.157803186Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 4 00:05:34.222402 kubelet[2730]: E0904 00:05:34.222344 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:34.223011 containerd[1582]: time="2025-09-04T00:05:34.222945511Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-vkpf4,Uid:cdf64e75-c0e8-4979-92ef-883d54dd53c4,Namespace:kube-system,Attempt:0,}" Sep 4 00:05:34.233396 kubelet[2730]: E0904 00:05:34.233357 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:34.233979 containerd[1582]: time="2025-09-04T00:05:34.233938001Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fq794,Uid:a61e66f7-5c59-4052-9364-69f00bf363d7,Namespace:kube-system,Attempt:0,}" Sep 4 00:05:34.252586 containerd[1582]: time="2025-09-04T00:05:34.252447095Z" level=info msg="connecting to shim 9f9b3d5b4fe6d5365835b898c253553395b7e87d8abc432bf56e1c024199a03c" address="unix:///run/containerd/s/d2465e9b5bd2a84ced74d11d6e86f59d7ec451d213e33cd7c901ba20970070a3" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:05:34.265718 containerd[1582]: time="2025-09-04T00:05:34.265668820Z" level=info msg="connecting to shim 693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b" address="unix:///run/containerd/s/95425d4e06495a3444fb4952d37032e9277083280889ada29a9d17eb231cd942" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:05:34.281681 systemd[1]: Started cri-containerd-9f9b3d5b4fe6d5365835b898c253553395b7e87d8abc432bf56e1c024199a03c.scope - libcontainer container 9f9b3d5b4fe6d5365835b898c253553395b7e87d8abc432bf56e1c024199a03c. Sep 4 00:05:34.289415 systemd[1]: Started cri-containerd-693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b.scope - libcontainer container 693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b. Sep 4 00:05:34.318941 containerd[1582]: time="2025-09-04T00:05:34.318899799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-vkpf4,Uid:cdf64e75-c0e8-4979-92ef-883d54dd53c4,Namespace:kube-system,Attempt:0,} returns sandbox id \"9f9b3d5b4fe6d5365835b898c253553395b7e87d8abc432bf56e1c024199a03c\"" Sep 4 00:05:34.320244 kubelet[2730]: E0904 00:05:34.320200 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:34.321297 containerd[1582]: time="2025-09-04T00:05:34.320837684Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fq794,Uid:a61e66f7-5c59-4052-9364-69f00bf363d7,Namespace:kube-system,Attempt:0,} returns sandbox id \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\"" Sep 4 00:05:34.321356 kubelet[2730]: E0904 00:05:34.321172 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:34.323454 containerd[1582]: time="2025-09-04T00:05:34.323428564Z" level=info msg="CreateContainer within sandbox \"9f9b3d5b4fe6d5365835b898c253553395b7e87d8abc432bf56e1c024199a03c\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 4 00:05:34.335074 containerd[1582]: time="2025-09-04T00:05:34.335021137Z" level=info msg="Container 7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:34.347586 containerd[1582]: time="2025-09-04T00:05:34.347527629Z" level=info msg="CreateContainer within sandbox \"9f9b3d5b4fe6d5365835b898c253553395b7e87d8abc432bf56e1c024199a03c\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f\"" Sep 4 00:05:34.348112 containerd[1582]: time="2025-09-04T00:05:34.348079573Z" level=info msg="StartContainer for \"7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f\"" Sep 4 00:05:34.349457 containerd[1582]: time="2025-09-04T00:05:34.349418816Z" level=info msg="connecting to shim 7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f" address="unix:///run/containerd/s/d2465e9b5bd2a84ced74d11d6e86f59d7ec451d213e33cd7c901ba20970070a3" protocol=ttrpc version=3 Sep 4 00:05:34.371773 systemd[1]: Started cri-containerd-7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f.scope - libcontainer container 7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f. Sep 4 00:05:34.432135 containerd[1582]: time="2025-09-04T00:05:34.432073331Z" level=info msg="StartContainer for \"7ec132391165481b9c56d90bed9d299a68b6fc4e1a4e27cbb038887362da4f3f\" returns successfully" Sep 4 00:05:34.718304 kubelet[2730]: E0904 00:05:34.718274 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:34.799313 kubelet[2730]: I0904 00:05:34.799245 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-vkpf4" podStartSLOduration=1.7992241610000002 podStartE2EDuration="1.799224161s" podCreationTimestamp="2025-09-04 00:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:05:34.799103443 +0000 UTC m=+7.527394346" watchObservedRunningTime="2025-09-04 00:05:34.799224161 +0000 UTC m=+7.527515044" Sep 4 00:05:34.870401 kubelet[2730]: E0904 00:05:34.870346 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:35.328906 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2806003223.mount: Deactivated successfully. Sep 4 00:05:35.721023 kubelet[2730]: E0904 00:05:35.720974 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:36.340693 containerd[1582]: time="2025-09-04T00:05:36.340616366Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:36.342841 containerd[1582]: time="2025-09-04T00:05:36.342785323Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Sep 4 00:05:36.344386 containerd[1582]: time="2025-09-04T00:05:36.344280128Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:36.345822 containerd[1582]: time="2025-09-04T00:05:36.345762158Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.187917414s" Sep 4 00:05:36.345822 containerd[1582]: time="2025-09-04T00:05:36.345810890Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 4 00:05:36.346814 containerd[1582]: time="2025-09-04T00:05:36.346783999Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 4 00:05:36.348085 containerd[1582]: time="2025-09-04T00:05:36.348046624Z" level=info msg="CreateContainer within sandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 4 00:05:36.359788 containerd[1582]: time="2025-09-04T00:05:36.359724657Z" level=info msg="Container 6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:36.377089 containerd[1582]: time="2025-09-04T00:05:36.377031976Z" level=info msg="CreateContainer within sandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\"" Sep 4 00:05:36.377786 containerd[1582]: time="2025-09-04T00:05:36.377724655Z" level=info msg="StartContainer for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\"" Sep 4 00:05:36.378819 containerd[1582]: time="2025-09-04T00:05:36.378790880Z" level=info msg="connecting to shim 6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f" address="unix:///run/containerd/s/6f0400d6c08286a17b8ab9fdc5e8519109e8e607cd422fc308bf8d0b39784ff9" protocol=ttrpc version=3 Sep 4 00:05:36.401797 systemd[1]: Started cri-containerd-6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f.scope - libcontainer container 6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f. Sep 4 00:05:36.554856 containerd[1582]: time="2025-09-04T00:05:36.554808021Z" level=info msg="StartContainer for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" returns successfully" Sep 4 00:05:36.723888 kubelet[2730]: E0904 00:05:36.723856 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:36.724438 kubelet[2730]: E0904 00:05:36.723925 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:36.733374 kubelet[2730]: I0904 00:05:36.733310 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-vdjlm" podStartSLOduration=1.544009629 podStartE2EDuration="3.733293555s" podCreationTimestamp="2025-09-04 00:05:33 +0000 UTC" firstStartedPulling="2025-09-04 00:05:34.157332896 +0000 UTC m=+6.885623779" lastFinishedPulling="2025-09-04 00:05:36.346616822 +0000 UTC m=+9.074907705" observedRunningTime="2025-09-04 00:05:36.732934427 +0000 UTC m=+9.461225310" watchObservedRunningTime="2025-09-04 00:05:36.733293555 +0000 UTC m=+9.461584448" Sep 4 00:05:37.411561 kubelet[2730]: E0904 00:05:37.411507 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:37.476954 kubelet[2730]: E0904 00:05:37.476909 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:37.725761 kubelet[2730]: E0904 00:05:37.725727 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:37.726286 kubelet[2730]: E0904 00:05:37.725841 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:37.726286 kubelet[2730]: E0904 00:05:37.725843 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:38.729030 kubelet[2730]: E0904 00:05:38.728997 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:44.963864 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3915210582.mount: Deactivated successfully. Sep 4 00:05:51.178585 kubelet[2730]: E0904 00:05:51.178515 2730 kubelet.go:2573] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.484s" Sep 4 00:05:52.097972 containerd[1582]: time="2025-09-04T00:05:52.097904919Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:52.114684 containerd[1582]: time="2025-09-04T00:05:52.114605746Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Sep 4 00:05:52.132228 containerd[1582]: time="2025-09-04T00:05:52.132156000Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 00:05:52.133779 containerd[1582]: time="2025-09-04T00:05:52.133747463Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 15.786896569s" Sep 4 00:05:52.133779 containerd[1582]: time="2025-09-04T00:05:52.133780866Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 4 00:05:52.148623 containerd[1582]: time="2025-09-04T00:05:52.148570929Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 4 00:05:52.442558 containerd[1582]: time="2025-09-04T00:05:52.442499178Z" level=info msg="Container a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:52.446678 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2535349056.mount: Deactivated successfully. Sep 4 00:05:52.632947 containerd[1582]: time="2025-09-04T00:05:52.632906138Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\"" Sep 4 00:05:52.633539 containerd[1582]: time="2025-09-04T00:05:52.633489205Z" level=info msg="StartContainer for \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\"" Sep 4 00:05:52.634564 containerd[1582]: time="2025-09-04T00:05:52.634526155Z" level=info msg="connecting to shim a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa" address="unix:///run/containerd/s/95425d4e06495a3444fb4952d37032e9277083280889ada29a9d17eb231cd942" protocol=ttrpc version=3 Sep 4 00:05:52.657780 systemd[1]: Started cri-containerd-a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa.scope - libcontainer container a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa. Sep 4 00:05:52.805020 containerd[1582]: time="2025-09-04T00:05:52.804724032Z" level=info msg="StartContainer for \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" returns successfully" Sep 4 00:05:52.809843 systemd[1]: cri-containerd-a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa.scope: Deactivated successfully. Sep 4 00:05:52.821879 containerd[1582]: time="2025-09-04T00:05:52.821809162Z" level=info msg="received exit event container_id:\"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" id:\"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" pid:3213 exited_at:{seconds:1756944352 nanos:812302464}" Sep 4 00:05:52.822042 containerd[1582]: time="2025-09-04T00:05:52.821958373Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" id:\"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" pid:3213 exited_at:{seconds:1756944352 nanos:812302464}" Sep 4 00:05:52.842622 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa-rootfs.mount: Deactivated successfully. Sep 4 00:05:53.810383 kubelet[2730]: E0904 00:05:53.810335 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:54.813056 kubelet[2730]: E0904 00:05:54.813002 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:54.815696 containerd[1582]: time="2025-09-04T00:05:54.815653458Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 4 00:05:54.945147 containerd[1582]: time="2025-09-04T00:05:54.945096477Z" level=info msg="Container 1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:54.951920 containerd[1582]: time="2025-09-04T00:05:54.951863709Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\"" Sep 4 00:05:54.952327 containerd[1582]: time="2025-09-04T00:05:54.952288848Z" level=info msg="StartContainer for \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\"" Sep 4 00:05:54.953373 containerd[1582]: time="2025-09-04T00:05:54.953338281Z" level=info msg="connecting to shim 1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e" address="unix:///run/containerd/s/95425d4e06495a3444fb4952d37032e9277083280889ada29a9d17eb231cd942" protocol=ttrpc version=3 Sep 4 00:05:54.979764 systemd[1]: Started cri-containerd-1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e.scope - libcontainer container 1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e. Sep 4 00:05:55.014284 containerd[1582]: time="2025-09-04T00:05:55.014223200Z" level=info msg="StartContainer for \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" returns successfully" Sep 4 00:05:55.026928 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 00:05:55.027222 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 00:05:55.027688 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 4 00:05:55.029327 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 00:05:55.031587 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 4 00:05:55.032372 systemd[1]: cri-containerd-1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e.scope: Deactivated successfully. Sep 4 00:05:55.033565 containerd[1582]: time="2025-09-04T00:05:55.032616223Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" id:\"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" pid:3256 exited_at:{seconds:1756944355 nanos:31995987}" Sep 4 00:05:55.033565 containerd[1582]: time="2025-09-04T00:05:55.032721331Z" level=info msg="received exit event container_id:\"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" id:\"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" pid:3256 exited_at:{seconds:1756944355 nanos:31995987}" Sep 4 00:05:55.066573 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 00:05:55.816575 kubelet[2730]: E0904 00:05:55.816528 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:55.818743 containerd[1582]: time="2025-09-04T00:05:55.818691681Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 4 00:05:55.945884 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e-rootfs.mount: Deactivated successfully. Sep 4 00:05:56.051256 containerd[1582]: time="2025-09-04T00:05:56.051203003Z" level=info msg="Container 673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:56.062265 containerd[1582]: time="2025-09-04T00:05:56.062202075Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\"" Sep 4 00:05:56.062911 containerd[1582]: time="2025-09-04T00:05:56.062885550Z" level=info msg="StartContainer for \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\"" Sep 4 00:05:56.064521 containerd[1582]: time="2025-09-04T00:05:56.064494293Z" level=info msg="connecting to shim 673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e" address="unix:///run/containerd/s/95425d4e06495a3444fb4952d37032e9277083280889ada29a9d17eb231cd942" protocol=ttrpc version=3 Sep 4 00:05:56.083705 systemd[1]: Started cri-containerd-673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e.scope - libcontainer container 673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e. Sep 4 00:05:56.125200 systemd[1]: cri-containerd-673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e.scope: Deactivated successfully. Sep 4 00:05:56.126492 containerd[1582]: time="2025-09-04T00:05:56.126457670Z" level=info msg="StartContainer for \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" returns successfully" Sep 4 00:05:56.126591 containerd[1582]: time="2025-09-04T00:05:56.126462359Z" level=info msg="received exit event container_id:\"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" id:\"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" pid:3305 exited_at:{seconds:1756944356 nanos:126284645}" Sep 4 00:05:56.126646 containerd[1582]: time="2025-09-04T00:05:56.126540436Z" level=info msg="TaskExit event in podsandbox handler container_id:\"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" id:\"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" pid:3305 exited_at:{seconds:1756944356 nanos:126284645}" Sep 4 00:05:56.154914 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e-rootfs.mount: Deactivated successfully. Sep 4 00:05:56.821451 kubelet[2730]: E0904 00:05:56.821411 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:05:56.823157 containerd[1582]: time="2025-09-04T00:05:56.823072896Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 4 00:05:57.346503 containerd[1582]: time="2025-09-04T00:05:57.346447614Z" level=info msg="Container 2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:05:57.350728 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1867225990.mount: Deactivated successfully. Sep 4 00:05:57.708237 containerd[1582]: time="2025-09-04T00:05:57.708181623Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\"" Sep 4 00:05:57.708634 containerd[1582]: time="2025-09-04T00:05:57.708602474Z" level=info msg="StartContainer for \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\"" Sep 4 00:05:57.709629 containerd[1582]: time="2025-09-04T00:05:57.709593816Z" level=info msg="connecting to shim 2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16" address="unix:///run/containerd/s/95425d4e06495a3444fb4952d37032e9277083280889ada29a9d17eb231cd942" protocol=ttrpc version=3 Sep 4 00:05:57.737739 systemd[1]: Started cri-containerd-2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16.scope - libcontainer container 2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16. Sep 4 00:05:57.771265 systemd[1]: cri-containerd-2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16.scope: Deactivated successfully. Sep 4 00:05:57.771787 containerd[1582]: time="2025-09-04T00:05:57.771728030Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" id:\"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" pid:3344 exited_at:{seconds:1756944357 nanos:771353196}" Sep 4 00:05:58.185939 containerd[1582]: time="2025-09-04T00:05:58.185624181Z" level=info msg="received exit event container_id:\"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" id:\"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" pid:3344 exited_at:{seconds:1756944357 nanos:771353196}" Sep 4 00:05:58.187840 containerd[1582]: time="2025-09-04T00:05:58.187784161Z" level=info msg="StartContainer for \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" returns successfully" Sep 4 00:05:58.209022 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16-rootfs.mount: Deactivated successfully. Sep 4 00:05:59.193441 kubelet[2730]: E0904 00:05:59.193410 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:00.198787 kubelet[2730]: E0904 00:06:00.198753 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:00.202362 containerd[1582]: time="2025-09-04T00:06:00.202302398Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 4 00:06:00.873087 containerd[1582]: time="2025-09-04T00:06:00.873030319Z" level=info msg="Container ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:06:01.209589 containerd[1582]: time="2025-09-04T00:06:01.208983294Z" level=info msg="CreateContainer within sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\"" Sep 4 00:06:01.210358 containerd[1582]: time="2025-09-04T00:06:01.210303254Z" level=info msg="StartContainer for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\"" Sep 4 00:06:01.212400 containerd[1582]: time="2025-09-04T00:06:01.212370817Z" level=info msg="connecting to shim ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2" address="unix:///run/containerd/s/95425d4e06495a3444fb4952d37032e9277083280889ada29a9d17eb231cd942" protocol=ttrpc version=3 Sep 4 00:06:01.235689 systemd[1]: Started cri-containerd-ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2.scope - libcontainer container ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2. Sep 4 00:06:01.371689 containerd[1582]: time="2025-09-04T00:06:01.371649680Z" level=info msg="StartContainer for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" returns successfully" Sep 4 00:06:01.444140 containerd[1582]: time="2025-09-04T00:06:01.444069235Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" id:\"6e9778582e839293de0c4aae55a5cc517993497041d793d1fd3d497ee39fc9ed\" pid:3420 exited_at:{seconds:1756944361 nanos:443751779}" Sep 4 00:06:01.486437 kubelet[2730]: I0904 00:06:01.486320 2730 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 4 00:06:02.151681 systemd[1]: Created slice kubepods-burstable-pod276f582e_cfe7_4333_a0fd_541b29092367.slice - libcontainer container kubepods-burstable-pod276f582e_cfe7_4333_a0fd_541b29092367.slice. Sep 4 00:06:02.206540 kubelet[2730]: E0904 00:06:02.206510 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:02.213621 kubelet[2730]: I0904 00:06:02.213579 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt67v\" (UniqueName: \"kubernetes.io/projected/276f582e-cfe7-4333-a0fd-541b29092367-kube-api-access-qt67v\") pod \"coredns-668d6bf9bc-gndv6\" (UID: \"276f582e-cfe7-4333-a0fd-541b29092367\") " pod="kube-system/coredns-668d6bf9bc-gndv6" Sep 4 00:06:02.213621 kubelet[2730]: I0904 00:06:02.213616 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/276f582e-cfe7-4333-a0fd-541b29092367-config-volume\") pod \"coredns-668d6bf9bc-gndv6\" (UID: \"276f582e-cfe7-4333-a0fd-541b29092367\") " pod="kube-system/coredns-668d6bf9bc-gndv6" Sep 4 00:06:02.594201 systemd[1]: Created slice kubepods-burstable-pod1768467c_8b64_4401_bfe8_c37479a2e9f5.slice - libcontainer container kubepods-burstable-pod1768467c_8b64_4401_bfe8_c37479a2e9f5.slice. Sep 4 00:06:02.616032 kubelet[2730]: I0904 00:06:02.615895 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt4n7\" (UniqueName: \"kubernetes.io/projected/1768467c-8b64-4401-bfe8-c37479a2e9f5-kube-api-access-nt4n7\") pod \"coredns-668d6bf9bc-kc9d5\" (UID: \"1768467c-8b64-4401-bfe8-c37479a2e9f5\") " pod="kube-system/coredns-668d6bf9bc-kc9d5" Sep 4 00:06:02.616032 kubelet[2730]: I0904 00:06:02.615989 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1768467c-8b64-4401-bfe8-c37479a2e9f5-config-volume\") pod \"coredns-668d6bf9bc-kc9d5\" (UID: \"1768467c-8b64-4401-bfe8-c37479a2e9f5\") " pod="kube-system/coredns-668d6bf9bc-kc9d5" Sep 4 00:06:02.755161 kubelet[2730]: E0904 00:06:02.755117 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:02.755906 containerd[1582]: time="2025-09-04T00:06:02.755863702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gndv6,Uid:276f582e-cfe7-4333-a0fd-541b29092367,Namespace:kube-system,Attempt:0,}" Sep 4 00:06:03.197803 kubelet[2730]: E0904 00:06:03.197745 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:03.198389 containerd[1582]: time="2025-09-04T00:06:03.198347188Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kc9d5,Uid:1768467c-8b64-4401-bfe8-c37479a2e9f5,Namespace:kube-system,Attempt:0,}" Sep 4 00:06:03.208471 kubelet[2730]: E0904 00:06:03.208426 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:03.282384 kubelet[2730]: I0904 00:06:03.282269 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-fq794" podStartSLOduration=12.469337490000001 podStartE2EDuration="30.282254393s" podCreationTimestamp="2025-09-04 00:05:33 +0000 UTC" firstStartedPulling="2025-09-04 00:05:34.321687752 +0000 UTC m=+7.049978635" lastFinishedPulling="2025-09-04 00:05:52.134604655 +0000 UTC m=+24.862895538" observedRunningTime="2025-09-04 00:06:03.282211463 +0000 UTC m=+36.010502346" watchObservedRunningTime="2025-09-04 00:06:03.282254393 +0000 UTC m=+36.010545276" Sep 4 00:06:04.210083 kubelet[2730]: E0904 00:06:04.210048 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:05.152766 systemd-networkd[1461]: cilium_host: Link UP Sep 4 00:06:05.152961 systemd-networkd[1461]: cilium_net: Link UP Sep 4 00:06:05.153220 systemd-networkd[1461]: cilium_net: Gained carrier Sep 4 00:06:05.153412 systemd-networkd[1461]: cilium_host: Gained carrier Sep 4 00:06:05.199114 systemd[1]: Started sshd@9-10.0.0.91:22-10.0.0.1:42748.service - OpenSSH per-connection server daemon (10.0.0.1:42748). Sep 4 00:06:05.224744 systemd-networkd[1461]: cilium_net: Gained IPv6LL Sep 4 00:06:05.264163 sshd[3546]: Accepted publickey for core from 10.0.0.1 port 42748 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:05.265976 sshd-session[3546]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:05.270855 systemd-logind[1561]: New session 10 of user core. Sep 4 00:06:05.278803 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 4 00:06:05.281431 systemd-networkd[1461]: cilium_vxlan: Link UP Sep 4 00:06:05.281442 systemd-networkd[1461]: cilium_vxlan: Gained carrier Sep 4 00:06:05.456459 sshd[3596]: Connection closed by 10.0.0.1 port 42748 Sep 4 00:06:05.456859 sshd-session[3546]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:05.463041 systemd[1]: sshd@9-10.0.0.91:22-10.0.0.1:42748.service: Deactivated successfully. Sep 4 00:06:05.467148 systemd[1]: session-10.scope: Deactivated successfully. Sep 4 00:06:05.468382 systemd-logind[1561]: Session 10 logged out. Waiting for processes to exit. Sep 4 00:06:05.470592 systemd-logind[1561]: Removed session 10. Sep 4 00:06:05.515597 kernel: NET: Registered PF_ALG protocol family Sep 4 00:06:05.809800 systemd-networkd[1461]: cilium_host: Gained IPv6LL Sep 4 00:06:06.220776 systemd-networkd[1461]: lxc_health: Link UP Sep 4 00:06:06.222201 systemd-networkd[1461]: lxc_health: Gained carrier Sep 4 00:06:06.235979 kubelet[2730]: E0904 00:06:06.235935 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:06.406596 kernel: eth0: renamed from tmp2b05b Sep 4 00:06:06.406669 systemd-networkd[1461]: lxc76cbfe68226f: Link UP Sep 4 00:06:06.409860 systemd-networkd[1461]: lxc76cbfe68226f: Gained carrier Sep 4 00:06:06.537600 systemd-networkd[1461]: lxce926251fa6e4: Link UP Sep 4 00:06:06.544144 kernel: eth0: renamed from tmpba11a Sep 4 00:06:06.545129 systemd-networkd[1461]: lxce926251fa6e4: Gained carrier Sep 4 00:06:07.152765 systemd-networkd[1461]: cilium_vxlan: Gained IPv6LL Sep 4 00:06:07.216283 kubelet[2730]: E0904 00:06:07.216244 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:07.472889 systemd-networkd[1461]: lxc76cbfe68226f: Gained IPv6LL Sep 4 00:06:07.920776 systemd-networkd[1461]: lxc_health: Gained IPv6LL Sep 4 00:06:08.217492 kubelet[2730]: E0904 00:06:08.217463 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:08.368888 systemd-networkd[1461]: lxce926251fa6e4: Gained IPv6LL Sep 4 00:06:10.330733 containerd[1582]: time="2025-09-04T00:06:10.330660284Z" level=info msg="connecting to shim 2b05baba5f86cf37a702a065255dd65811463b4f9b88404b637864c3d8d41bb3" address="unix:///run/containerd/s/7e94949a0cf412cd35f3f74873ef544b775a00e5ed101e6f0469c391a2994683" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:06:10.346071 containerd[1582]: time="2025-09-04T00:06:10.346010710Z" level=info msg="connecting to shim ba11a6736461e5ef3e68451d3f16f3fdaa8a8f42d521863ebec0b871e3a2507b" address="unix:///run/containerd/s/442028a37ecd9a5b769b818337e98da846d755bd69ff99a417341838b1098598" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:06:10.383872 systemd[1]: Started cri-containerd-2b05baba5f86cf37a702a065255dd65811463b4f9b88404b637864c3d8d41bb3.scope - libcontainer container 2b05baba5f86cf37a702a065255dd65811463b4f9b88404b637864c3d8d41bb3. Sep 4 00:06:10.387390 systemd[1]: Started cri-containerd-ba11a6736461e5ef3e68451d3f16f3fdaa8a8f42d521863ebec0b871e3a2507b.scope - libcontainer container ba11a6736461e5ef3e68451d3f16f3fdaa8a8f42d521863ebec0b871e3a2507b. Sep 4 00:06:10.400991 systemd-resolved[1434]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 00:06:10.403039 systemd-resolved[1434]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 00:06:10.441897 containerd[1582]: time="2025-09-04T00:06:10.441832941Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gndv6,Uid:276f582e-cfe7-4333-a0fd-541b29092367,Namespace:kube-system,Attempt:0,} returns sandbox id \"ba11a6736461e5ef3e68451d3f16f3fdaa8a8f42d521863ebec0b871e3a2507b\"" Sep 4 00:06:10.442597 kubelet[2730]: E0904 00:06:10.442501 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:10.444482 containerd[1582]: time="2025-09-04T00:06:10.444438633Z" level=info msg="CreateContainer within sandbox \"ba11a6736461e5ef3e68451d3f16f3fdaa8a8f42d521863ebec0b871e3a2507b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 00:06:10.463464 containerd[1582]: time="2025-09-04T00:06:10.463411288Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kc9d5,Uid:1768467c-8b64-4401-bfe8-c37479a2e9f5,Namespace:kube-system,Attempt:0,} returns sandbox id \"2b05baba5f86cf37a702a065255dd65811463b4f9b88404b637864c3d8d41bb3\"" Sep 4 00:06:10.464045 containerd[1582]: time="2025-09-04T00:06:10.464009010Z" level=info msg="Container 3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:06:10.465620 kubelet[2730]: E0904 00:06:10.465440 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:10.474304 systemd[1]: Started sshd@10-10.0.0.91:22-10.0.0.1:54274.service - OpenSSH per-connection server daemon (10.0.0.1:54274). Sep 4 00:06:10.476242 containerd[1582]: time="2025-09-04T00:06:10.476188184Z" level=info msg="CreateContainer within sandbox \"2b05baba5f86cf37a702a065255dd65811463b4f9b88404b637864c3d8d41bb3\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 00:06:10.476422 containerd[1582]: time="2025-09-04T00:06:10.476389732Z" level=info msg="CreateContainer within sandbox \"ba11a6736461e5ef3e68451d3f16f3fdaa8a8f42d521863ebec0b871e3a2507b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751\"" Sep 4 00:06:10.476914 containerd[1582]: time="2025-09-04T00:06:10.476875374Z" level=info msg="StartContainer for \"3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751\"" Sep 4 00:06:10.479099 containerd[1582]: time="2025-09-04T00:06:10.479051168Z" level=info msg="connecting to shim 3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751" address="unix:///run/containerd/s/442028a37ecd9a5b769b818337e98da846d755bd69ff99a417341838b1098598" protocol=ttrpc version=3 Sep 4 00:06:10.492574 containerd[1582]: time="2025-09-04T00:06:10.492341917Z" level=info msg="Container bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:06:10.502193 containerd[1582]: time="2025-09-04T00:06:10.502124261Z" level=info msg="CreateContainer within sandbox \"2b05baba5f86cf37a702a065255dd65811463b4f9b88404b637864c3d8d41bb3\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147\"" Sep 4 00:06:10.503722 systemd[1]: Started cri-containerd-3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751.scope - libcontainer container 3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751. Sep 4 00:06:10.504019 containerd[1582]: time="2025-09-04T00:06:10.503781683Z" level=info msg="StartContainer for \"bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147\"" Sep 4 00:06:10.506120 containerd[1582]: time="2025-09-04T00:06:10.506060080Z" level=info msg="connecting to shim bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147" address="unix:///run/containerd/s/7e94949a0cf412cd35f3f74873ef544b775a00e5ed101e6f0469c391a2994683" protocol=ttrpc version=3 Sep 4 00:06:10.534771 sshd[4006]: Accepted publickey for core from 10.0.0.1 port 54274 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:10.536366 sshd-session[4006]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:10.536898 systemd[1]: Started cri-containerd-bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147.scope - libcontainer container bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147. Sep 4 00:06:10.546164 systemd-logind[1561]: New session 11 of user core. Sep 4 00:06:10.552741 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 4 00:06:10.571269 containerd[1582]: time="2025-09-04T00:06:10.571160088Z" level=info msg="StartContainer for \"3b87864fad4f6a63ff71f18cbc35ca6dcb14b738823624549fd7d210f3405751\" returns successfully" Sep 4 00:06:10.582490 containerd[1582]: time="2025-09-04T00:06:10.582380341Z" level=info msg="StartContainer for \"bdbe6a0993b7dc3266f49f20741b6014098a317031da3eefabbd2f531639a147\" returns successfully" Sep 4 00:06:10.717033 sshd[4049]: Connection closed by 10.0.0.1 port 54274 Sep 4 00:06:10.717447 sshd-session[4006]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:10.722414 systemd[1]: sshd@10-10.0.0.91:22-10.0.0.1:54274.service: Deactivated successfully. Sep 4 00:06:10.724687 systemd[1]: session-11.scope: Deactivated successfully. Sep 4 00:06:10.725579 systemd-logind[1561]: Session 11 logged out. Waiting for processes to exit. Sep 4 00:06:10.727471 systemd-logind[1561]: Removed session 11. Sep 4 00:06:11.224226 kubelet[2730]: E0904 00:06:11.224161 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:11.226656 kubelet[2730]: E0904 00:06:11.226628 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:11.939382 kubelet[2730]: I0904 00:06:11.939101 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-gndv6" podStartSLOduration=38.939077143 podStartE2EDuration="38.939077143s" podCreationTimestamp="2025-09-04 00:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:06:11.589817025 +0000 UTC m=+44.318107918" watchObservedRunningTime="2025-09-04 00:06:11.939077143 +0000 UTC m=+44.667368026" Sep 4 00:06:11.953024 kubelet[2730]: I0904 00:06:11.952943 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-kc9d5" podStartSLOduration=38.952898238 podStartE2EDuration="38.952898238s" podCreationTimestamp="2025-09-04 00:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:06:11.951477411 +0000 UTC m=+44.679768294" watchObservedRunningTime="2025-09-04 00:06:11.952898238 +0000 UTC m=+44.681189151" Sep 4 00:06:12.228443 kubelet[2730]: E0904 00:06:12.228394 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:12.228642 kubelet[2730]: E0904 00:06:12.228394 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:13.230382 kubelet[2730]: E0904 00:06:13.230237 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:13.231187 kubelet[2730]: E0904 00:06:13.231160 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:15.737060 systemd[1]: Started sshd@11-10.0.0.91:22-10.0.0.1:54276.service - OpenSSH per-connection server daemon (10.0.0.1:54276). Sep 4 00:06:15.777382 sshd[4099]: Accepted publickey for core from 10.0.0.1 port 54276 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:15.779077 sshd-session[4099]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:15.784157 systemd-logind[1561]: New session 12 of user core. Sep 4 00:06:15.793692 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 4 00:06:15.920170 sshd[4101]: Connection closed by 10.0.0.1 port 54276 Sep 4 00:06:15.920542 sshd-session[4099]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:15.924861 systemd[1]: sshd@11-10.0.0.91:22-10.0.0.1:54276.service: Deactivated successfully. Sep 4 00:06:15.926851 systemd[1]: session-12.scope: Deactivated successfully. Sep 4 00:06:15.927698 systemd-logind[1561]: Session 12 logged out. Waiting for processes to exit. Sep 4 00:06:15.929436 systemd-logind[1561]: Removed session 12. Sep 4 00:06:20.944812 systemd[1]: Started sshd@12-10.0.0.91:22-10.0.0.1:35598.service - OpenSSH per-connection server daemon (10.0.0.1:35598). Sep 4 00:06:21.012261 sshd[4117]: Accepted publickey for core from 10.0.0.1 port 35598 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:21.014034 sshd-session[4117]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:21.019104 systemd-logind[1561]: New session 13 of user core. Sep 4 00:06:21.036834 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 4 00:06:21.252303 sshd[4119]: Connection closed by 10.0.0.1 port 35598 Sep 4 00:06:21.252526 sshd-session[4117]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:21.256115 systemd[1]: sshd@12-10.0.0.91:22-10.0.0.1:35598.service: Deactivated successfully. Sep 4 00:06:21.258353 systemd[1]: session-13.scope: Deactivated successfully. Sep 4 00:06:21.260626 systemd-logind[1561]: Session 13 logged out. Waiting for processes to exit. Sep 4 00:06:21.262250 systemd-logind[1561]: Removed session 13. Sep 4 00:06:26.270907 systemd[1]: Started sshd@13-10.0.0.91:22-10.0.0.1:35602.service - OpenSSH per-connection server daemon (10.0.0.1:35602). Sep 4 00:06:26.329448 sshd[4133]: Accepted publickey for core from 10.0.0.1 port 35602 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:26.331289 sshd-session[4133]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:26.336163 systemd-logind[1561]: New session 14 of user core. Sep 4 00:06:26.346679 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 4 00:06:26.468066 sshd[4135]: Connection closed by 10.0.0.1 port 35602 Sep 4 00:06:26.468443 sshd-session[4133]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:26.473818 systemd[1]: sshd@13-10.0.0.91:22-10.0.0.1:35602.service: Deactivated successfully. Sep 4 00:06:26.476266 systemd[1]: session-14.scope: Deactivated successfully. Sep 4 00:06:26.477118 systemd-logind[1561]: Session 14 logged out. Waiting for processes to exit. Sep 4 00:06:26.478663 systemd-logind[1561]: Removed session 14. Sep 4 00:06:31.486058 systemd[1]: Started sshd@14-10.0.0.91:22-10.0.0.1:37090.service - OpenSSH per-connection server daemon (10.0.0.1:37090). Sep 4 00:06:31.542327 sshd[4152]: Accepted publickey for core from 10.0.0.1 port 37090 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:31.544231 sshd-session[4152]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:31.549511 systemd-logind[1561]: New session 15 of user core. Sep 4 00:06:31.557704 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 4 00:06:31.800775 sshd[4154]: Connection closed by 10.0.0.1 port 37090 Sep 4 00:06:31.801017 sshd-session[4152]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:31.817448 systemd[1]: sshd@14-10.0.0.91:22-10.0.0.1:37090.service: Deactivated successfully. Sep 4 00:06:31.819983 systemd[1]: session-15.scope: Deactivated successfully. Sep 4 00:06:31.820853 systemd-logind[1561]: Session 15 logged out. Waiting for processes to exit. Sep 4 00:06:31.824638 systemd[1]: Started sshd@15-10.0.0.91:22-10.0.0.1:37096.service - OpenSSH per-connection server daemon (10.0.0.1:37096). Sep 4 00:06:31.825331 systemd-logind[1561]: Removed session 15. Sep 4 00:06:31.869631 sshd[4169]: Accepted publickey for core from 10.0.0.1 port 37096 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:31.871288 sshd-session[4169]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:31.876024 systemd-logind[1561]: New session 16 of user core. Sep 4 00:06:31.890683 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 4 00:06:32.172469 sshd[4171]: Connection closed by 10.0.0.1 port 37096 Sep 4 00:06:32.172826 sshd-session[4169]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:32.185007 systemd[1]: sshd@15-10.0.0.91:22-10.0.0.1:37096.service: Deactivated successfully. Sep 4 00:06:32.187791 systemd[1]: session-16.scope: Deactivated successfully. Sep 4 00:06:32.188745 systemd-logind[1561]: Session 16 logged out. Waiting for processes to exit. Sep 4 00:06:32.192459 systemd[1]: Started sshd@16-10.0.0.91:22-10.0.0.1:37110.service - OpenSSH per-connection server daemon (10.0.0.1:37110). Sep 4 00:06:32.193571 systemd-logind[1561]: Removed session 16. Sep 4 00:06:32.244939 sshd[4189]: Accepted publickey for core from 10.0.0.1 port 37110 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:32.246735 sshd-session[4189]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:32.252031 systemd-logind[1561]: New session 17 of user core. Sep 4 00:06:32.265759 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 4 00:06:32.395757 sshd[4191]: Connection closed by 10.0.0.1 port 37110 Sep 4 00:06:32.396129 sshd-session[4189]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:32.401980 systemd[1]: sshd@16-10.0.0.91:22-10.0.0.1:37110.service: Deactivated successfully. Sep 4 00:06:32.404381 systemd[1]: session-17.scope: Deactivated successfully. Sep 4 00:06:32.405509 systemd-logind[1561]: Session 17 logged out. Waiting for processes to exit. Sep 4 00:06:32.407081 systemd-logind[1561]: Removed session 17. Sep 4 00:06:37.408672 systemd[1]: Started sshd@17-10.0.0.91:22-10.0.0.1:37112.service - OpenSSH per-connection server daemon (10.0.0.1:37112). Sep 4 00:06:37.466206 sshd[4207]: Accepted publickey for core from 10.0.0.1 port 37112 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:37.468233 sshd-session[4207]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:37.474883 systemd-logind[1561]: New session 18 of user core. Sep 4 00:06:37.489854 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 4 00:06:37.611037 sshd[4209]: Connection closed by 10.0.0.1 port 37112 Sep 4 00:06:37.611434 sshd-session[4207]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:37.615314 systemd[1]: sshd@17-10.0.0.91:22-10.0.0.1:37112.service: Deactivated successfully. Sep 4 00:06:37.617420 systemd[1]: session-18.scope: Deactivated successfully. Sep 4 00:06:37.620102 systemd-logind[1561]: Session 18 logged out. Waiting for processes to exit. Sep 4 00:06:37.621345 systemd-logind[1561]: Removed session 18. Sep 4 00:06:37.695429 kubelet[2730]: E0904 00:06:37.695376 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:40.694826 kubelet[2730]: E0904 00:06:40.694682 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:42.635670 systemd[1]: Started sshd@18-10.0.0.91:22-10.0.0.1:52168.service - OpenSSH per-connection server daemon (10.0.0.1:52168). Sep 4 00:06:42.690648 sshd[4223]: Accepted publickey for core from 10.0.0.1 port 52168 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:42.692258 sshd-session[4223]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:42.697299 systemd-logind[1561]: New session 19 of user core. Sep 4 00:06:42.712697 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 4 00:06:42.841572 sshd[4225]: Connection closed by 10.0.0.1 port 52168 Sep 4 00:06:42.842775 sshd-session[4223]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:42.851984 systemd[1]: sshd@18-10.0.0.91:22-10.0.0.1:52168.service: Deactivated successfully. Sep 4 00:06:42.854018 systemd[1]: session-19.scope: Deactivated successfully. Sep 4 00:06:42.856390 systemd-logind[1561]: Session 19 logged out. Waiting for processes to exit. Sep 4 00:06:42.857968 systemd-logind[1561]: Removed session 19. Sep 4 00:06:47.695459 kubelet[2730]: E0904 00:06:47.695406 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:06:47.859782 systemd[1]: Started sshd@19-10.0.0.91:22-10.0.0.1:52170.service - OpenSSH per-connection server daemon (10.0.0.1:52170). Sep 4 00:06:47.917467 sshd[4238]: Accepted publickey for core from 10.0.0.1 port 52170 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:47.919869 sshd-session[4238]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:47.925806 systemd-logind[1561]: New session 20 of user core. Sep 4 00:06:47.935726 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 4 00:06:48.046726 sshd[4240]: Connection closed by 10.0.0.1 port 52170 Sep 4 00:06:48.046963 sshd-session[4238]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:48.056812 systemd[1]: sshd@19-10.0.0.91:22-10.0.0.1:52170.service: Deactivated successfully. Sep 4 00:06:48.059299 systemd[1]: session-20.scope: Deactivated successfully. Sep 4 00:06:48.060340 systemd-logind[1561]: Session 20 logged out. Waiting for processes to exit. Sep 4 00:06:48.064025 systemd[1]: Started sshd@20-10.0.0.91:22-10.0.0.1:52174.service - OpenSSH per-connection server daemon (10.0.0.1:52174). Sep 4 00:06:48.065394 systemd-logind[1561]: Removed session 20. Sep 4 00:06:48.119160 sshd[4253]: Accepted publickey for core from 10.0.0.1 port 52174 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:48.120914 sshd-session[4253]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:48.125815 systemd-logind[1561]: New session 21 of user core. Sep 4 00:06:48.136777 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 4 00:06:48.759574 sshd[4255]: Connection closed by 10.0.0.1 port 52174 Sep 4 00:06:48.760122 sshd-session[4253]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:48.779076 systemd[1]: sshd@20-10.0.0.91:22-10.0.0.1:52174.service: Deactivated successfully. Sep 4 00:06:48.781508 systemd[1]: session-21.scope: Deactivated successfully. Sep 4 00:06:48.782499 systemd-logind[1561]: Session 21 logged out. Waiting for processes to exit. Sep 4 00:06:48.785821 systemd[1]: Started sshd@21-10.0.0.91:22-10.0.0.1:52178.service - OpenSSH per-connection server daemon (10.0.0.1:52178). Sep 4 00:06:48.786792 systemd-logind[1561]: Removed session 21. Sep 4 00:06:48.839357 sshd[4266]: Accepted publickey for core from 10.0.0.1 port 52178 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:48.840882 sshd-session[4266]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:48.845827 systemd-logind[1561]: New session 22 of user core. Sep 4 00:06:48.853697 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 4 00:06:49.452413 sshd[4268]: Connection closed by 10.0.0.1 port 52178 Sep 4 00:06:49.452918 sshd-session[4266]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:49.465099 systemd[1]: sshd@21-10.0.0.91:22-10.0.0.1:52178.service: Deactivated successfully. Sep 4 00:06:49.468439 systemd[1]: session-22.scope: Deactivated successfully. Sep 4 00:06:49.469746 systemd-logind[1561]: Session 22 logged out. Waiting for processes to exit. Sep 4 00:06:49.474465 systemd[1]: Started sshd@22-10.0.0.91:22-10.0.0.1:52182.service - OpenSSH per-connection server daemon (10.0.0.1:52182). Sep 4 00:06:49.477161 systemd-logind[1561]: Removed session 22. Sep 4 00:06:49.527509 sshd[4290]: Accepted publickey for core from 10.0.0.1 port 52182 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:49.529392 sshd-session[4290]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:49.534638 systemd-logind[1561]: New session 23 of user core. Sep 4 00:06:49.545742 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 4 00:06:49.818294 sshd[4292]: Connection closed by 10.0.0.1 port 52182 Sep 4 00:06:49.818607 sshd-session[4290]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:49.833300 systemd[1]: sshd@22-10.0.0.91:22-10.0.0.1:52182.service: Deactivated successfully. Sep 4 00:06:49.835974 systemd[1]: session-23.scope: Deactivated successfully. Sep 4 00:06:49.837230 systemd-logind[1561]: Session 23 logged out. Waiting for processes to exit. Sep 4 00:06:49.840736 systemd-logind[1561]: Removed session 23. Sep 4 00:06:49.842801 systemd[1]: Started sshd@23-10.0.0.91:22-10.0.0.1:52194.service - OpenSSH per-connection server daemon (10.0.0.1:52194). Sep 4 00:06:49.891765 sshd[4303]: Accepted publickey for core from 10.0.0.1 port 52194 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:49.893334 sshd-session[4303]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:49.898656 systemd-logind[1561]: New session 24 of user core. Sep 4 00:06:49.905684 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 4 00:06:50.026372 sshd[4305]: Connection closed by 10.0.0.1 port 52194 Sep 4 00:06:50.026756 sshd-session[4303]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:50.032263 systemd[1]: sshd@23-10.0.0.91:22-10.0.0.1:52194.service: Deactivated successfully. Sep 4 00:06:50.034871 systemd[1]: session-24.scope: Deactivated successfully. Sep 4 00:06:50.036248 systemd-logind[1561]: Session 24 logged out. Waiting for processes to exit. Sep 4 00:06:50.037650 systemd-logind[1561]: Removed session 24. Sep 4 00:06:55.041930 systemd[1]: Started sshd@24-10.0.0.91:22-10.0.0.1:54716.service - OpenSSH per-connection server daemon (10.0.0.1:54716). Sep 4 00:06:55.104099 sshd[4318]: Accepted publickey for core from 10.0.0.1 port 54716 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:06:55.105842 sshd-session[4318]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:06:55.111050 systemd-logind[1561]: New session 25 of user core. Sep 4 00:06:55.120853 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 4 00:06:55.289301 sshd[4320]: Connection closed by 10.0.0.1 port 54716 Sep 4 00:06:55.289716 sshd-session[4318]: pam_unix(sshd:session): session closed for user core Sep 4 00:06:55.294569 systemd[1]: sshd@24-10.0.0.91:22-10.0.0.1:54716.service: Deactivated successfully. Sep 4 00:06:55.297343 systemd[1]: session-25.scope: Deactivated successfully. Sep 4 00:06:55.298267 systemd-logind[1561]: Session 25 logged out. Waiting for processes to exit. Sep 4 00:06:55.300578 systemd-logind[1561]: Removed session 25. Sep 4 00:07:00.316145 systemd[1]: Started sshd@25-10.0.0.91:22-10.0.0.1:41468.service - OpenSSH per-connection server daemon (10.0.0.1:41468). Sep 4 00:07:00.372606 sshd[4335]: Accepted publickey for core from 10.0.0.1 port 41468 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:00.374979 sshd-session[4335]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:00.380998 systemd-logind[1561]: New session 26 of user core. Sep 4 00:07:00.387767 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 4 00:07:00.507442 sshd[4337]: Connection closed by 10.0.0.1 port 41468 Sep 4 00:07:00.507874 sshd-session[4335]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:00.513590 systemd[1]: sshd@25-10.0.0.91:22-10.0.0.1:41468.service: Deactivated successfully. Sep 4 00:07:00.516684 systemd[1]: session-26.scope: Deactivated successfully. Sep 4 00:07:00.517620 systemd-logind[1561]: Session 26 logged out. Waiting for processes to exit. Sep 4 00:07:00.519427 systemd-logind[1561]: Removed session 26. Sep 4 00:07:05.520967 systemd[1]: Started sshd@26-10.0.0.91:22-10.0.0.1:41472.service - OpenSSH per-connection server daemon (10.0.0.1:41472). Sep 4 00:07:05.568086 sshd[4353]: Accepted publickey for core from 10.0.0.1 port 41472 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:05.569724 sshd-session[4353]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:05.574883 systemd-logind[1561]: New session 27 of user core. Sep 4 00:07:05.579787 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 4 00:07:05.692383 sshd[4355]: Connection closed by 10.0.0.1 port 41472 Sep 4 00:07:05.692730 sshd-session[4353]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:05.697292 systemd[1]: sshd@26-10.0.0.91:22-10.0.0.1:41472.service: Deactivated successfully. Sep 4 00:07:05.699357 systemd[1]: session-27.scope: Deactivated successfully. Sep 4 00:07:05.700356 systemd-logind[1561]: Session 27 logged out. Waiting for processes to exit. Sep 4 00:07:05.701808 systemd-logind[1561]: Removed session 27. Sep 4 00:07:06.695157 kubelet[2730]: E0904 00:07:06.695099 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:07.698439 kubelet[2730]: E0904 00:07:07.698399 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:10.704751 systemd[1]: Started sshd@27-10.0.0.91:22-10.0.0.1:52904.service - OpenSSH per-connection server daemon (10.0.0.1:52904). Sep 4 00:07:10.744363 sshd[4368]: Accepted publickey for core from 10.0.0.1 port 52904 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:10.746149 sshd-session[4368]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:10.750659 systemd-logind[1561]: New session 28 of user core. Sep 4 00:07:10.758692 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 4 00:07:10.865719 sshd[4370]: Connection closed by 10.0.0.1 port 52904 Sep 4 00:07:10.866071 sshd-session[4368]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:10.875492 systemd[1]: sshd@27-10.0.0.91:22-10.0.0.1:52904.service: Deactivated successfully. Sep 4 00:07:10.877648 systemd[1]: session-28.scope: Deactivated successfully. Sep 4 00:07:10.878661 systemd-logind[1561]: Session 28 logged out. Waiting for processes to exit. Sep 4 00:07:10.881880 systemd[1]: Started sshd@28-10.0.0.91:22-10.0.0.1:52914.service - OpenSSH per-connection server daemon (10.0.0.1:52914). Sep 4 00:07:10.882713 systemd-logind[1561]: Removed session 28. Sep 4 00:07:10.932414 sshd[4383]: Accepted publickey for core from 10.0.0.1 port 52914 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:10.933950 sshd-session[4383]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:10.938923 systemd-logind[1561]: New session 29 of user core. Sep 4 00:07:10.946749 systemd[1]: Started session-29.scope - Session 29 of User core. Sep 4 00:07:12.657824 containerd[1582]: time="2025-09-04T00:07:12.657754556Z" level=info msg="StopContainer for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" with timeout 30 (s)" Sep 4 00:07:12.660225 containerd[1582]: time="2025-09-04T00:07:12.660131858Z" level=info msg="Stop container \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" with signal terminated" Sep 4 00:07:12.681485 systemd[1]: cri-containerd-6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f.scope: Deactivated successfully. Sep 4 00:07:12.685125 containerd[1582]: time="2025-09-04T00:07:12.685077380Z" level=info msg="received exit event container_id:\"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" id:\"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" pid:3147 exited_at:{seconds:1756944432 nanos:684720484}" Sep 4 00:07:12.685991 containerd[1582]: time="2025-09-04T00:07:12.685884358Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" id:\"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" pid:3147 exited_at:{seconds:1756944432 nanos:684720484}" Sep 4 00:07:12.713701 containerd[1582]: time="2025-09-04T00:07:12.713638048Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 00:07:12.719986 containerd[1582]: time="2025-09-04T00:07:12.719919072Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" id:\"98b18b95f39557f1a5922c29db342f689f58509ce0dddcdce5d353ee8ec5b724\" pid:4414 exited_at:{seconds:1756944432 nanos:714804567}" Sep 4 00:07:12.721247 containerd[1582]: time="2025-09-04T00:07:12.721196761Z" level=info msg="StopContainer for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" with timeout 2 (s)" Sep 4 00:07:12.723109 containerd[1582]: time="2025-09-04T00:07:12.721708982Z" level=info msg="Stop container \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" with signal terminated" Sep 4 00:07:12.721567 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f-rootfs.mount: Deactivated successfully. Sep 4 00:07:12.732249 systemd-networkd[1461]: lxc_health: Link DOWN Sep 4 00:07:12.732261 systemd-networkd[1461]: lxc_health: Lost carrier Sep 4 00:07:12.740381 containerd[1582]: time="2025-09-04T00:07:12.740335346Z" level=info msg="StopContainer for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" returns successfully" Sep 4 00:07:12.750288 containerd[1582]: time="2025-09-04T00:07:12.750231758Z" level=info msg="StopPodSandbox for \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\"" Sep 4 00:07:12.750459 containerd[1582]: time="2025-09-04T00:07:12.750338159Z" level=info msg="Container to stop \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 00:07:12.756035 kubelet[2730]: E0904 00:07:12.755980 2730 kubelet.go:3002] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 4 00:07:12.757133 systemd[1]: cri-containerd-ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2.scope: Deactivated successfully. Sep 4 00:07:12.757635 systemd[1]: cri-containerd-ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2.scope: Consumed 6.980s CPU time, 122.9M memory peak, 156K read from disk, 13.3M written to disk. Sep 4 00:07:12.759405 containerd[1582]: time="2025-09-04T00:07:12.759352230Z" level=info msg="received exit event container_id:\"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" id:\"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" pid:3381 exited_at:{seconds:1756944432 nanos:759070537}" Sep 4 00:07:12.759636 containerd[1582]: time="2025-09-04T00:07:12.759605399Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" id:\"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" pid:3381 exited_at:{seconds:1756944432 nanos:759070537}" Sep 4 00:07:12.762844 systemd[1]: cri-containerd-3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6.scope: Deactivated successfully. Sep 4 00:07:12.770126 containerd[1582]: time="2025-09-04T00:07:12.769998631Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" id:\"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" pid:2858 exit_status:137 exited_at:{seconds:1756944432 nanos:769574428}" Sep 4 00:07:12.792002 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2-rootfs.mount: Deactivated successfully. Sep 4 00:07:12.809295 containerd[1582]: time="2025-09-04T00:07:12.809241319Z" level=info msg="StopContainer for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" returns successfully" Sep 4 00:07:12.809466 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6-rootfs.mount: Deactivated successfully. Sep 4 00:07:12.810212 containerd[1582]: time="2025-09-04T00:07:12.810075939Z" level=info msg="StopPodSandbox for \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\"" Sep 4 00:07:12.810419 containerd[1582]: time="2025-09-04T00:07:12.810394051Z" level=info msg="Container to stop \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 00:07:12.810584 containerd[1582]: time="2025-09-04T00:07:12.810482840Z" level=info msg="Container to stop \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 00:07:12.810584 containerd[1582]: time="2025-09-04T00:07:12.810526733Z" level=info msg="Container to stop \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 00:07:12.810728 containerd[1582]: time="2025-09-04T00:07:12.810538174Z" level=info msg="Container to stop \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 00:07:12.810728 containerd[1582]: time="2025-09-04T00:07:12.810682177Z" level=info msg="Container to stop \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 00:07:12.813565 containerd[1582]: time="2025-09-04T00:07:12.813211958Z" level=info msg="shim disconnected" id=3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6 namespace=k8s.io Sep 4 00:07:12.813565 containerd[1582]: time="2025-09-04T00:07:12.813237767Z" level=warning msg="cleaning up after shim disconnected" id=3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6 namespace=k8s.io Sep 4 00:07:12.818711 systemd[1]: cri-containerd-693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b.scope: Deactivated successfully. Sep 4 00:07:12.844561 containerd[1582]: time="2025-09-04T00:07:12.813247365Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 00:07:12.847925 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b-rootfs.mount: Deactivated successfully. Sep 4 00:07:12.861511 containerd[1582]: time="2025-09-04T00:07:12.861281350Z" level=info msg="shim disconnected" id=693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b namespace=k8s.io Sep 4 00:07:12.861511 containerd[1582]: time="2025-09-04T00:07:12.861318331Z" level=warning msg="cleaning up after shim disconnected" id=693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b namespace=k8s.io Sep 4 00:07:12.861511 containerd[1582]: time="2025-09-04T00:07:12.861327277Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 00:07:12.879814 containerd[1582]: time="2025-09-04T00:07:12.879762119Z" level=info msg="TaskExit event in podsandbox handler container_id:\"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" id:\"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" pid:2940 exit_status:137 exited_at:{seconds:1756944432 nanos:820504658}" Sep 4 00:07:12.884451 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b-shm.mount: Deactivated successfully. Sep 4 00:07:12.884683 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6-shm.mount: Deactivated successfully. Sep 4 00:07:12.902487 containerd[1582]: time="2025-09-04T00:07:12.902415840Z" level=info msg="received exit event sandbox_id:\"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" exit_status:137 exited_at:{seconds:1756944432 nanos:769574428}" Sep 4 00:07:12.902762 containerd[1582]: time="2025-09-04T00:07:12.902723102Z" level=info msg="received exit event sandbox_id:\"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" exit_status:137 exited_at:{seconds:1756944432 nanos:820504658}" Sep 4 00:07:12.907504 containerd[1582]: time="2025-09-04T00:07:12.907444013Z" level=info msg="TearDown network for sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" successfully" Sep 4 00:07:12.907504 containerd[1582]: time="2025-09-04T00:07:12.907501451Z" level=info msg="StopPodSandbox for \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" returns successfully" Sep 4 00:07:12.907608 containerd[1582]: time="2025-09-04T00:07:12.907526148Z" level=info msg="TearDown network for sandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" successfully" Sep 4 00:07:12.907608 containerd[1582]: time="2025-09-04T00:07:12.907571293Z" level=info msg="StopPodSandbox for \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" returns successfully" Sep 4 00:07:13.076778 kubelet[2730]: I0904 00:07:13.076670 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-run\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.076778 kubelet[2730]: I0904 00:07:13.076748 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ln5n\" (UniqueName: \"kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-kube-api-access-6ln5n\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.076778 kubelet[2730]: I0904 00:07:13.076772 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-xtables-lock\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.076778 kubelet[2730]: I0904 00:07:13.076791 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-hostproc\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077143 kubelet[2730]: I0904 00:07:13.076817 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-hubble-tls\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077143 kubelet[2730]: I0904 00:07:13.076812 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.077143 kubelet[2730]: I0904 00:07:13.076882 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.077143 kubelet[2730]: I0904 00:07:13.076837 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-cgroup\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077143 kubelet[2730]: I0904 00:07:13.076887 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.077309 kubelet[2730]: I0904 00:07:13.076919 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-hostproc" (OuterVolumeSpecName: "hostproc") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.077309 kubelet[2730]: I0904 00:07:13.076938 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/a61e66f7-5c59-4052-9364-69f00bf363d7-clustermesh-secrets\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077309 kubelet[2730]: I0904 00:07:13.076988 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-config-path\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077309 kubelet[2730]: I0904 00:07:13.077010 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-kernel\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077309 kubelet[2730]: I0904 00:07:13.077039 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-bpf-maps\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077309 kubelet[2730]: I0904 00:07:13.077065 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-net\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077086 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsnds\" (UniqueName: \"kubernetes.io/projected/84438b01-62a9-44cf-9db8-0c9d2cca272c-kube-api-access-zsnds\") pod \"84438b01-62a9-44cf-9db8-0c9d2cca272c\" (UID: \"84438b01-62a9-44cf-9db8-0c9d2cca272c\") " Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077106 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-lib-modules\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077118 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cni-path\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077134 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/84438b01-62a9-44cf-9db8-0c9d2cca272c-cilium-config-path\") pod \"84438b01-62a9-44cf-9db8-0c9d2cca272c\" (UID: \"84438b01-62a9-44cf-9db8-0c9d2cca272c\") " Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077147 2730 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-etc-cni-netd\") pod \"a61e66f7-5c59-4052-9364-69f00bf363d7\" (UID: \"a61e66f7-5c59-4052-9364-69f00bf363d7\") " Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077194 2730 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.077504 kubelet[2730]: I0904 00:07:13.077217 2730 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.077874 kubelet[2730]: I0904 00:07:13.077225 2730 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.077874 kubelet[2730]: I0904 00:07:13.077233 2730 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.077874 kubelet[2730]: I0904 00:07:13.077256 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.077874 kubelet[2730]: I0904 00:07:13.077276 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.077874 kubelet[2730]: I0904 00:07:13.077293 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.078066 kubelet[2730]: I0904 00:07:13.077308 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.082274 kubelet[2730]: I0904 00:07:13.081639 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cni-path" (OuterVolumeSpecName: "cni-path") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.082274 kubelet[2730]: I0904 00:07:13.081639 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 4 00:07:13.082274 kubelet[2730]: I0904 00:07:13.081719 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 4 00:07:13.082274 kubelet[2730]: I0904 00:07:13.081729 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 4 00:07:13.082643 kubelet[2730]: I0904 00:07:13.082526 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84438b01-62a9-44cf-9db8-0c9d2cca272c-kube-api-access-zsnds" (OuterVolumeSpecName: "kube-api-access-zsnds") pod "84438b01-62a9-44cf-9db8-0c9d2cca272c" (UID: "84438b01-62a9-44cf-9db8-0c9d2cca272c"). InnerVolumeSpecName "kube-api-access-zsnds". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 4 00:07:13.083747 kubelet[2730]: I0904 00:07:13.083698 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-kube-api-access-6ln5n" (OuterVolumeSpecName: "kube-api-access-6ln5n") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "kube-api-access-6ln5n". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 4 00:07:13.084942 kubelet[2730]: I0904 00:07:13.084903 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a61e66f7-5c59-4052-9364-69f00bf363d7-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "a61e66f7-5c59-4052-9364-69f00bf363d7" (UID: "a61e66f7-5c59-4052-9364-69f00bf363d7"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 4 00:07:13.085620 kubelet[2730]: I0904 00:07:13.085590 2730 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84438b01-62a9-44cf-9db8-0c9d2cca272c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "84438b01-62a9-44cf-9db8-0c9d2cca272c" (UID: "84438b01-62a9-44cf-9db8-0c9d2cca272c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178045 2730 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178091 2730 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/84438b01-62a9-44cf-9db8-0c9d2cca272c-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178105 2730 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178121 2730 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178131 2730 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6ln5n\" (UniqueName: \"kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-kube-api-access-6ln5n\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178145 2730 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/a61e66f7-5c59-4052-9364-69f00bf363d7-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178114 kubelet[2730]: I0904 00:07:13.178156 2730 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/a61e66f7-5c59-4052-9364-69f00bf363d7-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178625 kubelet[2730]: I0904 00:07:13.178167 2730 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a61e66f7-5c59-4052-9364-69f00bf363d7-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178625 kubelet[2730]: I0904 00:07:13.178178 2730 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178625 kubelet[2730]: I0904 00:07:13.178189 2730 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178625 kubelet[2730]: I0904 00:07:13.178199 2730 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/a61e66f7-5c59-4052-9364-69f00bf363d7-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.178625 kubelet[2730]: I0904 00:07:13.178209 2730 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zsnds\" (UniqueName: \"kubernetes.io/projected/84438b01-62a9-44cf-9db8-0c9d2cca272c-kube-api-access-zsnds\") on node \"localhost\" DevicePath \"\"" Sep 4 00:07:13.366276 kubelet[2730]: I0904 00:07:13.366090 2730 scope.go:117] "RemoveContainer" containerID="ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2" Sep 4 00:07:13.371635 containerd[1582]: time="2025-09-04T00:07:13.370872934Z" level=info msg="RemoveContainer for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\"" Sep 4 00:07:13.377524 systemd[1]: Removed slice kubepods-besteffort-pod84438b01_62a9_44cf_9db8_0c9d2cca272c.slice - libcontainer container kubepods-besteffort-pod84438b01_62a9_44cf_9db8_0c9d2cca272c.slice. Sep 4 00:07:13.379489 systemd[1]: Removed slice kubepods-burstable-poda61e66f7_5c59_4052_9364_69f00bf363d7.slice - libcontainer container kubepods-burstable-poda61e66f7_5c59_4052_9364_69f00bf363d7.slice. Sep 4 00:07:13.379637 systemd[1]: kubepods-burstable-poda61e66f7_5c59_4052_9364_69f00bf363d7.slice: Consumed 7.088s CPU time, 123.2M memory peak, 160K read from disk, 13.3M written to disk. Sep 4 00:07:13.381712 containerd[1582]: time="2025-09-04T00:07:13.381680939Z" level=info msg="RemoveContainer for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" returns successfully" Sep 4 00:07:13.382021 kubelet[2730]: I0904 00:07:13.381911 2730 scope.go:117] "RemoveContainer" containerID="2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16" Sep 4 00:07:13.383314 containerd[1582]: time="2025-09-04T00:07:13.383277562Z" level=info msg="RemoveContainer for \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\"" Sep 4 00:07:13.394332 containerd[1582]: time="2025-09-04T00:07:13.394271679Z" level=info msg="RemoveContainer for \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" returns successfully" Sep 4 00:07:13.394513 kubelet[2730]: I0904 00:07:13.394484 2730 scope.go:117] "RemoveContainer" containerID="673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e" Sep 4 00:07:13.397165 containerd[1582]: time="2025-09-04T00:07:13.397116305Z" level=info msg="RemoveContainer for \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\"" Sep 4 00:07:13.403640 containerd[1582]: time="2025-09-04T00:07:13.403592567Z" level=info msg="RemoveContainer for \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" returns successfully" Sep 4 00:07:13.404666 kubelet[2730]: I0904 00:07:13.403795 2730 scope.go:117] "RemoveContainer" containerID="1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e" Sep 4 00:07:13.406024 containerd[1582]: time="2025-09-04T00:07:13.405938960Z" level=info msg="RemoveContainer for \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\"" Sep 4 00:07:13.410401 containerd[1582]: time="2025-09-04T00:07:13.410356664Z" level=info msg="RemoveContainer for \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" returns successfully" Sep 4 00:07:13.410562 kubelet[2730]: I0904 00:07:13.410523 2730 scope.go:117] "RemoveContainer" containerID="a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa" Sep 4 00:07:13.412439 containerd[1582]: time="2025-09-04T00:07:13.412393721Z" level=info msg="RemoveContainer for \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\"" Sep 4 00:07:13.416490 containerd[1582]: time="2025-09-04T00:07:13.416441916Z" level=info msg="RemoveContainer for \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" returns successfully" Sep 4 00:07:13.416681 kubelet[2730]: I0904 00:07:13.416623 2730 scope.go:117] "RemoveContainer" containerID="ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2" Sep 4 00:07:13.416930 containerd[1582]: time="2025-09-04T00:07:13.416871350Z" level=error msg="ContainerStatus for \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\": not found" Sep 4 00:07:13.417102 kubelet[2730]: E0904 00:07:13.417074 2730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\": not found" containerID="ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2" Sep 4 00:07:13.417179 kubelet[2730]: I0904 00:07:13.417101 2730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2"} err="failed to get container status \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\": rpc error: code = NotFound desc = an error occurred when try to find container \"ae949b1fe90ffb43061a9b07c48258fdbc720a1b1664c2e43ae5d0f25e6512e2\": not found" Sep 4 00:07:13.417179 kubelet[2730]: I0904 00:07:13.417174 2730 scope.go:117] "RemoveContainer" containerID="2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16" Sep 4 00:07:13.417346 containerd[1582]: time="2025-09-04T00:07:13.417318757Z" level=error msg="ContainerStatus for \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\": not found" Sep 4 00:07:13.417482 kubelet[2730]: E0904 00:07:13.417455 2730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\": not found" containerID="2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16" Sep 4 00:07:13.417523 kubelet[2730]: I0904 00:07:13.417487 2730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16"} err="failed to get container status \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\": rpc error: code = NotFound desc = an error occurred when try to find container \"2d2a0ffba29289c246c2f6df7ec6edf38b2d298cc20c007eaa3bad5c13b82b16\": not found" Sep 4 00:07:13.417523 kubelet[2730]: I0904 00:07:13.417505 2730 scope.go:117] "RemoveContainer" containerID="673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e" Sep 4 00:07:13.417702 containerd[1582]: time="2025-09-04T00:07:13.417669781Z" level=error msg="ContainerStatus for \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\": not found" Sep 4 00:07:13.417772 kubelet[2730]: E0904 00:07:13.417749 2730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\": not found" containerID="673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e" Sep 4 00:07:13.417807 kubelet[2730]: I0904 00:07:13.417774 2730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e"} err="failed to get container status \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\": rpc error: code = NotFound desc = an error occurred when try to find container \"673ac07ad69b7dfa820987011c95c1aff717f0433d9fdd774e679ae087c8899e\": not found" Sep 4 00:07:13.417807 kubelet[2730]: I0904 00:07:13.417792 2730 scope.go:117] "RemoveContainer" containerID="1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e" Sep 4 00:07:13.417972 containerd[1582]: time="2025-09-04T00:07:13.417930455Z" level=error msg="ContainerStatus for \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\": not found" Sep 4 00:07:13.418072 kubelet[2730]: E0904 00:07:13.418051 2730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\": not found" containerID="1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e" Sep 4 00:07:13.418101 kubelet[2730]: I0904 00:07:13.418071 2730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e"} err="failed to get container status \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\": rpc error: code = NotFound desc = an error occurred when try to find container \"1cf466436dbe85326b618aa9f2bddd7d712eb0886d2db4c90ef33df5b095719e\": not found" Sep 4 00:07:13.418101 kubelet[2730]: I0904 00:07:13.418084 2730 scope.go:117] "RemoveContainer" containerID="a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa" Sep 4 00:07:13.418226 containerd[1582]: time="2025-09-04T00:07:13.418196749Z" level=error msg="ContainerStatus for \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\": not found" Sep 4 00:07:13.418580 kubelet[2730]: E0904 00:07:13.418526 2730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\": not found" containerID="a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa" Sep 4 00:07:13.418646 kubelet[2730]: I0904 00:07:13.418581 2730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa"} err="failed to get container status \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\": rpc error: code = NotFound desc = an error occurred when try to find container \"a5a384fff065705aa81badfa14c057e45361459f0c4aeee53caf88fa3a104faa\": not found" Sep 4 00:07:13.418646 kubelet[2730]: I0904 00:07:13.418610 2730 scope.go:117] "RemoveContainer" containerID="6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f" Sep 4 00:07:13.420233 containerd[1582]: time="2025-09-04T00:07:13.420191877Z" level=info msg="RemoveContainer for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\"" Sep 4 00:07:13.425628 containerd[1582]: time="2025-09-04T00:07:13.425585539Z" level=info msg="RemoveContainer for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" returns successfully" Sep 4 00:07:13.425819 kubelet[2730]: I0904 00:07:13.425771 2730 scope.go:117] "RemoveContainer" containerID="6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f" Sep 4 00:07:13.426031 containerd[1582]: time="2025-09-04T00:07:13.425983523Z" level=error msg="ContainerStatus for \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\": not found" Sep 4 00:07:13.426131 kubelet[2730]: E0904 00:07:13.426103 2730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\": not found" containerID="6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f" Sep 4 00:07:13.426164 kubelet[2730]: I0904 00:07:13.426140 2730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f"} err="failed to get container status \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\": rpc error: code = NotFound desc = an error occurred when try to find container \"6788703d07af13afdb003ff8256371af086dbe2ef25c743437cdff0c5e5a387f\": not found" Sep 4 00:07:13.697243 kubelet[2730]: I0904 00:07:13.697168 2730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84438b01-62a9-44cf-9db8-0c9d2cca272c" path="/var/lib/kubelet/pods/84438b01-62a9-44cf-9db8-0c9d2cca272c/volumes" Sep 4 00:07:13.697905 kubelet[2730]: I0904 00:07:13.697865 2730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a61e66f7-5c59-4052-9364-69f00bf363d7" path="/var/lib/kubelet/pods/a61e66f7-5c59-4052-9364-69f00bf363d7/volumes" Sep 4 00:07:13.720871 systemd[1]: var-lib-kubelet-pods-a61e66f7\x2d5c59\x2d4052\x2d9364\x2d69f00bf363d7-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d6ln5n.mount: Deactivated successfully. Sep 4 00:07:13.721020 systemd[1]: var-lib-kubelet-pods-84438b01\x2d62a9\x2d44cf\x2d9db8\x2d0c9d2cca272c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dzsnds.mount: Deactivated successfully. Sep 4 00:07:13.721124 systemd[1]: var-lib-kubelet-pods-a61e66f7\x2d5c59\x2d4052\x2d9364\x2d69f00bf363d7-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 4 00:07:13.721207 systemd[1]: var-lib-kubelet-pods-a61e66f7\x2d5c59\x2d4052\x2d9364\x2d69f00bf363d7-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 4 00:07:14.571135 sshd[4385]: Connection closed by 10.0.0.1 port 52914 Sep 4 00:07:14.571793 sshd-session[4383]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:14.584606 systemd[1]: sshd@28-10.0.0.91:22-10.0.0.1:52914.service: Deactivated successfully. Sep 4 00:07:14.586800 systemd[1]: session-29.scope: Deactivated successfully. Sep 4 00:07:14.587899 systemd-logind[1561]: Session 29 logged out. Waiting for processes to exit. Sep 4 00:07:14.591240 systemd[1]: Started sshd@29-10.0.0.91:22-10.0.0.1:52922.service - OpenSSH per-connection server daemon (10.0.0.1:52922). Sep 4 00:07:14.591833 systemd-logind[1561]: Removed session 29. Sep 4 00:07:14.653031 sshd[4538]: Accepted publickey for core from 10.0.0.1 port 52922 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:14.654673 sshd-session[4538]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:14.659933 systemd-logind[1561]: New session 30 of user core. Sep 4 00:07:14.668766 systemd[1]: Started session-30.scope - Session 30 of User core. Sep 4 00:07:15.377066 sshd[4540]: Connection closed by 10.0.0.1 port 52922 Sep 4 00:07:15.378783 sshd-session[4538]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:15.391447 systemd[1]: sshd@29-10.0.0.91:22-10.0.0.1:52922.service: Deactivated successfully. Sep 4 00:07:15.394966 systemd[1]: session-30.scope: Deactivated successfully. Sep 4 00:07:15.399453 systemd-logind[1561]: Session 30 logged out. Waiting for processes to exit. Sep 4 00:07:15.402965 systemd-logind[1561]: Removed session 30. Sep 4 00:07:15.405235 systemd[1]: Started sshd@30-10.0.0.91:22-10.0.0.1:52936.service - OpenSSH per-connection server daemon (10.0.0.1:52936). Sep 4 00:07:15.436627 kubelet[2730]: I0904 00:07:15.436516 2730 memory_manager.go:355] "RemoveStaleState removing state" podUID="84438b01-62a9-44cf-9db8-0c9d2cca272c" containerName="cilium-operator" Sep 4 00:07:15.438222 kubelet[2730]: I0904 00:07:15.438174 2730 memory_manager.go:355] "RemoveStaleState removing state" podUID="a61e66f7-5c59-4052-9364-69f00bf363d7" containerName="cilium-agent" Sep 4 00:07:15.454449 systemd[1]: Created slice kubepods-burstable-pod7152c988_5573_40dd_b647_3e057ef095dc.slice - libcontainer container kubepods-burstable-pod7152c988_5573_40dd_b647_3e057ef095dc.slice. Sep 4 00:07:15.478966 sshd[4553]: Accepted publickey for core from 10.0.0.1 port 52936 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:15.481901 sshd-session[4553]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:15.491690 systemd-logind[1561]: New session 31 of user core. Sep 4 00:07:15.498702 systemd[1]: Started session-31.scope - Session 31 of User core. Sep 4 00:07:15.552055 sshd[4555]: Connection closed by 10.0.0.1 port 52936 Sep 4 00:07:15.552511 sshd-session[4553]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:15.567640 systemd[1]: sshd@30-10.0.0.91:22-10.0.0.1:52936.service: Deactivated successfully. Sep 4 00:07:15.569647 systemd[1]: session-31.scope: Deactivated successfully. Sep 4 00:07:15.570575 systemd-logind[1561]: Session 31 logged out. Waiting for processes to exit. Sep 4 00:07:15.573898 systemd[1]: Started sshd@31-10.0.0.91:22-10.0.0.1:52940.service - OpenSSH per-connection server daemon (10.0.0.1:52940). Sep 4 00:07:15.574598 systemd-logind[1561]: Removed session 31. Sep 4 00:07:15.592859 kubelet[2730]: I0904 00:07:15.592815 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7152c988-5573-40dd-b647-3e057ef095dc-hubble-tls\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.592859 kubelet[2730]: I0904 00:07:15.592860 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-bpf-maps\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593075 kubelet[2730]: I0904 00:07:15.592883 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-cilium-cgroup\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593075 kubelet[2730]: I0904 00:07:15.592905 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-etc-cni-netd\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593075 kubelet[2730]: I0904 00:07:15.592925 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7152c988-5573-40dd-b647-3e057ef095dc-cilium-config-path\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593075 kubelet[2730]: I0904 00:07:15.593000 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-host-proc-sys-kernel\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593075 kubelet[2730]: I0904 00:07:15.593045 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-hostproc\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593075 kubelet[2730]: I0904 00:07:15.593073 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-xtables-lock\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593249 kubelet[2730]: I0904 00:07:15.593099 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-host-proc-sys-net\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593249 kubelet[2730]: I0904 00:07:15.593122 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-cni-path\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593249 kubelet[2730]: I0904 00:07:15.593148 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc4mr\" (UniqueName: \"kubernetes.io/projected/7152c988-5573-40dd-b647-3e057ef095dc-kube-api-access-pc4mr\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593249 kubelet[2730]: I0904 00:07:15.593171 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-lib-modules\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593249 kubelet[2730]: I0904 00:07:15.593193 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7152c988-5573-40dd-b647-3e057ef095dc-clustermesh-secrets\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593397 kubelet[2730]: I0904 00:07:15.593219 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/7152c988-5573-40dd-b647-3e057ef095dc-cilium-ipsec-secrets\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.593397 kubelet[2730]: I0904 00:07:15.593245 2730 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7152c988-5573-40dd-b647-3e057ef095dc-cilium-run\") pod \"cilium-nhqzf\" (UID: \"7152c988-5573-40dd-b647-3e057ef095dc\") " pod="kube-system/cilium-nhqzf" Sep 4 00:07:15.631895 sshd[4562]: Accepted publickey for core from 10.0.0.1 port 52940 ssh2: RSA SHA256:1o0Rn/iFE2HG+o4C2c8UWdMz6TCxmTa3FwGAPCIw01A Sep 4 00:07:15.633504 sshd-session[4562]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 00:07:15.640046 systemd-logind[1561]: New session 32 of user core. Sep 4 00:07:15.644781 systemd[1]: Started session-32.scope - Session 32 of User core. Sep 4 00:07:15.759034 kubelet[2730]: E0904 00:07:15.758951 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:15.761418 containerd[1582]: time="2025-09-04T00:07:15.761364074Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nhqzf,Uid:7152c988-5573-40dd-b647-3e057ef095dc,Namespace:kube-system,Attempt:0,}" Sep 4 00:07:15.784558 containerd[1582]: time="2025-09-04T00:07:15.784442607Z" level=info msg="connecting to shim d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c" address="unix:///run/containerd/s/f3f8ef96c8203e06329ce763dd6588265bda2ab0d190eb6221f210cf5a5e1326" namespace=k8s.io protocol=ttrpc version=3 Sep 4 00:07:15.816839 systemd[1]: Started cri-containerd-d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c.scope - libcontainer container d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c. Sep 4 00:07:15.849797 containerd[1582]: time="2025-09-04T00:07:15.849751161Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nhqzf,Uid:7152c988-5573-40dd-b647-3e057ef095dc,Namespace:kube-system,Attempt:0,} returns sandbox id \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\"" Sep 4 00:07:15.850448 kubelet[2730]: E0904 00:07:15.850422 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:15.857355 containerd[1582]: time="2025-09-04T00:07:15.857268029Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 4 00:07:15.868498 containerd[1582]: time="2025-09-04T00:07:15.868442510Z" level=info msg="Container 7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:07:15.924210 containerd[1582]: time="2025-09-04T00:07:15.924079227Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\"" Sep 4 00:07:15.925572 containerd[1582]: time="2025-09-04T00:07:15.924656028Z" level=info msg="StartContainer for \"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\"" Sep 4 00:07:15.925865 containerd[1582]: time="2025-09-04T00:07:15.925818869Z" level=info msg="connecting to shim 7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845" address="unix:///run/containerd/s/f3f8ef96c8203e06329ce763dd6588265bda2ab0d190eb6221f210cf5a5e1326" protocol=ttrpc version=3 Sep 4 00:07:15.952805 systemd[1]: Started cri-containerd-7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845.scope - libcontainer container 7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845. Sep 4 00:07:15.988946 containerd[1582]: time="2025-09-04T00:07:15.988897523Z" level=info msg="StartContainer for \"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\" returns successfully" Sep 4 00:07:15.995698 systemd[1]: cri-containerd-7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845.scope: Deactivated successfully. Sep 4 00:07:15.996988 containerd[1582]: time="2025-09-04T00:07:15.996958090Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\" id:\"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\" pid:4633 exited_at:{seconds:1756944435 nanos:996645238}" Sep 4 00:07:16.003197 containerd[1582]: time="2025-09-04T00:07:16.003160719Z" level=info msg="received exit event container_id:\"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\" id:\"7ea51e1bac8dd639433abab70e9a9bd271b9eba9f754c9482bd16635391a3845\" pid:4633 exited_at:{seconds:1756944435 nanos:996645238}" Sep 4 00:07:16.384000 kubelet[2730]: E0904 00:07:16.383963 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:16.386043 containerd[1582]: time="2025-09-04T00:07:16.386007309Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 4 00:07:16.394166 containerd[1582]: time="2025-09-04T00:07:16.394111556Z" level=info msg="Container c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:07:16.401526 containerd[1582]: time="2025-09-04T00:07:16.401477125Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\"" Sep 4 00:07:16.402147 containerd[1582]: time="2025-09-04T00:07:16.402114411Z" level=info msg="StartContainer for \"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\"" Sep 4 00:07:16.403169 containerd[1582]: time="2025-09-04T00:07:16.403141726Z" level=info msg="connecting to shim c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027" address="unix:///run/containerd/s/f3f8ef96c8203e06329ce763dd6588265bda2ab0d190eb6221f210cf5a5e1326" protocol=ttrpc version=3 Sep 4 00:07:16.424695 systemd[1]: Started cri-containerd-c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027.scope - libcontainer container c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027. Sep 4 00:07:16.458293 containerd[1582]: time="2025-09-04T00:07:16.458247270Z" level=info msg="StartContainer for \"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\" returns successfully" Sep 4 00:07:16.462255 systemd[1]: cri-containerd-c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027.scope: Deactivated successfully. Sep 4 00:07:16.463021 containerd[1582]: time="2025-09-04T00:07:16.462983332Z" level=info msg="received exit event container_id:\"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\" id:\"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\" pid:4681 exited_at:{seconds:1756944436 nanos:462780489}" Sep 4 00:07:16.463277 containerd[1582]: time="2025-09-04T00:07:16.463241772Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\" id:\"c788de942f23cdc5ef678476455e71b92c9897736cef49f0cad04c1bcd89c027\" pid:4681 exited_at:{seconds:1756944436 nanos:462780489}" Sep 4 00:07:16.695065 kubelet[2730]: E0904 00:07:16.694997 2730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-kc9d5" podUID="1768467c-8b64-4401-bfe8-c37479a2e9f5" Sep 4 00:07:16.701838 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount574854224.mount: Deactivated successfully. Sep 4 00:07:17.388731 kubelet[2730]: E0904 00:07:17.388673 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:17.391425 containerd[1582]: time="2025-09-04T00:07:17.391370285Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 4 00:07:17.404197 containerd[1582]: time="2025-09-04T00:07:17.404122224Z" level=info msg="Container 07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:07:17.418842 containerd[1582]: time="2025-09-04T00:07:17.418771783Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\"" Sep 4 00:07:17.419445 containerd[1582]: time="2025-09-04T00:07:17.419403258Z" level=info msg="StartContainer for \"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\"" Sep 4 00:07:17.421061 containerd[1582]: time="2025-09-04T00:07:17.421022832Z" level=info msg="connecting to shim 07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256" address="unix:///run/containerd/s/f3f8ef96c8203e06329ce763dd6588265bda2ab0d190eb6221f210cf5a5e1326" protocol=ttrpc version=3 Sep 4 00:07:17.447702 systemd[1]: Started cri-containerd-07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256.scope - libcontainer container 07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256. Sep 4 00:07:17.508059 containerd[1582]: time="2025-09-04T00:07:17.508007380Z" level=info msg="StartContainer for \"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\" returns successfully" Sep 4 00:07:17.510311 systemd[1]: cri-containerd-07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256.scope: Deactivated successfully. Sep 4 00:07:17.512831 containerd[1582]: time="2025-09-04T00:07:17.512783837Z" level=info msg="received exit event container_id:\"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\" id:\"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\" pid:4725 exited_at:{seconds:1756944437 nanos:512133738}" Sep 4 00:07:17.514673 containerd[1582]: time="2025-09-04T00:07:17.514632756Z" level=info msg="TaskExit event in podsandbox handler container_id:\"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\" id:\"07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256\" pid:4725 exited_at:{seconds:1756944437 nanos:512133738}" Sep 4 00:07:17.548183 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-07b676b58df49e0f458fe950ab002cd330fab75a006d435174b567cd9df55256-rootfs.mount: Deactivated successfully. Sep 4 00:07:17.757781 kubelet[2730]: E0904 00:07:17.757738 2730 kubelet.go:3002] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 4 00:07:18.392472 kubelet[2730]: E0904 00:07:18.392433 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:18.400849 containerd[1582]: time="2025-09-04T00:07:18.400807654Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 4 00:07:18.654724 containerd[1582]: time="2025-09-04T00:07:18.654589633Z" level=info msg="Container 3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:07:18.694880 kubelet[2730]: E0904 00:07:18.694814 2730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-kc9d5" podUID="1768467c-8b64-4401-bfe8-c37479a2e9f5" Sep 4 00:07:18.765008 containerd[1582]: time="2025-09-04T00:07:18.764945210Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\"" Sep 4 00:07:18.765541 containerd[1582]: time="2025-09-04T00:07:18.765503917Z" level=info msg="StartContainer for \"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\"" Sep 4 00:07:18.766790 containerd[1582]: time="2025-09-04T00:07:18.766763850Z" level=info msg="connecting to shim 3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676" address="unix:///run/containerd/s/f3f8ef96c8203e06329ce763dd6588265bda2ab0d190eb6221f210cf5a5e1326" protocol=ttrpc version=3 Sep 4 00:07:18.790771 systemd[1]: Started cri-containerd-3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676.scope - libcontainer container 3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676. Sep 4 00:07:18.820854 systemd[1]: cri-containerd-3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676.scope: Deactivated successfully. Sep 4 00:07:18.822316 containerd[1582]: time="2025-09-04T00:07:18.822259600Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\" id:\"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\" pid:4763 exited_at:{seconds:1756944438 nanos:821740438}" Sep 4 00:07:18.823596 containerd[1582]: time="2025-09-04T00:07:18.823559549Z" level=info msg="received exit event container_id:\"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\" id:\"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\" pid:4763 exited_at:{seconds:1756944438 nanos:821740438}" Sep 4 00:07:18.832160 containerd[1582]: time="2025-09-04T00:07:18.832121017Z" level=info msg="StartContainer for \"3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676\" returns successfully" Sep 4 00:07:18.845940 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3d0d6077b8e5f6a7feefd9440a4bc347b25312e28703d16e144d332c47567676-rootfs.mount: Deactivated successfully. Sep 4 00:07:19.398618 kubelet[2730]: E0904 00:07:19.398541 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:19.400194 containerd[1582]: time="2025-09-04T00:07:19.400136947Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 4 00:07:19.410567 containerd[1582]: time="2025-09-04T00:07:19.410277268Z" level=info msg="Container aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86: CDI devices from CRI Config.CDIDevices: []" Sep 4 00:07:19.423690 containerd[1582]: time="2025-09-04T00:07:19.423638582Z" level=info msg="CreateContainer within sandbox \"d20d7eb9bd3a78533941fd0c4d73ff425413745fd1373aae05ded1efa9be961c\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\"" Sep 4 00:07:19.424570 containerd[1582]: time="2025-09-04T00:07:19.424075398Z" level=info msg="StartContainer for \"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\"" Sep 4 00:07:19.425322 containerd[1582]: time="2025-09-04T00:07:19.425296237Z" level=info msg="connecting to shim aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86" address="unix:///run/containerd/s/f3f8ef96c8203e06329ce763dd6588265bda2ab0d190eb6221f210cf5a5e1326" protocol=ttrpc version=3 Sep 4 00:07:19.452862 systemd[1]: Started cri-containerd-aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86.scope - libcontainer container aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86. Sep 4 00:07:19.489080 containerd[1582]: time="2025-09-04T00:07:19.489042628Z" level=info msg="StartContainer for \"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" returns successfully" Sep 4 00:07:19.560983 containerd[1582]: time="2025-09-04T00:07:19.560936991Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" id:\"5319914c7b0038d6dc4bc55cfb77bf746dd3de434a78d46b40661a315322d3ec\" pid:4831 exited_at:{seconds:1756944439 nanos:560445060}" Sep 4 00:07:19.947700 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni-avx)) Sep 4 00:07:20.405981 kubelet[2730]: E0904 00:07:20.405840 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:20.444177 kubelet[2730]: I0904 00:07:20.444113 2730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-nhqzf" podStartSLOduration=5.444094683 podStartE2EDuration="5.444094683s" podCreationTimestamp="2025-09-04 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-04 00:07:20.4440671 +0000 UTC m=+113.172357993" watchObservedRunningTime="2025-09-04 00:07:20.444094683 +0000 UTC m=+113.172385566" Sep 4 00:07:20.695036 kubelet[2730]: E0904 00:07:20.694973 2730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-kc9d5" podUID="1768467c-8b64-4401-bfe8-c37479a2e9f5" Sep 4 00:07:21.173402 kubelet[2730]: I0904 00:07:21.173243 2730 setters.go:602] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-04T00:07:21Z","lastTransitionTime":"2025-09-04T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 4 00:07:21.760182 kubelet[2730]: E0904 00:07:21.760051 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:22.291299 containerd[1582]: time="2025-09-04T00:07:22.291249870Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" id:\"960f17a2a8f2f4d3c42af9ce39211575508281ce519392d4fd6412641294e05c\" pid:5140 exit_status:1 exited_at:{seconds:1756944442 nanos:290896381}" Sep 4 00:07:22.696084 kubelet[2730]: E0904 00:07:22.695834 2730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-kc9d5" podUID="1768467c-8b64-4401-bfe8-c37479a2e9f5" Sep 4 00:07:23.610591 systemd-networkd[1461]: lxc_health: Link UP Sep 4 00:07:23.611071 systemd-networkd[1461]: lxc_health: Gained carrier Sep 4 00:07:23.762377 kubelet[2730]: E0904 00:07:23.762329 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:24.419334 kubelet[2730]: E0904 00:07:24.418263 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:24.469792 containerd[1582]: time="2025-09-04T00:07:24.469738043Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" id:\"04e283526f0e281088562c6f4cac282bdf656b26bd1d3db916f1f0dd8a06ae09\" pid:5360 exited_at:{seconds:1756944444 nanos:469092462}" Sep 4 00:07:24.695941 kubelet[2730]: E0904 00:07:24.695452 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:24.695941 kubelet[2730]: E0904 00:07:24.695583 2730 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 00:07:25.040886 systemd-networkd[1461]: lxc_health: Gained IPv6LL Sep 4 00:07:26.595644 containerd[1582]: time="2025-09-04T00:07:26.595582234Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" id:\"40d371ddc3108d5ac115ce96ee4fba9b68602a36fa9a3eb53afd6bc506d9f327\" pid:5398 exited_at:{seconds:1756944446 nanos:595171769}" Sep 4 00:07:27.726436 containerd[1582]: time="2025-09-04T00:07:27.726330704Z" level=info msg="StopPodSandbox for \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\"" Sep 4 00:07:27.726929 containerd[1582]: time="2025-09-04T00:07:27.726588902Z" level=info msg="TearDown network for sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" successfully" Sep 4 00:07:27.726929 containerd[1582]: time="2025-09-04T00:07:27.726606906Z" level=info msg="StopPodSandbox for \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" returns successfully" Sep 4 00:07:27.727284 containerd[1582]: time="2025-09-04T00:07:27.727256383Z" level=info msg="RemovePodSandbox for \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\"" Sep 4 00:07:27.727330 containerd[1582]: time="2025-09-04T00:07:27.727287642Z" level=info msg="Forcibly stopping sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\"" Sep 4 00:07:27.727377 containerd[1582]: time="2025-09-04T00:07:27.727350942Z" level=info msg="TearDown network for sandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" successfully" Sep 4 00:07:27.729352 containerd[1582]: time="2025-09-04T00:07:27.729315505Z" level=info msg="Ensure that sandbox 693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b in task-service has been cleanup successfully" Sep 4 00:07:27.735784 containerd[1582]: time="2025-09-04T00:07:27.735629184Z" level=info msg="RemovePodSandbox \"693a37e6396232fe0db2b682d31b85433b8fd3a658b29a59f7c37e00f506f93b\" returns successfully" Sep 4 00:07:27.736568 containerd[1582]: time="2025-09-04T00:07:27.736508166Z" level=info msg="StopPodSandbox for \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\"" Sep 4 00:07:27.736937 containerd[1582]: time="2025-09-04T00:07:27.736867865Z" level=info msg="TearDown network for sandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" successfully" Sep 4 00:07:27.736937 containerd[1582]: time="2025-09-04T00:07:27.736915375Z" level=info msg="StopPodSandbox for \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" returns successfully" Sep 4 00:07:27.737569 containerd[1582]: time="2025-09-04T00:07:27.737443663Z" level=info msg="RemovePodSandbox for \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\"" Sep 4 00:07:27.737569 containerd[1582]: time="2025-09-04T00:07:27.737466106Z" level=info msg="Forcibly stopping sandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\"" Sep 4 00:07:27.737569 containerd[1582]: time="2025-09-04T00:07:27.737530728Z" level=info msg="TearDown network for sandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" successfully" Sep 4 00:07:27.739567 containerd[1582]: time="2025-09-04T00:07:27.739518714Z" level=info msg="Ensure that sandbox 3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6 in task-service has been cleanup successfully" Sep 4 00:07:27.743604 containerd[1582]: time="2025-09-04T00:07:27.743575991Z" level=info msg="RemovePodSandbox \"3e9e5e339edc99b1ebf70e7b17178427a137be8bd1a04709853eb8cb9e181cb6\" returns successfully" Sep 4 00:07:28.912299 containerd[1582]: time="2025-09-04T00:07:28.912228940Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" id:\"3c182942b606462c88f8f8b20758c98808c1b0b09699736b4659bf1f12bd04cb\" pid:5431 exited_at:{seconds:1756944448 nanos:911766546}" Sep 4 00:07:31.051889 containerd[1582]: time="2025-09-04T00:07:31.051807057Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa324da6de3ea678d72cb8b9eb7ddff4355c54927d1b8201518be15900181f86\" id:\"f6865d39e377fb55a54e05a174b19ce41e08222d33ddc473909b0069b98d8bf1\" pid:5455 exited_at:{seconds:1756944451 nanos:51356386}" Sep 4 00:07:31.061149 sshd[4564]: Connection closed by 10.0.0.1 port 52940 Sep 4 00:07:31.061717 sshd-session[4562]: pam_unix(sshd:session): session closed for user core Sep 4 00:07:31.066887 systemd[1]: sshd@31-10.0.0.91:22-10.0.0.1:52940.service: Deactivated successfully. Sep 4 00:07:31.069379 systemd[1]: session-32.scope: Deactivated successfully. Sep 4 00:07:31.070211 systemd-logind[1561]: Session 32 logged out. Waiting for processes to exit. Sep 4 00:07:31.072163 systemd-logind[1561]: Removed session 32.