Sep 10 00:40:37.157509 kernel: Linux version 6.6.104-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Sep 9 22:56:44 -00 2025 Sep 10 00:40:37.157552 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a3dbdfb50e14c8de85dda26f853cdd6055239b4b8b15c08fb0eb00b67ce87a58 Sep 10 00:40:37.157566 kernel: BIOS-provided physical RAM map: Sep 10 00:40:37.157575 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 10 00:40:37.157583 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Sep 10 00:40:37.157591 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Sep 10 00:40:37.157601 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Sep 10 00:40:37.157610 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Sep 10 00:40:37.157618 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable Sep 10 00:40:37.157627 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS Sep 10 00:40:37.157638 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable Sep 10 00:40:37.157647 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009c9eefff] reserved Sep 10 00:40:37.157658 kernel: BIOS-e820: [mem 0x000000009c9ef000-0x000000009caeefff] type 20 Sep 10 00:40:37.157667 kernel: BIOS-e820: [mem 0x000000009caef000-0x000000009cb6efff] reserved Sep 10 00:40:37.157680 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data Sep 10 00:40:37.157690 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Sep 10 00:40:37.157702 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable Sep 10 00:40:37.157711 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved Sep 10 00:40:37.157720 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Sep 10 00:40:37.157729 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Sep 10 00:40:37.157739 kernel: NX (Execute Disable) protection: active Sep 10 00:40:37.157748 kernel: APIC: Static calls initialized Sep 10 00:40:37.157757 kernel: efi: EFI v2.7 by EDK II Sep 10 00:40:37.157766 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b675198 Sep 10 00:40:37.157775 kernel: SMBIOS 2.8 present. Sep 10 00:40:37.157784 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 0.0.0 02/06/2015 Sep 10 00:40:37.157793 kernel: Hypervisor detected: KVM Sep 10 00:40:37.157805 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 10 00:40:37.157814 kernel: kvm-clock: using sched offset of 6906849121 cycles Sep 10 00:40:37.157824 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 10 00:40:37.157833 kernel: tsc: Detected 2794.748 MHz processor Sep 10 00:40:37.157843 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 10 00:40:37.157853 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 10 00:40:37.157862 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x400000000 Sep 10 00:40:37.157872 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 10 00:40:37.157881 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 10 00:40:37.157894 kernel: Using GB pages for direct mapping Sep 10 00:40:37.157903 kernel: Secure boot disabled Sep 10 00:40:37.157912 kernel: ACPI: Early table checksum verification disabled Sep 10 00:40:37.157922 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Sep 10 00:40:37.157936 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Sep 10 00:40:37.157946 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:40:37.157955 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:40:37.157968 kernel: ACPI: FACS 0x000000009CBDD000 000040 Sep 10 00:40:37.157978 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:40:37.157991 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:40:37.158010 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:40:37.158020 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:40:37.158030 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Sep 10 00:40:37.158040 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Sep 10 00:40:37.158053 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Sep 10 00:40:37.158063 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Sep 10 00:40:37.158073 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Sep 10 00:40:37.158083 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Sep 10 00:40:37.158093 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Sep 10 00:40:37.158103 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Sep 10 00:40:37.158112 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Sep 10 00:40:37.158122 kernel: No NUMA configuration found Sep 10 00:40:37.158135 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] Sep 10 00:40:37.158148 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] Sep 10 00:40:37.158158 kernel: Zone ranges: Sep 10 00:40:37.158167 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 10 00:40:37.158177 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] Sep 10 00:40:37.158187 kernel: Normal empty Sep 10 00:40:37.158197 kernel: Movable zone start for each node Sep 10 00:40:37.158207 kernel: Early memory node ranges Sep 10 00:40:37.158216 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 10 00:40:37.158226 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Sep 10 00:40:37.158236 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Sep 10 00:40:37.158249 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] Sep 10 00:40:37.158273 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] Sep 10 00:40:37.158283 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] Sep 10 00:40:37.158296 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] Sep 10 00:40:37.158306 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 10 00:40:37.158316 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 10 00:40:37.158325 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Sep 10 00:40:37.158335 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 10 00:40:37.158345 kernel: On node 0, zone DMA: 240 pages in unavailable ranges Sep 10 00:40:37.158358 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Sep 10 00:40:37.158368 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges Sep 10 00:40:37.158378 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 10 00:40:37.158388 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 10 00:40:37.158398 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 10 00:40:37.158408 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 10 00:40:37.158418 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 10 00:40:37.158427 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 10 00:40:37.158437 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 10 00:40:37.158450 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 10 00:40:37.158460 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 10 00:40:37.158470 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 10 00:40:37.158479 kernel: TSC deadline timer available Sep 10 00:40:37.158489 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Sep 10 00:40:37.158499 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 10 00:40:37.158509 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 10 00:40:37.158518 kernel: kvm-guest: setup PV sched yield Sep 10 00:40:37.158528 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Sep 10 00:40:37.158541 kernel: Booting paravirtualized kernel on KVM Sep 10 00:40:37.158551 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 10 00:40:37.158561 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 10 00:40:37.158571 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u524288 Sep 10 00:40:37.158581 kernel: pcpu-alloc: s197160 r8192 d32216 u524288 alloc=1*2097152 Sep 10 00:40:37.158590 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 10 00:40:37.158600 kernel: kvm-guest: PV spinlocks enabled Sep 10 00:40:37.158610 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 10 00:40:37.158621 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a3dbdfb50e14c8de85dda26f853cdd6055239b4b8b15c08fb0eb00b67ce87a58 Sep 10 00:40:37.158637 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 10 00:40:37.158647 kernel: random: crng init done Sep 10 00:40:37.158657 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 10 00:40:37.158667 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 10 00:40:37.158677 kernel: Fallback order for Node 0: 0 Sep 10 00:40:37.158686 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 Sep 10 00:40:37.158696 kernel: Policy zone: DMA32 Sep 10 00:40:37.158706 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 10 00:40:37.158716 kernel: Memory: 2400600K/2567000K available (12288K kernel code, 2293K rwdata, 22744K rodata, 42884K init, 2312K bss, 166140K reserved, 0K cma-reserved) Sep 10 00:40:37.158729 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 10 00:40:37.158739 kernel: ftrace: allocating 37969 entries in 149 pages Sep 10 00:40:37.158749 kernel: ftrace: allocated 149 pages with 4 groups Sep 10 00:40:37.158759 kernel: Dynamic Preempt: voluntary Sep 10 00:40:37.158778 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 10 00:40:37.158792 kernel: rcu: RCU event tracing is enabled. Sep 10 00:40:37.158803 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 10 00:40:37.158813 kernel: Trampoline variant of Tasks RCU enabled. Sep 10 00:40:37.158823 kernel: Rude variant of Tasks RCU enabled. Sep 10 00:40:37.158834 kernel: Tracing variant of Tasks RCU enabled. Sep 10 00:40:37.158844 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 10 00:40:37.158857 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 10 00:40:37.158868 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 10 00:40:37.158881 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 10 00:40:37.158891 kernel: Console: colour dummy device 80x25 Sep 10 00:40:37.158901 kernel: printk: console [ttyS0] enabled Sep 10 00:40:37.158915 kernel: ACPI: Core revision 20230628 Sep 10 00:40:37.158925 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 10 00:40:37.158935 kernel: APIC: Switch to symmetric I/O mode setup Sep 10 00:40:37.158946 kernel: x2apic enabled Sep 10 00:40:37.158956 kernel: APIC: Switched APIC routing to: physical x2apic Sep 10 00:40:37.158966 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 10 00:40:37.158977 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 10 00:40:37.158987 kernel: kvm-guest: setup PV IPIs Sep 10 00:40:37.160147 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 10 00:40:37.160166 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Sep 10 00:40:37.160174 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Sep 10 00:40:37.160182 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 10 00:40:37.160190 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 10 00:40:37.160198 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 10 00:40:37.160209 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 10 00:40:37.160220 kernel: Spectre V2 : Mitigation: Retpolines Sep 10 00:40:37.160231 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 10 00:40:37.160241 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 10 00:40:37.160265 kernel: active return thunk: retbleed_return_thunk Sep 10 00:40:37.160273 kernel: RETBleed: Mitigation: untrained return thunk Sep 10 00:40:37.160281 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 10 00:40:37.160289 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 10 00:40:37.160300 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 10 00:40:37.160309 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 10 00:40:37.160317 kernel: active return thunk: srso_return_thunk Sep 10 00:40:37.160325 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 10 00:40:37.160336 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 10 00:40:37.160344 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 10 00:40:37.160352 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 10 00:40:37.160359 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 10 00:40:37.160367 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 10 00:40:37.160375 kernel: Freeing SMP alternatives memory: 32K Sep 10 00:40:37.160385 kernel: pid_max: default: 32768 minimum: 301 Sep 10 00:40:37.160395 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 10 00:40:37.160405 kernel: landlock: Up and running. Sep 10 00:40:37.160417 kernel: SELinux: Initializing. Sep 10 00:40:37.160427 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 10 00:40:37.160437 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 10 00:40:37.160447 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 10 00:40:37.160456 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 10 00:40:37.160466 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 10 00:40:37.160475 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 10 00:40:37.160482 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 10 00:40:37.160490 kernel: ... version: 0 Sep 10 00:40:37.160500 kernel: ... bit width: 48 Sep 10 00:40:37.160508 kernel: ... generic registers: 6 Sep 10 00:40:37.160516 kernel: ... value mask: 0000ffffffffffff Sep 10 00:40:37.160523 kernel: ... max period: 00007fffffffffff Sep 10 00:40:37.160531 kernel: ... fixed-purpose events: 0 Sep 10 00:40:37.160539 kernel: ... event mask: 000000000000003f Sep 10 00:40:37.160546 kernel: signal: max sigframe size: 1776 Sep 10 00:40:37.160554 kernel: rcu: Hierarchical SRCU implementation. Sep 10 00:40:37.160562 kernel: rcu: Max phase no-delay instances is 400. Sep 10 00:40:37.160573 kernel: smp: Bringing up secondary CPUs ... Sep 10 00:40:37.160581 kernel: smpboot: x86: Booting SMP configuration: Sep 10 00:40:37.160589 kernel: .... node #0, CPUs: #1 #2 #3 Sep 10 00:40:37.160597 kernel: smp: Brought up 1 node, 4 CPUs Sep 10 00:40:37.160605 kernel: smpboot: Max logical packages: 1 Sep 10 00:40:37.160613 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Sep 10 00:40:37.160621 kernel: devtmpfs: initialized Sep 10 00:40:37.160629 kernel: x86/mm: Memory block size: 128MB Sep 10 00:40:37.160637 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Sep 10 00:40:37.160645 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Sep 10 00:40:37.160656 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) Sep 10 00:40:37.160665 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Sep 10 00:40:37.160673 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Sep 10 00:40:37.160681 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 10 00:40:37.160689 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 10 00:40:37.160697 kernel: pinctrl core: initialized pinctrl subsystem Sep 10 00:40:37.160705 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 10 00:40:37.160713 kernel: audit: initializing netlink subsys (disabled) Sep 10 00:40:37.160724 kernel: audit: type=2000 audit(1757464836.088:1): state=initialized audit_enabled=0 res=1 Sep 10 00:40:37.160732 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 10 00:40:37.160740 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 10 00:40:37.160748 kernel: cpuidle: using governor menu Sep 10 00:40:37.160756 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 10 00:40:37.160764 kernel: dca service started, version 1.12.1 Sep 10 00:40:37.160773 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Sep 10 00:40:37.160781 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Sep 10 00:40:37.160789 kernel: PCI: Using configuration type 1 for base access Sep 10 00:40:37.160799 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 10 00:40:37.160808 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 10 00:40:37.160816 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 10 00:40:37.160824 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 10 00:40:37.160832 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 10 00:40:37.160840 kernel: ACPI: Added _OSI(Module Device) Sep 10 00:40:37.160848 kernel: ACPI: Added _OSI(Processor Device) Sep 10 00:40:37.160856 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 10 00:40:37.160864 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 10 00:40:37.160875 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 10 00:40:37.160883 kernel: ACPI: Interpreter enabled Sep 10 00:40:37.160891 kernel: ACPI: PM: (supports S0 S3 S5) Sep 10 00:40:37.160899 kernel: ACPI: Using IOAPIC for interrupt routing Sep 10 00:40:37.160907 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 10 00:40:37.160915 kernel: PCI: Using E820 reservations for host bridge windows Sep 10 00:40:37.160923 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 10 00:40:37.160931 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 10 00:40:37.161194 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 10 00:40:37.161384 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 10 00:40:37.161542 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 10 00:40:37.161555 kernel: PCI host bridge to bus 0000:00 Sep 10 00:40:37.161720 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 10 00:40:37.161850 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 10 00:40:37.161992 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 10 00:40:37.162151 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Sep 10 00:40:37.162306 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Sep 10 00:40:37.162469 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0xfffffffff window] Sep 10 00:40:37.162605 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 10 00:40:37.162785 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Sep 10 00:40:37.162970 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Sep 10 00:40:37.163131 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Sep 10 00:40:37.163303 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Sep 10 00:40:37.163468 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Sep 10 00:40:37.163722 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Sep 10 00:40:37.163873 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 10 00:40:37.164070 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Sep 10 00:40:37.164226 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Sep 10 00:40:37.164422 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Sep 10 00:40:37.164570 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] Sep 10 00:40:37.164720 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Sep 10 00:40:37.164851 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Sep 10 00:40:37.164978 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Sep 10 00:40:37.166264 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] Sep 10 00:40:37.166507 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Sep 10 00:40:37.166668 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Sep 10 00:40:37.166815 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Sep 10 00:40:37.166948 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] Sep 10 00:40:37.167088 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Sep 10 00:40:37.167313 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Sep 10 00:40:37.167454 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 10 00:40:37.167641 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Sep 10 00:40:37.167784 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Sep 10 00:40:37.167910 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Sep 10 00:40:37.168065 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Sep 10 00:40:37.168193 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Sep 10 00:40:37.168210 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 10 00:40:37.168226 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 10 00:40:37.168235 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 10 00:40:37.168248 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 10 00:40:37.168271 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 10 00:40:37.168279 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 10 00:40:37.168287 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 10 00:40:37.168295 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 10 00:40:37.168302 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 10 00:40:37.168311 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 10 00:40:37.168318 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 10 00:40:37.168326 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 10 00:40:37.168334 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 10 00:40:37.168346 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 10 00:40:37.168354 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 10 00:40:37.168362 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 10 00:40:37.168370 kernel: iommu: Default domain type: Translated Sep 10 00:40:37.168378 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 10 00:40:37.168386 kernel: efivars: Registered efivars operations Sep 10 00:40:37.168394 kernel: PCI: Using ACPI for IRQ routing Sep 10 00:40:37.168402 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 10 00:40:37.168410 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Sep 10 00:40:37.168421 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] Sep 10 00:40:37.168429 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] Sep 10 00:40:37.168437 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] Sep 10 00:40:37.168580 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 10 00:40:37.168709 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 10 00:40:37.168835 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 10 00:40:37.168846 kernel: vgaarb: loaded Sep 10 00:40:37.168854 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 10 00:40:37.168866 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 10 00:40:37.168874 kernel: clocksource: Switched to clocksource kvm-clock Sep 10 00:40:37.168882 kernel: VFS: Disk quotas dquot_6.6.0 Sep 10 00:40:37.168890 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 10 00:40:37.168898 kernel: pnp: PnP ACPI init Sep 10 00:40:37.169076 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Sep 10 00:40:37.169089 kernel: pnp: PnP ACPI: found 6 devices Sep 10 00:40:37.169097 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 10 00:40:37.169109 kernel: NET: Registered PF_INET protocol family Sep 10 00:40:37.169116 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 10 00:40:37.169125 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 10 00:40:37.169133 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 10 00:40:37.169141 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 10 00:40:37.169149 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 10 00:40:37.169157 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 10 00:40:37.169165 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 10 00:40:37.169173 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 10 00:40:37.169186 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 10 00:40:37.169202 kernel: NET: Registered PF_XDP protocol family Sep 10 00:40:37.169390 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Sep 10 00:40:37.169533 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Sep 10 00:40:37.169656 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 10 00:40:37.169774 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 10 00:40:37.169899 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 10 00:40:37.170026 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Sep 10 00:40:37.170150 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Sep 10 00:40:37.170347 kernel: pci_bus 0000:00: resource 9 [mem 0x800000000-0xfffffffff window] Sep 10 00:40:37.170359 kernel: PCI: CLS 0 bytes, default 64 Sep 10 00:40:37.170367 kernel: Initialise system trusted keyrings Sep 10 00:40:37.170375 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 10 00:40:37.170383 kernel: Key type asymmetric registered Sep 10 00:40:37.170391 kernel: Asymmetric key parser 'x509' registered Sep 10 00:40:37.170400 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 10 00:40:37.170408 kernel: io scheduler mq-deadline registered Sep 10 00:40:37.170420 kernel: io scheduler kyber registered Sep 10 00:40:37.170428 kernel: io scheduler bfq registered Sep 10 00:40:37.170436 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 10 00:40:37.170444 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 10 00:40:37.170452 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 10 00:40:37.170460 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 10 00:40:37.170468 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 10 00:40:37.170476 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 10 00:40:37.170484 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 10 00:40:37.170495 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 10 00:40:37.170503 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 10 00:40:37.170511 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 10 00:40:37.170661 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 10 00:40:37.170784 kernel: rtc_cmos 00:04: registered as rtc0 Sep 10 00:40:37.170904 kernel: rtc_cmos 00:04: setting system clock to 2025-09-10T00:40:36 UTC (1757464836) Sep 10 00:40:37.171035 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Sep 10 00:40:37.171046 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 10 00:40:37.171058 kernel: efifb: probing for efifb Sep 10 00:40:37.171066 kernel: efifb: framebuffer at 0xc0000000, using 1408k, total 1408k Sep 10 00:40:37.171074 kernel: efifb: mode is 800x600x24, linelength=2400, pages=1 Sep 10 00:40:37.171082 kernel: efifb: scrolling: redraw Sep 10 00:40:37.171090 kernel: efifb: Truecolor: size=0:8:8:8, shift=0:16:8:0 Sep 10 00:40:37.171099 kernel: Console: switching to colour frame buffer device 100x37 Sep 10 00:40:37.171124 kernel: fb0: EFI VGA frame buffer device Sep 10 00:40:37.171135 kernel: pstore: Using crash dump compression: deflate Sep 10 00:40:37.171144 kernel: pstore: Registered efi_pstore as persistent store backend Sep 10 00:40:37.171155 kernel: NET: Registered PF_INET6 protocol family Sep 10 00:40:37.171163 kernel: Segment Routing with IPv6 Sep 10 00:40:37.171171 kernel: In-situ OAM (IOAM) with IPv6 Sep 10 00:40:37.171179 kernel: NET: Registered PF_PACKET protocol family Sep 10 00:40:37.171187 kernel: Key type dns_resolver registered Sep 10 00:40:37.171196 kernel: IPI shorthand broadcast: enabled Sep 10 00:40:37.171204 kernel: sched_clock: Marking stable (1085010145, 121840823)->(1336801410, -129950442) Sep 10 00:40:37.171212 kernel: registered taskstats version 1 Sep 10 00:40:37.171221 kernel: Loading compiled-in X.509 certificates Sep 10 00:40:37.171232 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.104-flatcar: a614f1c62f27a560d677bbf0283703118c9005ec' Sep 10 00:40:37.171241 kernel: Key type .fscrypt registered Sep 10 00:40:37.171249 kernel: Key type fscrypt-provisioning registered Sep 10 00:40:37.171271 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 10 00:40:37.171279 kernel: ima: Allocated hash algorithm: sha1 Sep 10 00:40:37.171288 kernel: ima: No architecture policies found Sep 10 00:40:37.171296 kernel: clk: Disabling unused clocks Sep 10 00:40:37.171304 kernel: Freeing unused kernel image (initmem) memory: 42884K Sep 10 00:40:37.171313 kernel: Write protecting the kernel read-only data: 36864k Sep 10 00:40:37.171324 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Sep 10 00:40:37.171333 kernel: Run /init as init process Sep 10 00:40:37.171343 kernel: with arguments: Sep 10 00:40:37.171353 kernel: /init Sep 10 00:40:37.171363 kernel: with environment: Sep 10 00:40:37.171373 kernel: HOME=/ Sep 10 00:40:37.171384 kernel: TERM=linux Sep 10 00:40:37.171394 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 10 00:40:37.171410 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 10 00:40:37.171424 systemd[1]: Detected virtualization kvm. Sep 10 00:40:37.171435 systemd[1]: Detected architecture x86-64. Sep 10 00:40:37.171446 systemd[1]: Running in initrd. Sep 10 00:40:37.171462 systemd[1]: No hostname configured, using default hostname. Sep 10 00:40:37.171472 systemd[1]: Hostname set to . Sep 10 00:40:37.171484 systemd[1]: Initializing machine ID from VM UUID. Sep 10 00:40:37.171494 systemd[1]: Queued start job for default target initrd.target. Sep 10 00:40:37.171503 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 10 00:40:37.171511 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 10 00:40:37.171521 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 10 00:40:37.171529 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 10 00:40:37.171541 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 10 00:40:37.171550 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 10 00:40:37.171560 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 10 00:40:37.171569 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 10 00:40:37.171578 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 10 00:40:37.171587 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 10 00:40:37.171595 systemd[1]: Reached target paths.target - Path Units. Sep 10 00:40:37.171607 systemd[1]: Reached target slices.target - Slice Units. Sep 10 00:40:37.171615 systemd[1]: Reached target swap.target - Swaps. Sep 10 00:40:37.171624 systemd[1]: Reached target timers.target - Timer Units. Sep 10 00:40:37.171633 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 10 00:40:37.171641 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 10 00:40:37.171650 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 10 00:40:37.171659 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 10 00:40:37.171667 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 10 00:40:37.171676 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 10 00:40:37.171688 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 10 00:40:37.171696 systemd[1]: Reached target sockets.target - Socket Units. Sep 10 00:40:37.171705 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 10 00:40:37.171714 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 10 00:40:37.171723 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 10 00:40:37.171731 systemd[1]: Starting systemd-fsck-usr.service... Sep 10 00:40:37.171740 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 10 00:40:37.171749 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 10 00:40:37.171760 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:40:37.171769 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 10 00:40:37.171777 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 10 00:40:37.171786 systemd[1]: Finished systemd-fsck-usr.service. Sep 10 00:40:37.171816 systemd-journald[192]: Collecting audit messages is disabled. Sep 10 00:40:37.171840 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 10 00:40:37.171849 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:40:37.171858 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 10 00:40:37.171867 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 10 00:40:37.171878 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 10 00:40:37.171887 systemd-journald[192]: Journal started Sep 10 00:40:37.171905 systemd-journald[192]: Runtime Journal (/run/log/journal/a38dacc556de4cb9877b5cf329df81ce) is 6.0M, max 48.3M, 42.2M free. Sep 10 00:40:37.128406 systemd-modules-load[194]: Inserted module 'overlay' Sep 10 00:40:37.210381 systemd[1]: Started systemd-journald.service - Journal Service. Sep 10 00:40:37.210747 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:40:37.212285 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 10 00:40:37.219278 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 10 00:40:37.248818 systemd-modules-load[194]: Inserted module 'br_netfilter' Sep 10 00:40:37.249963 kernel: Bridge firewalling registered Sep 10 00:40:37.250460 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 10 00:40:37.251512 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 10 00:40:37.254036 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 10 00:40:37.259302 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:40:37.265053 dracut-cmdline[219]: dracut-dracut-053 Sep 10 00:40:37.268558 dracut-cmdline[219]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a3dbdfb50e14c8de85dda26f853cdd6055239b4b8b15c08fb0eb00b67ce87a58 Sep 10 00:40:37.273248 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:40:37.278495 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 10 00:40:37.288450 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 10 00:40:37.324055 systemd-resolved[247]: Positive Trust Anchors: Sep 10 00:40:37.324072 systemd-resolved[247]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 10 00:40:37.324104 systemd-resolved[247]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 10 00:40:37.327131 systemd-resolved[247]: Defaulting to hostname 'linux'. Sep 10 00:40:37.328429 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 10 00:40:37.333588 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 10 00:40:37.391341 kernel: SCSI subsystem initialized Sep 10 00:40:37.404308 kernel: Loading iSCSI transport class v2.0-870. Sep 10 00:40:37.417302 kernel: iscsi: registered transport (tcp) Sep 10 00:40:37.439291 kernel: iscsi: registered transport (qla4xxx) Sep 10 00:40:37.439366 kernel: QLogic iSCSI HBA Driver Sep 10 00:40:37.501634 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 10 00:40:37.513611 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 10 00:40:37.548806 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 10 00:40:37.548932 kernel: device-mapper: uevent: version 1.0.3 Sep 10 00:40:37.548950 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 10 00:40:37.598308 kernel: raid6: avx2x4 gen() 29405 MB/s Sep 10 00:40:37.615300 kernel: raid6: avx2x2 gen() 23639 MB/s Sep 10 00:40:37.632674 kernel: raid6: avx2x1 gen() 20828 MB/s Sep 10 00:40:37.632770 kernel: raid6: using algorithm avx2x4 gen() 29405 MB/s Sep 10 00:40:37.650484 kernel: raid6: .... xor() 6163 MB/s, rmw enabled Sep 10 00:40:37.650562 kernel: raid6: using avx2x2 recovery algorithm Sep 10 00:40:37.676301 kernel: xor: automatically using best checksumming function avx Sep 10 00:40:37.857295 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 10 00:40:37.877693 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 10 00:40:37.886667 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 10 00:40:37.907048 systemd-udevd[413]: Using default interface naming scheme 'v255'. Sep 10 00:40:37.912960 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 10 00:40:37.922817 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 10 00:40:37.941058 dracut-pre-trigger[418]: rd.md=0: removing MD RAID activation Sep 10 00:40:37.988662 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 10 00:40:38.001619 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 10 00:40:38.094955 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 10 00:40:38.101503 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 10 00:40:38.123959 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 10 00:40:38.129085 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 10 00:40:38.132094 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 10 00:40:38.134751 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 10 00:40:38.146647 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 10 00:40:38.145470 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 10 00:40:38.152843 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 10 00:40:38.160378 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 10 00:40:38.169290 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 10 00:40:38.169365 kernel: GPT:9289727 != 19775487 Sep 10 00:40:38.169380 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 10 00:40:38.169392 kernel: GPT:9289727 != 19775487 Sep 10 00:40:38.169403 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 10 00:40:38.170329 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:40:38.176276 kernel: cryptd: max_cpu_qlen set to 1000 Sep 10 00:40:38.176310 kernel: libata version 3.00 loaded. Sep 10 00:40:38.180698 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 10 00:40:38.181715 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:40:38.190004 kernel: ahci 0000:00:1f.2: version 3.0 Sep 10 00:40:38.190368 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 10 00:40:38.186324 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 10 00:40:38.199405 kernel: AVX2 version of gcm_enc/dec engaged. Sep 10 00:40:38.199448 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Sep 10 00:40:38.199723 kernel: AES CTR mode by8 optimization enabled Sep 10 00:40:38.199751 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 10 00:40:38.199973 kernel: scsi host0: ahci Sep 10 00:40:38.200399 kernel: scsi host1: ahci Sep 10 00:40:38.201864 kernel: scsi host2: ahci Sep 10 00:40:38.202124 kernel: scsi host3: ahci Sep 10 00:40:38.202313 kernel: scsi host4: ahci Sep 10 00:40:38.187698 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 10 00:40:38.208167 kernel: scsi host5: ahci Sep 10 00:40:38.208473 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Sep 10 00:40:38.187983 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:40:38.227320 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Sep 10 00:40:38.227346 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Sep 10 00:40:38.227359 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Sep 10 00:40:38.227370 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Sep 10 00:40:38.227382 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Sep 10 00:40:38.227393 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (467) Sep 10 00:40:38.227412 kernel: BTRFS: device fsid 47ffa5df-7ab2-4f1a-b68f-595717991426 devid 1 transid 34 /dev/vda3 scanned by (udev-worker) (470) Sep 10 00:40:38.194719 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:40:38.205713 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:40:38.240068 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 10 00:40:38.243371 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:40:38.255402 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 10 00:40:38.270923 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 10 00:40:38.274235 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 10 00:40:38.279312 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 10 00:40:38.299564 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 10 00:40:38.302976 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 10 00:40:38.330943 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:40:38.344634 disk-uuid[556]: Primary Header is updated. Sep 10 00:40:38.344634 disk-uuid[556]: Secondary Entries is updated. Sep 10 00:40:38.344634 disk-uuid[556]: Secondary Header is updated. Sep 10 00:40:38.349287 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:40:38.355296 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:40:38.526322 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 10 00:40:38.526422 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 10 00:40:38.527587 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 10 00:40:38.528554 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 10 00:40:38.528650 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 10 00:40:38.529296 kernel: ata3.00: applying bridge limits Sep 10 00:40:38.530289 kernel: ata3.00: configured for UDMA/100 Sep 10 00:40:38.532306 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 10 00:40:38.535288 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 10 00:40:38.535333 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 10 00:40:38.580329 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 10 00:40:38.580730 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 10 00:40:38.594325 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 10 00:40:39.358296 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:40:39.358903 disk-uuid[565]: The operation has completed successfully. Sep 10 00:40:39.393406 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 10 00:40:39.393548 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 10 00:40:39.421474 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 10 00:40:39.426645 sh[593]: Success Sep 10 00:40:39.440270 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Sep 10 00:40:39.478920 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 10 00:40:39.529268 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 10 00:40:39.531520 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 10 00:40:39.546325 kernel: BTRFS info (device dm-0): first mount of filesystem 47ffa5df-7ab2-4f1a-b68f-595717991426 Sep 10 00:40:39.546389 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:40:39.561889 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 10 00:40:39.561942 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 10 00:40:39.562816 kernel: BTRFS info (device dm-0): using free space tree Sep 10 00:40:39.569214 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 10 00:40:39.572551 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 10 00:40:39.583564 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 10 00:40:39.586973 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 10 00:40:39.625771 kernel: BTRFS info (device vda6): first mount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:40:39.625870 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:40:39.625883 kernel: BTRFS info (device vda6): using free space tree Sep 10 00:40:39.630465 kernel: BTRFS info (device vda6): auto enabling async discard Sep 10 00:40:39.643417 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 10 00:40:39.646295 kernel: BTRFS info (device vda6): last unmount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:40:39.657554 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 10 00:40:39.662457 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 10 00:40:39.737212 ignition[680]: Ignition 2.19.0 Sep 10 00:40:39.737229 ignition[680]: Stage: fetch-offline Sep 10 00:40:39.737298 ignition[680]: no configs at "/usr/lib/ignition/base.d" Sep 10 00:40:39.737313 ignition[680]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:40:39.737470 ignition[680]: parsed url from cmdline: "" Sep 10 00:40:39.737476 ignition[680]: no config URL provided Sep 10 00:40:39.737484 ignition[680]: reading system config file "/usr/lib/ignition/user.ign" Sep 10 00:40:39.737497 ignition[680]: no config at "/usr/lib/ignition/user.ign" Sep 10 00:40:39.737536 ignition[680]: op(1): [started] loading QEMU firmware config module Sep 10 00:40:39.737579 ignition[680]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 10 00:40:39.752491 ignition[680]: op(1): [finished] loading QEMU firmware config module Sep 10 00:40:39.752534 ignition[680]: QEMU firmware config was not found. Ignoring... Sep 10 00:40:39.799270 ignition[680]: parsing config with SHA512: 7d427afba2ee03891920f130b314bca6ee533ef1d0822cf3505c109c152d039d1e69e07c42ee810ba058d409b4599e7d3e049896debaaabafb7ad9bd651f0929 Sep 10 00:40:39.807011 unknown[680]: fetched base config from "system" Sep 10 00:40:39.807884 ignition[680]: fetch-offline: fetch-offline passed Sep 10 00:40:39.807033 unknown[680]: fetched user config from "qemu" Sep 10 00:40:39.808379 ignition[680]: Ignition finished successfully Sep 10 00:40:39.810766 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 10 00:40:39.914001 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 10 00:40:39.928591 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 10 00:40:39.954347 systemd-networkd[783]: lo: Link UP Sep 10 00:40:39.954360 systemd-networkd[783]: lo: Gained carrier Sep 10 00:40:39.956233 systemd-networkd[783]: Enumeration completed Sep 10 00:40:39.956764 systemd-networkd[783]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:40:39.956769 systemd-networkd[783]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 10 00:40:39.958010 systemd-networkd[783]: eth0: Link UP Sep 10 00:40:39.958014 systemd-networkd[783]: eth0: Gained carrier Sep 10 00:40:39.958022 systemd-networkd[783]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:40:39.958198 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 10 00:40:39.963676 systemd[1]: Reached target network.target - Network. Sep 10 00:40:39.970321 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 10 00:40:39.971336 systemd-networkd[783]: eth0: DHCPv4 address 10.0.0.97/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 10 00:40:39.983769 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 10 00:40:40.000892 ignition[787]: Ignition 2.19.0 Sep 10 00:40:40.000905 ignition[787]: Stage: kargs Sep 10 00:40:40.001135 ignition[787]: no configs at "/usr/lib/ignition/base.d" Sep 10 00:40:40.001150 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:40:40.005967 ignition[787]: kargs: kargs passed Sep 10 00:40:40.006047 ignition[787]: Ignition finished successfully Sep 10 00:40:40.010557 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 10 00:40:40.031597 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 10 00:40:40.058073 ignition[795]: Ignition 2.19.0 Sep 10 00:40:40.058087 ignition[795]: Stage: disks Sep 10 00:40:40.058403 ignition[795]: no configs at "/usr/lib/ignition/base.d" Sep 10 00:40:40.058421 ignition[795]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:40:40.063092 ignition[795]: disks: disks passed Sep 10 00:40:40.063164 ignition[795]: Ignition finished successfully Sep 10 00:40:40.067442 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 10 00:40:40.070006 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 10 00:40:40.072498 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 10 00:40:40.075357 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 10 00:40:40.077607 systemd[1]: Reached target sysinit.target - System Initialization. Sep 10 00:40:40.079900 systemd[1]: Reached target basic.target - Basic System. Sep 10 00:40:40.095483 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 10 00:40:40.115066 systemd-fsck[805]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 10 00:40:40.123719 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 10 00:40:40.138574 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 10 00:40:40.262302 kernel: EXT4-fs (vda9): mounted filesystem 0a9bf3c7-f8cd-4d40-b949-283957ba2f96 r/w with ordered data mode. Quota mode: none. Sep 10 00:40:40.263213 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 10 00:40:40.264430 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 10 00:40:40.272468 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 10 00:40:40.274497 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 10 00:40:40.275011 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 10 00:40:40.280698 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (813) Sep 10 00:40:40.275058 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 10 00:40:40.285780 kernel: BTRFS info (device vda6): first mount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:40:40.285797 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:40:40.285809 kernel: BTRFS info (device vda6): using free space tree Sep 10 00:40:40.275083 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 10 00:40:40.288298 kernel: BTRFS info (device vda6): auto enabling async discard Sep 10 00:40:40.290613 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 10 00:40:40.300091 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 10 00:40:40.302468 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 10 00:40:40.346353 initrd-setup-root[837]: cut: /sysroot/etc/passwd: No such file or directory Sep 10 00:40:40.351562 initrd-setup-root[844]: cut: /sysroot/etc/group: No such file or directory Sep 10 00:40:40.357374 initrd-setup-root[851]: cut: /sysroot/etc/shadow: No such file or directory Sep 10 00:40:40.361790 initrd-setup-root[858]: cut: /sysroot/etc/gshadow: No such file or directory Sep 10 00:40:40.457842 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 10 00:40:40.468393 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 10 00:40:40.471559 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 10 00:40:40.480283 kernel: BTRFS info (device vda6): last unmount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:40:40.498674 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 10 00:40:40.544832 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 10 00:40:40.584814 ignition[929]: INFO : Ignition 2.19.0 Sep 10 00:40:40.584814 ignition[929]: INFO : Stage: mount Sep 10 00:40:40.587020 ignition[929]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 10 00:40:40.587020 ignition[929]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:40:40.587020 ignition[929]: INFO : mount: mount passed Sep 10 00:40:40.587020 ignition[929]: INFO : Ignition finished successfully Sep 10 00:40:40.590506 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 10 00:40:40.602379 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 10 00:40:40.610191 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 10 00:40:40.625133 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (939) Sep 10 00:40:40.625209 kernel: BTRFS info (device vda6): first mount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:40:40.625227 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:40:40.626021 kernel: BTRFS info (device vda6): using free space tree Sep 10 00:40:40.630323 kernel: BTRFS info (device vda6): auto enabling async discard Sep 10 00:40:40.631625 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 10 00:40:40.658939 ignition[956]: INFO : Ignition 2.19.0 Sep 10 00:40:40.658939 ignition[956]: INFO : Stage: files Sep 10 00:40:40.661001 ignition[956]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 10 00:40:40.661001 ignition[956]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:40:40.661001 ignition[956]: DEBUG : files: compiled without relabeling support, skipping Sep 10 00:40:40.664792 ignition[956]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 10 00:40:40.664792 ignition[956]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 10 00:40:40.667632 ignition[956]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 10 00:40:40.669137 ignition[956]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 10 00:40:40.669137 ignition[956]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 10 00:40:40.668493 unknown[956]: wrote ssh authorized keys file for user: core Sep 10 00:40:40.673439 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 10 00:40:40.673439 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Sep 10 00:40:40.714590 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 10 00:40:40.873546 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 10 00:40:40.873546 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 10 00:40:40.877321 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 10 00:40:41.161564 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 10 00:40:41.283623 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:40:41.285995 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Sep 10 00:40:41.556017 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 10 00:40:42.035508 systemd-networkd[783]: eth0: Gained IPv6LL Sep 10 00:40:42.640205 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:40:42.640205 ignition[956]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Sep 10 00:40:42.645000 ignition[956]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Sep 10 00:40:42.691957 ignition[956]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 10 00:40:42.700022 ignition[956]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 10 00:40:42.701921 ignition[956]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Sep 10 00:40:42.701921 ignition[956]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Sep 10 00:40:42.701921 ignition[956]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Sep 10 00:40:42.701921 ignition[956]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 10 00:40:42.701921 ignition[956]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 10 00:40:42.701921 ignition[956]: INFO : files: files passed Sep 10 00:40:42.701921 ignition[956]: INFO : Ignition finished successfully Sep 10 00:40:42.703828 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 10 00:40:42.719463 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 10 00:40:42.722680 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 10 00:40:42.724859 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 10 00:40:42.725011 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 10 00:40:42.734731 initrd-setup-root-after-ignition[984]: grep: /sysroot/oem/oem-release: No such file or directory Sep 10 00:40:42.737912 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 10 00:40:42.737912 initrd-setup-root-after-ignition[986]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 10 00:40:42.741624 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 10 00:40:42.741870 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 10 00:40:42.744923 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 10 00:40:42.760584 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 10 00:40:42.810632 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 10 00:40:42.810790 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 10 00:40:42.813207 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 10 00:40:42.815193 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 10 00:40:42.817174 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 10 00:40:42.818440 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 10 00:40:42.851455 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 10 00:40:42.865558 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 10 00:40:42.875990 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 10 00:40:42.877608 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 10 00:40:42.880062 systemd[1]: Stopped target timers.target - Timer Units. Sep 10 00:40:42.882211 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 10 00:40:42.882398 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 10 00:40:42.884759 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 10 00:40:42.886477 systemd[1]: Stopped target basic.target - Basic System. Sep 10 00:40:42.888880 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 10 00:40:42.890862 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 10 00:40:42.892896 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 10 00:40:42.895109 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 10 00:40:42.897250 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 10 00:40:42.899435 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 10 00:40:42.901394 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 10 00:40:42.903521 systemd[1]: Stopped target swap.target - Swaps. Sep 10 00:40:42.905352 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 10 00:40:42.905566 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 10 00:40:42.907660 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 10 00:40:42.909340 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 10 00:40:42.911381 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 10 00:40:42.911538 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 10 00:40:42.913589 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 10 00:40:42.913782 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 10 00:40:42.915890 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 10 00:40:42.916018 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 10 00:40:42.917974 systemd[1]: Stopped target paths.target - Path Units. Sep 10 00:40:42.919748 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 10 00:40:42.923381 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 10 00:40:42.925555 systemd[1]: Stopped target slices.target - Slice Units. Sep 10 00:40:42.927519 systemd[1]: Stopped target sockets.target - Socket Units. Sep 10 00:40:42.929338 systemd[1]: iscsid.socket: Deactivated successfully. Sep 10 00:40:42.929497 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 10 00:40:42.931288 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 10 00:40:42.931421 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 10 00:40:42.933671 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 10 00:40:42.933807 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 10 00:40:42.935696 systemd[1]: ignition-files.service: Deactivated successfully. Sep 10 00:40:42.935810 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 10 00:40:42.946630 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 10 00:40:42.948683 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 10 00:40:42.948858 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 10 00:40:42.952039 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 10 00:40:42.953043 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 10 00:40:42.953193 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 10 00:40:42.955275 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 10 00:40:42.955417 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 10 00:40:42.961116 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 10 00:40:42.961274 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 10 00:40:42.966935 ignition[1010]: INFO : Ignition 2.19.0 Sep 10 00:40:42.969449 ignition[1010]: INFO : Stage: umount Sep 10 00:40:42.969449 ignition[1010]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 10 00:40:42.969449 ignition[1010]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:40:42.969449 ignition[1010]: INFO : umount: umount passed Sep 10 00:40:42.969449 ignition[1010]: INFO : Ignition finished successfully Sep 10 00:40:42.972022 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 10 00:40:42.972173 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 10 00:40:42.973916 systemd[1]: Stopped target network.target - Network. Sep 10 00:40:42.975300 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 10 00:40:42.975388 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 10 00:40:42.977559 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 10 00:40:42.977630 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 10 00:40:42.979581 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 10 00:40:42.979655 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 10 00:40:42.981625 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 10 00:40:42.981681 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 10 00:40:42.983715 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 10 00:40:42.985985 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 10 00:40:42.989319 systemd-networkd[783]: eth0: DHCPv6 lease lost Sep 10 00:40:42.989948 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 10 00:40:42.993130 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 10 00:40:42.993339 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 10 00:40:42.995681 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 10 00:40:42.995740 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 10 00:40:43.006524 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 10 00:40:43.007830 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 10 00:40:43.008059 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 10 00:40:43.010943 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 10 00:40:43.013792 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 10 00:40:43.013987 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 10 00:40:43.021662 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 10 00:40:43.021813 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:40:43.022392 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 10 00:40:43.022446 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 10 00:40:43.023111 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 10 00:40:43.023166 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 10 00:40:43.034669 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 10 00:40:43.035702 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 10 00:40:43.037896 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 10 00:40:43.039036 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 10 00:40:43.042806 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 10 00:40:43.043818 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 10 00:40:43.045922 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 10 00:40:43.046886 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 10 00:40:43.049077 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 10 00:40:43.050071 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 10 00:40:43.052367 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 10 00:40:43.053330 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 10 00:40:43.055500 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 10 00:40:43.056498 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:40:43.070447 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 10 00:40:43.104841 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 10 00:40:43.104974 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 10 00:40:43.105562 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 10 00:40:43.105622 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:40:43.106444 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 10 00:40:43.106568 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 10 00:40:43.374147 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 10 00:40:43.374371 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 10 00:40:43.375792 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 10 00:40:43.379091 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 10 00:40:43.379197 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 10 00:40:43.391531 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 10 00:40:43.400868 systemd[1]: Switching root. Sep 10 00:40:43.444726 systemd-journald[192]: Journal stopped Sep 10 00:40:45.140626 systemd-journald[192]: Received SIGTERM from PID 1 (systemd). Sep 10 00:40:45.140712 kernel: SELinux: policy capability network_peer_controls=1 Sep 10 00:40:45.140733 kernel: SELinux: policy capability open_perms=1 Sep 10 00:40:45.140750 kernel: SELinux: policy capability extended_socket_class=1 Sep 10 00:40:45.140770 kernel: SELinux: policy capability always_check_network=0 Sep 10 00:40:45.140782 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 10 00:40:45.140806 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 10 00:40:45.140820 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 10 00:40:45.140838 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 10 00:40:45.140850 kernel: audit: type=1403 audit(1757464844.002:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 10 00:40:45.140863 systemd[1]: Successfully loaded SELinux policy in 41.488ms. Sep 10 00:40:45.140882 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 14.044ms. Sep 10 00:40:45.140895 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 10 00:40:45.140908 systemd[1]: Detected virtualization kvm. Sep 10 00:40:45.140920 systemd[1]: Detected architecture x86-64. Sep 10 00:40:45.140932 systemd[1]: Detected first boot. Sep 10 00:40:45.140944 systemd[1]: Initializing machine ID from VM UUID. Sep 10 00:40:45.140956 zram_generator::config[1054]: No configuration found. Sep 10 00:40:45.140970 systemd[1]: Populated /etc with preset unit settings. Sep 10 00:40:45.140982 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 10 00:40:45.140998 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 10 00:40:45.141010 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 10 00:40:45.141023 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 10 00:40:45.141035 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 10 00:40:45.141048 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 10 00:40:45.141060 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 10 00:40:45.141072 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 10 00:40:45.141085 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 10 00:40:45.141102 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 10 00:40:45.141114 systemd[1]: Created slice user.slice - User and Session Slice. Sep 10 00:40:45.141127 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 10 00:40:45.141139 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 10 00:40:45.141151 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 10 00:40:45.141164 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 10 00:40:45.141176 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 10 00:40:45.141189 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 10 00:40:45.141201 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 10 00:40:45.141216 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 10 00:40:45.141229 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 10 00:40:45.141241 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 10 00:40:45.141253 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 10 00:40:45.141279 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 10 00:40:45.141292 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 10 00:40:45.141305 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 10 00:40:45.141318 systemd[1]: Reached target slices.target - Slice Units. Sep 10 00:40:45.141335 systemd[1]: Reached target swap.target - Swaps. Sep 10 00:40:45.141347 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 10 00:40:45.141364 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 10 00:40:45.141376 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 10 00:40:45.141390 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 10 00:40:45.141402 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 10 00:40:45.141417 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 10 00:40:45.141430 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 10 00:40:45.141444 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 10 00:40:45.141459 systemd[1]: Mounting media.mount - External Media Directory... Sep 10 00:40:45.141472 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:45.141485 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 10 00:40:45.141497 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 10 00:40:45.141509 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 10 00:40:45.141522 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 10 00:40:45.141534 systemd[1]: Reached target machines.target - Containers. Sep 10 00:40:45.141546 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 10 00:40:45.141562 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:40:45.141574 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 10 00:40:45.141586 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 10 00:40:45.141599 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:40:45.141611 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 10 00:40:45.141623 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:40:45.141635 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 10 00:40:45.141647 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:40:45.141661 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 10 00:40:45.141676 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 10 00:40:45.141688 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 10 00:40:45.141700 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 10 00:40:45.141712 systemd[1]: Stopped systemd-fsck-usr.service. Sep 10 00:40:45.141724 kernel: loop: module loaded Sep 10 00:40:45.141735 kernel: fuse: init (API version 7.39) Sep 10 00:40:45.141747 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 10 00:40:45.141769 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 10 00:40:45.141782 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 10 00:40:45.141797 kernel: ACPI: bus type drm_connector registered Sep 10 00:40:45.141809 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 10 00:40:45.141821 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 10 00:40:45.141834 systemd[1]: verity-setup.service: Deactivated successfully. Sep 10 00:40:45.141846 systemd[1]: Stopped verity-setup.service. Sep 10 00:40:45.141865 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:45.141896 systemd-journald[1138]: Collecting audit messages is disabled. Sep 10 00:40:45.141922 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 10 00:40:45.141935 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 10 00:40:45.141949 systemd-journald[1138]: Journal started Sep 10 00:40:45.141974 systemd-journald[1138]: Runtime Journal (/run/log/journal/a38dacc556de4cb9877b5cf329df81ce) is 6.0M, max 48.3M, 42.2M free. Sep 10 00:40:44.674760 systemd[1]: Queued start job for default target multi-user.target. Sep 10 00:40:44.693499 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 10 00:40:44.694060 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 10 00:40:45.145288 systemd[1]: Started systemd-journald.service - Journal Service. Sep 10 00:40:45.146435 systemd[1]: Mounted media.mount - External Media Directory. Sep 10 00:40:45.147573 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 10 00:40:45.148798 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 10 00:40:45.150036 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 10 00:40:45.151377 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 10 00:40:45.152869 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 10 00:40:45.154494 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 10 00:40:45.154717 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 10 00:40:45.156416 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:40:45.156646 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:40:45.158442 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 10 00:40:45.158666 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 10 00:40:45.160112 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:40:45.160354 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:40:45.161915 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 10 00:40:45.162128 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 10 00:40:45.163586 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:40:45.163814 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:40:45.165197 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 10 00:40:45.166637 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 10 00:40:45.168155 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 10 00:40:45.185165 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 10 00:40:45.205357 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 10 00:40:45.207825 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 10 00:40:45.208971 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 10 00:40:45.209016 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 10 00:40:45.211415 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 10 00:40:45.214213 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 10 00:40:45.217771 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 10 00:40:45.219133 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:40:45.223197 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 10 00:40:45.252011 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 10 00:40:45.253607 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 10 00:40:45.257709 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 10 00:40:45.260977 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 10 00:40:45.263637 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:40:45.280444 systemd-journald[1138]: Time spent on flushing to /var/log/journal/a38dacc556de4cb9877b5cf329df81ce is 25.414ms for 993 entries. Sep 10 00:40:45.280444 systemd-journald[1138]: System Journal (/var/log/journal/a38dacc556de4cb9877b5cf329df81ce) is 8.0M, max 195.6M, 187.6M free. Sep 10 00:40:45.324088 systemd-journald[1138]: Received client request to flush runtime journal. Sep 10 00:40:45.324139 kernel: loop0: detected capacity change from 0 to 229808 Sep 10 00:40:45.324155 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 10 00:40:45.268630 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 10 00:40:45.273605 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 10 00:40:45.278196 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 10 00:40:45.288124 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 10 00:40:45.290929 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 10 00:40:45.294343 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 10 00:40:45.303159 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 10 00:40:45.316600 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 10 00:40:45.327747 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 10 00:40:45.357869 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 10 00:40:45.369590 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 10 00:40:45.380953 udevadm[1184]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Sep 10 00:40:45.421356 kernel: loop1: detected capacity change from 0 to 140768 Sep 10 00:40:45.424784 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:40:45.438999 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 10 00:40:45.450551 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 10 00:40:45.496292 kernel: loop2: detected capacity change from 0 to 142488 Sep 10 00:40:45.556125 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Sep 10 00:40:45.556146 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Sep 10 00:40:45.562871 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 10 00:40:45.595283 kernel: loop3: detected capacity change from 0 to 229808 Sep 10 00:40:45.611290 kernel: loop4: detected capacity change from 0 to 140768 Sep 10 00:40:45.683282 kernel: loop5: detected capacity change from 0 to 142488 Sep 10 00:40:45.694966 (sd-merge)[1193]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 10 00:40:45.695678 (sd-merge)[1193]: Merged extensions into '/usr'. Sep 10 00:40:45.768677 systemd[1]: Reloading requested from client PID 1168 ('systemd-sysext') (unit systemd-sysext.service)... Sep 10 00:40:45.768700 systemd[1]: Reloading... Sep 10 00:40:45.858313 zram_generator::config[1220]: No configuration found. Sep 10 00:40:45.985221 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:40:46.037813 ldconfig[1163]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 10 00:40:46.051194 systemd[1]: Reloading finished in 281 ms. Sep 10 00:40:46.094358 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 10 00:40:46.096351 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 10 00:40:46.098134 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 10 00:40:46.110696 systemd[1]: Starting ensure-sysext.service... Sep 10 00:40:46.113671 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 10 00:40:46.115861 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 10 00:40:46.122344 systemd[1]: Reloading requested from client PID 1257 ('systemctl') (unit ensure-sysext.service)... Sep 10 00:40:46.122512 systemd[1]: Reloading... Sep 10 00:40:46.139959 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 10 00:40:46.140858 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 10 00:40:46.142039 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 10 00:40:46.142454 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Sep 10 00:40:46.142612 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Sep 10 00:40:46.146356 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Sep 10 00:40:46.146483 systemd-tmpfiles[1258]: Skipping /boot Sep 10 00:40:46.158317 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Sep 10 00:40:46.161625 systemd-tmpfiles[1258]: Skipping /boot Sep 10 00:40:46.194297 zram_generator::config[1286]: No configuration found. Sep 10 00:40:46.331875 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:40:46.384734 systemd[1]: Reloading finished in 261 ms. Sep 10 00:40:46.403599 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 10 00:40:46.423869 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 10 00:40:46.444623 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 10 00:40:46.447207 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 10 00:40:46.452126 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 10 00:40:46.454671 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 10 00:40:46.517530 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 10 00:40:46.520155 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:46.520355 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:40:46.521842 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:40:46.524149 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:40:46.531360 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:40:46.533126 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:40:46.533294 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:46.534334 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:40:46.534571 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:40:46.536555 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:40:46.536743 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:40:46.539673 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:40:46.539899 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:40:46.544954 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:46.545277 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:40:46.554611 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:40:46.558576 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:40:46.565628 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:40:46.582908 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:40:46.583333 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:46.585404 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 10 00:40:46.587710 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:40:46.588073 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:40:46.589898 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:40:46.590094 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:40:46.591796 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:40:46.592019 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:40:46.600297 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 10 00:40:46.606118 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:46.606456 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:40:46.619788 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:40:46.623695 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 10 00:40:46.627667 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:40:46.633580 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:40:46.634842 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:40:46.635038 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:40:46.635853 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 10 00:40:46.639992 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:40:46.640192 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:40:46.642586 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:40:46.642776 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:40:46.644575 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 10 00:40:46.644802 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 10 00:40:46.646825 systemd[1]: Finished ensure-sysext.service. Sep 10 00:40:46.648135 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:40:46.648366 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:40:46.657421 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 10 00:40:46.657565 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 10 00:40:46.665621 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 10 00:40:46.668580 augenrules[1374]: No rules Sep 10 00:40:46.670970 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 10 00:40:46.690224 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 10 00:40:46.692651 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 10 00:40:46.702418 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 10 00:40:46.711469 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 10 00:40:46.715561 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 10 00:40:46.730323 systemd-resolved[1328]: Positive Trust Anchors: Sep 10 00:40:46.730742 systemd-resolved[1328]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 10 00:40:46.730826 systemd-resolved[1328]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 10 00:40:46.735622 systemd-resolved[1328]: Defaulting to hostname 'linux'. Sep 10 00:40:46.738178 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 10 00:40:46.740161 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 10 00:40:46.741636 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 10 00:40:46.751367 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 10 00:40:46.752906 systemd[1]: Reached target time-set.target - System Time Set. Sep 10 00:40:46.753984 systemd-udevd[1382]: Using default interface naming scheme 'v255'. Sep 10 00:40:46.778378 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 10 00:40:46.786559 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 10 00:40:46.826302 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (1392) Sep 10 00:40:46.826974 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 10 00:40:46.859230 systemd-networkd[1391]: lo: Link UP Sep 10 00:40:46.859249 systemd-networkd[1391]: lo: Gained carrier Sep 10 00:40:46.861429 systemd-networkd[1391]: Enumeration completed Sep 10 00:40:46.862109 systemd-networkd[1391]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:40:46.862123 systemd-networkd[1391]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 10 00:40:46.862824 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 10 00:40:46.863404 systemd-networkd[1391]: eth0: Link UP Sep 10 00:40:46.863419 systemd-networkd[1391]: eth0: Gained carrier Sep 10 00:40:46.863433 systemd-networkd[1391]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:40:46.866470 systemd[1]: Reached target network.target - Network. Sep 10 00:40:46.881559 systemd-networkd[1391]: eth0: DHCPv4 address 10.0.0.97/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 10 00:40:46.881673 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 10 00:40:46.884376 systemd-timesyncd[1372]: Network configuration changed, trying to establish connection. Sep 10 00:40:46.886195 systemd-networkd[1391]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:40:47.581482 systemd-timesyncd[1372]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 10 00:40:47.581535 systemd-timesyncd[1372]: Initial clock synchronization to Wed 2025-09-10 00:40:47.581347 UTC. Sep 10 00:40:47.582313 systemd-resolved[1328]: Clock change detected. Flushing caches. Sep 10 00:40:47.583717 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 10 00:40:47.590277 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Sep 10 00:40:47.593581 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 10 00:40:47.602322 kernel: ACPI: button: Power Button [PWRF] Sep 10 00:40:47.613091 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 10 00:40:47.616675 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Sep 10 00:40:47.617635 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 10 00:40:47.621586 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Sep 10 00:40:47.623105 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 10 00:40:47.664077 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Sep 10 00:40:47.722921 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:40:47.794372 kernel: mousedev: PS/2 mouse device common for all mice Sep 10 00:40:47.798995 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 10 00:40:47.799382 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:40:47.808067 kernel: kvm_amd: TSC scaling supported Sep 10 00:40:47.808111 kernel: kvm_amd: Nested Virtualization enabled Sep 10 00:40:47.808129 kernel: kvm_amd: Nested Paging enabled Sep 10 00:40:47.808145 kernel: kvm_amd: LBR virtualization supported Sep 10 00:40:47.809232 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 10 00:40:47.809277 kernel: kvm_amd: Virtual GIF supported Sep 10 00:40:47.812765 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:40:47.833281 kernel: EDAC MC: Ver: 3.0.0 Sep 10 00:40:47.867643 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 10 00:40:47.876655 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 10 00:40:47.878722 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:40:47.888214 lvm[1433]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 10 00:40:47.932992 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 10 00:40:47.934832 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 10 00:40:47.936033 systemd[1]: Reached target sysinit.target - System Initialization. Sep 10 00:40:47.937337 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 10 00:40:47.938727 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 10 00:40:47.940638 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 10 00:40:47.941992 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 10 00:40:47.943384 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 10 00:40:47.944761 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 10 00:40:47.944813 systemd[1]: Reached target paths.target - Path Units. Sep 10 00:40:47.945833 systemd[1]: Reached target timers.target - Timer Units. Sep 10 00:40:47.949369 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 10 00:40:47.953223 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 10 00:40:47.970288 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 10 00:40:47.973782 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 10 00:40:47.975880 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 10 00:40:47.977222 systemd[1]: Reached target sockets.target - Socket Units. Sep 10 00:40:47.978303 systemd[1]: Reached target basic.target - Basic System. Sep 10 00:40:47.979421 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 10 00:40:47.979457 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 10 00:40:47.981936 systemd[1]: Starting containerd.service - containerd container runtime... Sep 10 00:40:47.984980 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 10 00:40:47.987381 lvm[1438]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 10 00:40:47.990025 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 10 00:40:47.993433 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 10 00:40:47.995018 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 10 00:40:47.996424 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 10 00:40:47.999997 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 10 00:40:48.005433 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 10 00:40:48.005544 jq[1441]: false Sep 10 00:40:48.009905 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 10 00:40:48.014697 dbus-daemon[1440]: [system] SELinux support is enabled Sep 10 00:40:48.016711 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 10 00:40:48.020825 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 10 00:40:48.021578 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 10 00:40:48.023435 systemd[1]: Starting update-engine.service - Update Engine... Sep 10 00:40:48.026124 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 10 00:40:48.028433 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 10 00:40:48.033713 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 10 00:40:48.042666 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 10 00:40:48.042961 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 10 00:40:48.045422 extend-filesystems[1442]: Found loop3 Sep 10 00:40:48.076565 jq[1453]: true Sep 10 00:40:48.046214 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 10 00:40:48.076896 extend-filesystems[1442]: Found loop4 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found loop5 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found sr0 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda1 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda2 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda3 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found usr Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda4 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda6 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda7 Sep 10 00:40:48.076896 extend-filesystems[1442]: Found vda9 Sep 10 00:40:48.076896 extend-filesystems[1442]: Checking size of /dev/vda9 Sep 10 00:40:48.046487 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 10 00:40:48.078467 systemd[1]: motdgen.service: Deactivated successfully. Sep 10 00:40:48.078729 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 10 00:40:48.107230 update_engine[1451]: I20250910 00:40:48.101873 1451 main.cc:92] Flatcar Update Engine starting Sep 10 00:40:48.103441 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 10 00:40:48.103520 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 10 00:40:48.105300 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 10 00:40:48.105334 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 10 00:40:48.111875 jq[1462]: true Sep 10 00:40:48.111182 systemd-logind[1448]: Watching system buttons on /dev/input/event1 (Power Button) Sep 10 00:40:48.115957 tar[1460]: linux-amd64/LICENSE Sep 10 00:40:48.115957 tar[1460]: linux-amd64/helm Sep 10 00:40:48.116272 extend-filesystems[1442]: Resized partition /dev/vda9 Sep 10 00:40:48.129791 update_engine[1451]: I20250910 00:40:48.112353 1451 update_check_scheduler.cc:74] Next update check in 9m24s Sep 10 00:40:48.111204 systemd-logind[1448]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 10 00:40:48.130022 extend-filesystems[1477]: resize2fs 1.47.1 (20-May-2024) Sep 10 00:40:48.112264 systemd[1]: Started update-engine.service - Update Engine. Sep 10 00:40:48.113597 systemd-logind[1448]: New seat seat0. Sep 10 00:40:48.115750 (ntainerd)[1463]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 10 00:40:48.123203 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 10 00:40:48.124954 systemd[1]: Started systemd-logind.service - User Login Management. Sep 10 00:40:48.170584 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (1405) Sep 10 00:40:48.196489 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 10 00:40:48.272644 locksmithd[1479]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 10 00:40:48.287288 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 10 00:40:48.319979 extend-filesystems[1477]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 10 00:40:48.319979 extend-filesystems[1477]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 10 00:40:48.319979 extend-filesystems[1477]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 10 00:40:48.323738 extend-filesystems[1442]: Resized filesystem in /dev/vda9 Sep 10 00:40:48.324893 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 10 00:40:48.325134 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 10 00:40:48.329962 bash[1493]: Updated "/home/core/.ssh/authorized_keys" Sep 10 00:40:48.331594 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 10 00:40:48.335396 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 10 00:40:48.432147 sshd_keygen[1455]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 10 00:40:48.442463 containerd[1463]: time="2025-09-10T00:40:48.442346429Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Sep 10 00:40:48.464262 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 10 00:40:48.470702 containerd[1463]: time="2025-09-10T00:40:48.470634549Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.472721 containerd[1463]: time="2025-09-10T00:40:48.472657392Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.104-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:40:48.472721 containerd[1463]: time="2025-09-10T00:40:48.472717706Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 10 00:40:48.472785 containerd[1463]: time="2025-09-10T00:40:48.472736821Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 10 00:40:48.472978 containerd[1463]: time="2025-09-10T00:40:48.472954470Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 10 00:40:48.472978 containerd[1463]: time="2025-09-10T00:40:48.472977653Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473084 containerd[1463]: time="2025-09-10T00:40:48.473057533Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473084 containerd[1463]: time="2025-09-10T00:40:48.473076328Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473381 containerd[1463]: time="2025-09-10T00:40:48.473350041Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473381 containerd[1463]: time="2025-09-10T00:40:48.473379346Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473440 containerd[1463]: time="2025-09-10T00:40:48.473394415Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473440 containerd[1463]: time="2025-09-10T00:40:48.473405045Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473538 containerd[1463]: time="2025-09-10T00:40:48.473514580Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473809 containerd[1463]: time="2025-09-10T00:40:48.473785638Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473962 containerd[1463]: time="2025-09-10T00:40:48.473935600Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:40:48.473962 containerd[1463]: time="2025-09-10T00:40:48.473957631Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 10 00:40:48.474087 containerd[1463]: time="2025-09-10T00:40:48.474062277Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 10 00:40:48.474146 containerd[1463]: time="2025-09-10T00:40:48.474127459Z" level=info msg="metadata content store policy set" policy=shared Sep 10 00:40:48.479185 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 10 00:40:48.481710 containerd[1463]: time="2025-09-10T00:40:48.481647552Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 10 00:40:48.481779 containerd[1463]: time="2025-09-10T00:40:48.481748321Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 10 00:40:48.481801 containerd[1463]: time="2025-09-10T00:40:48.481779279Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 10 00:40:48.481820 containerd[1463]: time="2025-09-10T00:40:48.481802343Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 10 00:40:48.481841 containerd[1463]: time="2025-09-10T00:40:48.481820917Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 10 00:40:48.482067 containerd[1463]: time="2025-09-10T00:40:48.482040900Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 10 00:40:48.482452 containerd[1463]: time="2025-09-10T00:40:48.482415142Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 10 00:40:48.482605 containerd[1463]: time="2025-09-10T00:40:48.482580171Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 10 00:40:48.482634 containerd[1463]: time="2025-09-10T00:40:48.482604517Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 10 00:40:48.482634 containerd[1463]: time="2025-09-10T00:40:48.482618463Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 10 00:40:48.482670 containerd[1463]: time="2025-09-10T00:40:48.482633922Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482670 containerd[1463]: time="2025-09-10T00:40:48.482647738Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482670 containerd[1463]: time="2025-09-10T00:40:48.482659750Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482744 containerd[1463]: time="2025-09-10T00:40:48.482683395Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482744 containerd[1463]: time="2025-09-10T00:40:48.482698673Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482744 containerd[1463]: time="2025-09-10T00:40:48.482712730Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482744 containerd[1463]: time="2025-09-10T00:40:48.482725323Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482744 containerd[1463]: time="2025-09-10T00:40:48.482737486Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 10 00:40:48.482833 containerd[1463]: time="2025-09-10T00:40:48.482756812Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482833 containerd[1463]: time="2025-09-10T00:40:48.482770949Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482833 containerd[1463]: time="2025-09-10T00:40:48.482787720Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482833 containerd[1463]: time="2025-09-10T00:40:48.482799352Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482833 containerd[1463]: time="2025-09-10T00:40:48.482811014Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482833 containerd[1463]: time="2025-09-10T00:40:48.482824509Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482837043Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482849787Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482863793Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482878511Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482890774Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482902325Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.482945 containerd[1463]: time="2025-09-10T00:40:48.482915410Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.482948552Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.482970062Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.482980933Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.482991743Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.483042759Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.483064930Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 10 00:40:48.483080 containerd[1463]: time="2025-09-10T00:40:48.483080580Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 10 00:40:48.483206 containerd[1463]: time="2025-09-10T00:40:48.483097251Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 10 00:40:48.483206 containerd[1463]: time="2025-09-10T00:40:48.483111027Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.483206 containerd[1463]: time="2025-09-10T00:40:48.483127468Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 10 00:40:48.483206 containerd[1463]: time="2025-09-10T00:40:48.483165779Z" level=info msg="NRI interface is disabled by configuration." Sep 10 00:40:48.483206 containerd[1463]: time="2025-09-10T00:40:48.483178964Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 10 00:40:48.483812 containerd[1463]: time="2025-09-10T00:40:48.483712875Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 10 00:40:48.483957 containerd[1463]: time="2025-09-10T00:40:48.483811470Z" level=info msg="Connect containerd service" Sep 10 00:40:48.483957 containerd[1463]: time="2025-09-10T00:40:48.483869730Z" level=info msg="using legacy CRI server" Sep 10 00:40:48.483957 containerd[1463]: time="2025-09-10T00:40:48.483880640Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 10 00:40:48.484021 containerd[1463]: time="2025-09-10T00:40:48.484008600Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 10 00:40:48.485300 containerd[1463]: time="2025-09-10T00:40:48.484914870Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 10 00:40:48.485300 containerd[1463]: time="2025-09-10T00:40:48.485074930Z" level=info msg="Start subscribing containerd event" Sep 10 00:40:48.485300 containerd[1463]: time="2025-09-10T00:40:48.485143468Z" level=info msg="Start recovering state" Sep 10 00:40:48.485443 containerd[1463]: time="2025-09-10T00:40:48.485416831Z" level=info msg="Start event monitor" Sep 10 00:40:48.485443 containerd[1463]: time="2025-09-10T00:40:48.485439524Z" level=info msg="Start snapshots syncer" Sep 10 00:40:48.485505 containerd[1463]: time="2025-09-10T00:40:48.485452147Z" level=info msg="Start cni network conf syncer for default" Sep 10 00:40:48.485505 containerd[1463]: time="2025-09-10T00:40:48.485462567Z" level=info msg="Start streaming server" Sep 10 00:40:48.486207 containerd[1463]: time="2025-09-10T00:40:48.486160025Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 10 00:40:48.486283 containerd[1463]: time="2025-09-10T00:40:48.486233012Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 10 00:40:48.486405 containerd[1463]: time="2025-09-10T00:40:48.486379697Z" level=info msg="containerd successfully booted in 0.045356s" Sep 10 00:40:48.487662 systemd[1]: Started containerd.service - containerd container runtime. Sep 10 00:40:48.489540 systemd[1]: issuegen.service: Deactivated successfully. Sep 10 00:40:48.489805 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 10 00:40:48.501711 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 10 00:40:48.518497 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 10 00:40:48.528843 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 10 00:40:48.531732 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 10 00:40:48.533263 systemd[1]: Reached target getty.target - Login Prompts. Sep 10 00:40:48.637445 tar[1460]: linux-amd64/README.md Sep 10 00:40:48.652884 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 10 00:40:49.459481 systemd-networkd[1391]: eth0: Gained IPv6LL Sep 10 00:40:49.463226 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 10 00:40:49.465568 systemd[1]: Reached target network-online.target - Network is Online. Sep 10 00:40:49.481797 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 10 00:40:49.485355 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:40:49.488149 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 10 00:40:49.514759 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 10 00:40:49.515719 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 10 00:40:49.517949 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 10 00:40:49.520895 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 10 00:40:50.297436 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:40:50.299035 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 10 00:40:50.300326 systemd[1]: Startup finished in 1.279s (kernel) + 7.226s (initrd) + 5.645s (userspace) = 14.151s. Sep 10 00:40:50.303819 (kubelet)[1553]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:40:50.774724 kubelet[1553]: E0910 00:40:50.774536 1553 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:40:50.779709 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:40:50.779955 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:40:50.780389 systemd[1]: kubelet.service: Consumed 1.112s CPU time. Sep 10 00:40:51.281834 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 10 00:40:51.291767 systemd[1]: Started sshd@0-10.0.0.97:22-10.0.0.1:42562.service - OpenSSH per-connection server daemon (10.0.0.1:42562). Sep 10 00:40:51.339792 sshd[1567]: Accepted publickey for core from 10.0.0.1 port 42562 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:51.342236 sshd[1567]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:51.353447 systemd-logind[1448]: New session 1 of user core. Sep 10 00:40:51.355126 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 10 00:40:51.366759 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 10 00:40:51.382417 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 10 00:40:51.393659 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 10 00:40:51.397271 (systemd)[1571]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 10 00:40:51.512504 systemd[1571]: Queued start job for default target default.target. Sep 10 00:40:51.522995 systemd[1571]: Created slice app.slice - User Application Slice. Sep 10 00:40:51.523027 systemd[1571]: Reached target paths.target - Paths. Sep 10 00:40:51.523042 systemd[1571]: Reached target timers.target - Timers. Sep 10 00:40:51.524967 systemd[1571]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 10 00:40:51.541513 systemd[1571]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 10 00:40:51.541744 systemd[1571]: Reached target sockets.target - Sockets. Sep 10 00:40:51.541761 systemd[1571]: Reached target basic.target - Basic System. Sep 10 00:40:51.541815 systemd[1571]: Reached target default.target - Main User Target. Sep 10 00:40:51.541855 systemd[1571]: Startup finished in 136ms. Sep 10 00:40:51.542328 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 10 00:40:51.544534 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 10 00:40:51.611309 systemd[1]: Started sshd@1-10.0.0.97:22-10.0.0.1:42570.service - OpenSSH per-connection server daemon (10.0.0.1:42570). Sep 10 00:40:51.665282 sshd[1582]: Accepted publickey for core from 10.0.0.1 port 42570 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:51.667755 sshd[1582]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:51.673158 systemd-logind[1448]: New session 2 of user core. Sep 10 00:40:51.683402 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 10 00:40:51.742489 sshd[1582]: pam_unix(sshd:session): session closed for user core Sep 10 00:40:51.755065 systemd[1]: sshd@1-10.0.0.97:22-10.0.0.1:42570.service: Deactivated successfully. Sep 10 00:40:51.756966 systemd[1]: session-2.scope: Deactivated successfully. Sep 10 00:40:51.758524 systemd-logind[1448]: Session 2 logged out. Waiting for processes to exit. Sep 10 00:40:51.775530 systemd[1]: Started sshd@2-10.0.0.97:22-10.0.0.1:42582.service - OpenSSH per-connection server daemon (10.0.0.1:42582). Sep 10 00:40:51.776498 systemd-logind[1448]: Removed session 2. Sep 10 00:40:51.811312 sshd[1589]: Accepted publickey for core from 10.0.0.1 port 42582 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:51.813428 sshd[1589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:51.819204 systemd-logind[1448]: New session 3 of user core. Sep 10 00:40:51.830636 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 10 00:40:51.883835 sshd[1589]: pam_unix(sshd:session): session closed for user core Sep 10 00:40:51.901617 systemd[1]: sshd@2-10.0.0.97:22-10.0.0.1:42582.service: Deactivated successfully. Sep 10 00:40:51.903691 systemd[1]: session-3.scope: Deactivated successfully. Sep 10 00:40:51.905235 systemd-logind[1448]: Session 3 logged out. Waiting for processes to exit. Sep 10 00:40:51.914609 systemd[1]: Started sshd@3-10.0.0.97:22-10.0.0.1:42598.service - OpenSSH per-connection server daemon (10.0.0.1:42598). Sep 10 00:40:51.915621 systemd-logind[1448]: Removed session 3. Sep 10 00:40:51.948207 sshd[1596]: Accepted publickey for core from 10.0.0.1 port 42598 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:51.949857 sshd[1596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:51.954429 systemd-logind[1448]: New session 4 of user core. Sep 10 00:40:51.965393 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 10 00:40:52.023771 sshd[1596]: pam_unix(sshd:session): session closed for user core Sep 10 00:40:52.035088 systemd[1]: sshd@3-10.0.0.97:22-10.0.0.1:42598.service: Deactivated successfully. Sep 10 00:40:52.036888 systemd[1]: session-4.scope: Deactivated successfully. Sep 10 00:40:52.038720 systemd-logind[1448]: Session 4 logged out. Waiting for processes to exit. Sep 10 00:40:52.054771 systemd[1]: Started sshd@4-10.0.0.97:22-10.0.0.1:42612.service - OpenSSH per-connection server daemon (10.0.0.1:42612). Sep 10 00:40:52.056164 systemd-logind[1448]: Removed session 4. Sep 10 00:40:52.089112 sshd[1603]: Accepted publickey for core from 10.0.0.1 port 42612 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:52.091151 sshd[1603]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:52.095662 systemd-logind[1448]: New session 5 of user core. Sep 10 00:40:52.109383 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 10 00:40:52.173196 sudo[1606]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 10 00:40:52.173598 sudo[1606]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:40:52.197845 sudo[1606]: pam_unix(sudo:session): session closed for user root Sep 10 00:40:52.200512 sshd[1603]: pam_unix(sshd:session): session closed for user core Sep 10 00:40:52.214378 systemd[1]: sshd@4-10.0.0.97:22-10.0.0.1:42612.service: Deactivated successfully. Sep 10 00:40:52.217166 systemd[1]: session-5.scope: Deactivated successfully. Sep 10 00:40:52.219394 systemd-logind[1448]: Session 5 logged out. Waiting for processes to exit. Sep 10 00:40:52.237844 systemd[1]: Started sshd@5-10.0.0.97:22-10.0.0.1:42622.service - OpenSSH per-connection server daemon (10.0.0.1:42622). Sep 10 00:40:52.239264 systemd-logind[1448]: Removed session 5. Sep 10 00:40:52.273955 sshd[1611]: Accepted publickey for core from 10.0.0.1 port 42622 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:52.276106 sshd[1611]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:52.280786 systemd-logind[1448]: New session 6 of user core. Sep 10 00:40:52.296474 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 10 00:40:52.352595 sudo[1615]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 10 00:40:52.352972 sudo[1615]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:40:52.357854 sudo[1615]: pam_unix(sudo:session): session closed for user root Sep 10 00:40:52.365046 sudo[1614]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 10 00:40:52.365450 sudo[1614]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:40:52.385554 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 10 00:40:52.388109 auditctl[1618]: No rules Sep 10 00:40:52.389869 systemd[1]: audit-rules.service: Deactivated successfully. Sep 10 00:40:52.390208 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 10 00:40:52.392539 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 10 00:40:52.429276 augenrules[1636]: No rules Sep 10 00:40:52.431311 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 10 00:40:52.434030 sudo[1614]: pam_unix(sudo:session): session closed for user root Sep 10 00:40:52.436385 sshd[1611]: pam_unix(sshd:session): session closed for user core Sep 10 00:40:52.454061 systemd[1]: sshd@5-10.0.0.97:22-10.0.0.1:42622.service: Deactivated successfully. Sep 10 00:40:52.456590 systemd[1]: session-6.scope: Deactivated successfully. Sep 10 00:40:52.458872 systemd-logind[1448]: Session 6 logged out. Waiting for processes to exit. Sep 10 00:40:52.468614 systemd[1]: Started sshd@6-10.0.0.97:22-10.0.0.1:42634.service - OpenSSH per-connection server daemon (10.0.0.1:42634). Sep 10 00:40:52.469951 systemd-logind[1448]: Removed session 6. Sep 10 00:40:52.504410 sshd[1644]: Accepted publickey for core from 10.0.0.1 port 42634 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:40:52.506195 sshd[1644]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:40:52.519681 systemd-logind[1448]: New session 7 of user core. Sep 10 00:40:52.541633 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 10 00:40:52.607088 sudo[1647]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 10 00:40:52.607491 sudo[1647]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:40:54.763598 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 10 00:40:54.770378 (dockerd)[1665]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 10 00:40:56.423451 dockerd[1665]: time="2025-09-10T00:40:56.423359307Z" level=info msg="Starting up" Sep 10 00:40:57.385738 dockerd[1665]: time="2025-09-10T00:40:57.385630969Z" level=info msg="Loading containers: start." Sep 10 00:40:57.557275 kernel: Initializing XFRM netlink socket Sep 10 00:40:57.661764 systemd-networkd[1391]: docker0: Link UP Sep 10 00:40:57.692394 dockerd[1665]: time="2025-09-10T00:40:57.692302698Z" level=info msg="Loading containers: done." Sep 10 00:40:57.713835 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1877540148-merged.mount: Deactivated successfully. Sep 10 00:40:57.716832 dockerd[1665]: time="2025-09-10T00:40:57.716746209Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 10 00:40:57.716959 dockerd[1665]: time="2025-09-10T00:40:57.716929563Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 10 00:40:57.717177 dockerd[1665]: time="2025-09-10T00:40:57.717128275Z" level=info msg="Daemon has completed initialization" Sep 10 00:40:57.760073 dockerd[1665]: time="2025-09-10T00:40:57.759949310Z" level=info msg="API listen on /run/docker.sock" Sep 10 00:40:57.760379 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 10 00:40:58.854539 containerd[1463]: time="2025-09-10T00:40:58.854484113Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.4\"" Sep 10 00:40:59.837945 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount563459072.mount: Deactivated successfully. Sep 10 00:41:00.905356 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 10 00:41:00.914770 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:01.103163 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:01.111377 (kubelet)[1871]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:41:01.861146 kubelet[1871]: E0910 00:41:01.860928 1871 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:41:01.868802 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:41:01.869042 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:41:02.759320 containerd[1463]: time="2025-09-10T00:41:02.759206111Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:02.760590 containerd[1463]: time="2025-09-10T00:41:02.760495178Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.4: active requests=0, bytes read=30078664" Sep 10 00:41:02.761928 containerd[1463]: time="2025-09-10T00:41:02.761883292Z" level=info msg="ImageCreate event name:\"sha256:1f41885d0a91155d5a5e670b2862eed338c7f12b0e8a5bbc88b1ab4a2d505ae8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:02.765475 containerd[1463]: time="2025-09-10T00:41:02.765392152Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:0d441d0d347145b3f02f20cb313239cdae86067643d7f70803fab8bac2d28876\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:02.767023 containerd[1463]: time="2025-09-10T00:41:02.766962106Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.4\" with image id \"sha256:1f41885d0a91155d5a5e670b2862eed338c7f12b0e8a5bbc88b1ab4a2d505ae8\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:0d441d0d347145b3f02f20cb313239cdae86067643d7f70803fab8bac2d28876\", size \"30075464\" in 3.912426787s" Sep 10 00:41:02.767023 containerd[1463]: time="2025-09-10T00:41:02.767017781Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.4\" returns image reference \"sha256:1f41885d0a91155d5a5e670b2862eed338c7f12b0e8a5bbc88b1ab4a2d505ae8\"" Sep 10 00:41:02.767824 containerd[1463]: time="2025-09-10T00:41:02.767780581Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.4\"" Sep 10 00:41:05.300819 containerd[1463]: time="2025-09-10T00:41:05.300675463Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:05.301599 containerd[1463]: time="2025-09-10T00:41:05.301505279Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.4: active requests=0, bytes read=26018066" Sep 10 00:41:05.302877 containerd[1463]: time="2025-09-10T00:41:05.302832228Z" level=info msg="ImageCreate event name:\"sha256:358ab71c1a1ea4846ad0b3dff0d9db6b124236b64bc8a6b79dc874f65dc0d492\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:05.306750 containerd[1463]: time="2025-09-10T00:41:05.306668071Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:bd22c2af2f30a8f818568b4d5fe131098fdd38267e9e07872cfc33e8f5876bc3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:05.308015 containerd[1463]: time="2025-09-10T00:41:05.307959854Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.4\" with image id \"sha256:358ab71c1a1ea4846ad0b3dff0d9db6b124236b64bc8a6b79dc874f65dc0d492\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:bd22c2af2f30a8f818568b4d5fe131098fdd38267e9e07872cfc33e8f5876bc3\", size \"27646961\" in 2.540131714s" Sep 10 00:41:05.308074 containerd[1463]: time="2025-09-10T00:41:05.308012513Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.4\" returns image reference \"sha256:358ab71c1a1ea4846ad0b3dff0d9db6b124236b64bc8a6b79dc874f65dc0d492\"" Sep 10 00:41:05.308682 containerd[1463]: time="2025-09-10T00:41:05.308642664Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.4\"" Sep 10 00:41:07.883835 containerd[1463]: time="2025-09-10T00:41:07.883727337Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:07.884686 containerd[1463]: time="2025-09-10T00:41:07.884616084Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.4: active requests=0, bytes read=20153911" Sep 10 00:41:07.886011 containerd[1463]: time="2025-09-10T00:41:07.885960716Z" level=info msg="ImageCreate event name:\"sha256:ab4ad8a84c3c69c18494ef32fa087b32f7c44d71e6acba463d2c7dda798c3d66\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:07.889605 containerd[1463]: time="2025-09-10T00:41:07.889565917Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:71533e5a960e2955a54164905e92dac516ec874a23e0bf31304db82650101a4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:07.891012 containerd[1463]: time="2025-09-10T00:41:07.890937890Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.4\" with image id \"sha256:ab4ad8a84c3c69c18494ef32fa087b32f7c44d71e6acba463d2c7dda798c3d66\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:71533e5a960e2955a54164905e92dac516ec874a23e0bf31304db82650101a4a\", size \"21782824\" in 2.582256763s" Sep 10 00:41:07.891012 containerd[1463]: time="2025-09-10T00:41:07.890988154Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.4\" returns image reference \"sha256:ab4ad8a84c3c69c18494ef32fa087b32f7c44d71e6acba463d2c7dda798c3d66\"" Sep 10 00:41:07.891716 containerd[1463]: time="2025-09-10T00:41:07.891663991Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.4\"" Sep 10 00:41:09.362218 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2638172118.mount: Deactivated successfully. Sep 10 00:41:10.578520 containerd[1463]: time="2025-09-10T00:41:10.578406833Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:10.581460 containerd[1463]: time="2025-09-10T00:41:10.581337049Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.4: active requests=0, bytes read=31899626" Sep 10 00:41:10.585653 containerd[1463]: time="2025-09-10T00:41:10.585427349Z" level=info msg="ImageCreate event name:\"sha256:1b2ea5e018dbbbd2efb8e5c540a6d3c463d77f250d3904429402ee057f09c64e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:10.588871 containerd[1463]: time="2025-09-10T00:41:10.588788322Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bb04e9247da3aaeb96406b4d530a79fc865695b6807353dd1a28871df0d7f837\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:10.590054 containerd[1463]: time="2025-09-10T00:41:10.589898074Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.4\" with image id \"sha256:1b2ea5e018dbbbd2efb8e5c540a6d3c463d77f250d3904429402ee057f09c64e\", repo tag \"registry.k8s.io/kube-proxy:v1.33.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:bb04e9247da3aaeb96406b4d530a79fc865695b6807353dd1a28871df0d7f837\", size \"31898645\" in 2.698178638s" Sep 10 00:41:10.590054 containerd[1463]: time="2025-09-10T00:41:10.589991879Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.4\" returns image reference \"sha256:1b2ea5e018dbbbd2efb8e5c540a6d3c463d77f250d3904429402ee057f09c64e\"" Sep 10 00:41:10.590813 containerd[1463]: time="2025-09-10T00:41:10.590678056Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 10 00:41:11.533098 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount255638861.mount: Deactivated successfully. Sep 10 00:41:11.905593 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 10 00:41:11.923657 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:12.109181 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:12.115176 (kubelet)[1919]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:41:12.396084 kubelet[1919]: E0910 00:41:12.395850 1919 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:41:12.401715 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:41:12.401985 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:41:15.165120 containerd[1463]: time="2025-09-10T00:41:15.164988020Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:15.166851 containerd[1463]: time="2025-09-10T00:41:15.166792825Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Sep 10 00:41:15.168702 containerd[1463]: time="2025-09-10T00:41:15.168615954Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:15.173581 containerd[1463]: time="2025-09-10T00:41:15.173476249Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:15.175096 containerd[1463]: time="2025-09-10T00:41:15.175047766Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 4.58432688s" Sep 10 00:41:15.175096 containerd[1463]: time="2025-09-10T00:41:15.175088923Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Sep 10 00:41:15.175796 containerd[1463]: time="2025-09-10T00:41:15.175709307Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 10 00:41:16.088911 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1803166864.mount: Deactivated successfully. Sep 10 00:41:16.098036 containerd[1463]: time="2025-09-10T00:41:16.097968412Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:16.098953 containerd[1463]: time="2025-09-10T00:41:16.098831420Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 10 00:41:16.108442 containerd[1463]: time="2025-09-10T00:41:16.108370459Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:16.111134 containerd[1463]: time="2025-09-10T00:41:16.111088687Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:16.111938 containerd[1463]: time="2025-09-10T00:41:16.111892865Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 936.095823ms" Sep 10 00:41:16.111938 containerd[1463]: time="2025-09-10T00:41:16.111938931Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 10 00:41:16.112547 containerd[1463]: time="2025-09-10T00:41:16.112523107Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 10 00:41:16.747033 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3016934742.mount: Deactivated successfully. Sep 10 00:41:19.288733 containerd[1463]: time="2025-09-10T00:41:19.288622447Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:19.289672 containerd[1463]: time="2025-09-10T00:41:19.289628624Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58377871" Sep 10 00:41:19.291869 containerd[1463]: time="2025-09-10T00:41:19.291794445Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:19.296203 containerd[1463]: time="2025-09-10T00:41:19.296131569Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:41:19.298001 containerd[1463]: time="2025-09-10T00:41:19.297915404Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 3.185359666s" Sep 10 00:41:19.298001 containerd[1463]: time="2025-09-10T00:41:19.297995154Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Sep 10 00:41:22.405214 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Sep 10 00:41:22.414532 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:22.616754 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:22.624753 (kubelet)[2057]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:41:22.684044 kubelet[2057]: E0910 00:41:22.683736 2057 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:41:22.689366 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:41:22.689629 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:41:23.967198 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:23.981603 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:24.010341 systemd[1]: Reloading requested from client PID 2072 ('systemctl') (unit session-7.scope)... Sep 10 00:41:24.010380 systemd[1]: Reloading... Sep 10 00:41:24.121281 zram_generator::config[2114]: No configuration found. Sep 10 00:41:26.601551 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:41:26.692890 systemd[1]: Reloading finished in 2682 ms. Sep 10 00:41:26.751230 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 10 00:41:26.751388 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 10 00:41:26.751752 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:26.754320 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:26.947721 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:26.952298 (kubelet)[2160]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 10 00:41:27.062394 kubelet[2160]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:41:27.062394 kubelet[2160]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 10 00:41:27.062394 kubelet[2160]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:41:27.062912 kubelet[2160]: I0910 00:41:27.062424 2160 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 10 00:41:27.782510 kubelet[2160]: I0910 00:41:27.782443 2160 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 10 00:41:27.782510 kubelet[2160]: I0910 00:41:27.782480 2160 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 10 00:41:27.787175 kubelet[2160]: I0910 00:41:27.786991 2160 server.go:956] "Client rotation is on, will bootstrap in background" Sep 10 00:41:27.821032 kubelet[2160]: I0910 00:41:27.820961 2160 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 10 00:41:27.835910 kubelet[2160]: E0910 00:41:27.835838 2160 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.97:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 10 00:41:27.840336 kubelet[2160]: E0910 00:41:27.840281 2160 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 10 00:41:27.840336 kubelet[2160]: I0910 00:41:27.840332 2160 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 10 00:41:27.848732 kubelet[2160]: I0910 00:41:27.848689 2160 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 10 00:41:27.849039 kubelet[2160]: I0910 00:41:27.848981 2160 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 10 00:41:27.849210 kubelet[2160]: I0910 00:41:27.849024 2160 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 10 00:41:27.849407 kubelet[2160]: I0910 00:41:27.849213 2160 topology_manager.go:138] "Creating topology manager with none policy" Sep 10 00:41:27.849407 kubelet[2160]: I0910 00:41:27.849225 2160 container_manager_linux.go:303] "Creating device plugin manager" Sep 10 00:41:27.849456 kubelet[2160]: I0910 00:41:27.849434 2160 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:41:27.888965 kubelet[2160]: I0910 00:41:27.888877 2160 kubelet.go:480] "Attempting to sync node with API server" Sep 10 00:41:27.888965 kubelet[2160]: I0910 00:41:27.888955 2160 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 10 00:41:27.889182 kubelet[2160]: I0910 00:41:27.888996 2160 kubelet.go:386] "Adding apiserver pod source" Sep 10 00:41:27.891544 kubelet[2160]: E0910 00:41:27.891481 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.97:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 10 00:41:27.902042 kubelet[2160]: I0910 00:41:27.901989 2160 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 10 00:41:27.905847 kubelet[2160]: E0910 00:41:27.905799 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.97:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 10 00:41:27.907998 kubelet[2160]: I0910 00:41:27.907960 2160 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 10 00:41:27.908603 kubelet[2160]: I0910 00:41:27.908572 2160 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 10 00:41:27.909749 kubelet[2160]: W0910 00:41:27.909719 2160 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 10 00:41:27.913419 kubelet[2160]: I0910 00:41:27.913384 2160 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 10 00:41:27.913488 kubelet[2160]: I0910 00:41:27.913467 2160 server.go:1289] "Started kubelet" Sep 10 00:41:27.916281 kubelet[2160]: I0910 00:41:27.913698 2160 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 10 00:41:27.916281 kubelet[2160]: I0910 00:41:27.914673 2160 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 10 00:41:27.916281 kubelet[2160]: I0910 00:41:27.914660 2160 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 10 00:41:27.916281 kubelet[2160]: I0910 00:41:27.915928 2160 server.go:317] "Adding debug handlers to kubelet server" Sep 10 00:41:27.916591 kubelet[2160]: I0910 00:41:27.916561 2160 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 10 00:41:27.918528 kubelet[2160]: I0910 00:41:27.918491 2160 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 10 00:41:27.919230 kubelet[2160]: E0910 00:41:27.917564 2160 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.97:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.97:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1863c5010d4d0e28 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-10 00:41:27.91341828 +0000 UTC m=+0.896392713,LastTimestamp:2025-09-10 00:41:27.91341828 +0000 UTC m=+0.896392713,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 10 00:41:27.920388 kubelet[2160]: E0910 00:41:27.919804 2160 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 10 00:41:27.920388 kubelet[2160]: E0910 00:41:27.919819 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:27.920388 kubelet[2160]: I0910 00:41:27.919856 2160 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 10 00:41:27.920388 kubelet[2160]: I0910 00:41:27.920014 2160 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 10 00:41:27.920388 kubelet[2160]: I0910 00:41:27.920143 2160 reconciler.go:26] "Reconciler: start to sync state" Sep 10 00:41:27.921177 kubelet[2160]: E0910 00:41:27.921143 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.97:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 10 00:41:27.921473 kubelet[2160]: E0910 00:41:27.921429 2160 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.97:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.97:6443: connect: connection refused" interval="200ms" Sep 10 00:41:27.921727 kubelet[2160]: I0910 00:41:27.921694 2160 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 10 00:41:27.922966 kubelet[2160]: I0910 00:41:27.922942 2160 factory.go:223] Registration of the containerd container factory successfully Sep 10 00:41:27.923142 kubelet[2160]: I0910 00:41:27.923098 2160 factory.go:223] Registration of the systemd container factory successfully Sep 10 00:41:27.943422 kubelet[2160]: I0910 00:41:27.943389 2160 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 10 00:41:27.943852 kubelet[2160]: I0910 00:41:27.943653 2160 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 10 00:41:27.943852 kubelet[2160]: I0910 00:41:27.943714 2160 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:41:27.944448 kubelet[2160]: I0910 00:41:27.944408 2160 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 10 00:41:27.946357 kubelet[2160]: I0910 00:41:27.945988 2160 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 10 00:41:27.946357 kubelet[2160]: I0910 00:41:27.946024 2160 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 10 00:41:27.946357 kubelet[2160]: I0910 00:41:27.946050 2160 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 10 00:41:27.946357 kubelet[2160]: I0910 00:41:27.946057 2160 kubelet.go:2436] "Starting kubelet main sync loop" Sep 10 00:41:27.946357 kubelet[2160]: E0910 00:41:27.946104 2160 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 10 00:41:27.948315 kubelet[2160]: E0910 00:41:27.946894 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.97:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 10 00:41:28.020704 kubelet[2160]: E0910 00:41:28.020626 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.047057 kubelet[2160]: E0910 00:41:28.046942 2160 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 10 00:41:28.121510 kubelet[2160]: E0910 00:41:28.121429 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.122074 kubelet[2160]: E0910 00:41:28.122005 2160 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.97:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.97:6443: connect: connection refused" interval="400ms" Sep 10 00:41:28.222549 kubelet[2160]: E0910 00:41:28.222454 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.247708 kubelet[2160]: E0910 00:41:28.247649 2160 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 10 00:41:28.323378 kubelet[2160]: E0910 00:41:28.323177 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.424329 kubelet[2160]: E0910 00:41:28.424230 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.523454 kubelet[2160]: E0910 00:41:28.523383 2160 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.97:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.97:6443: connect: connection refused" interval="800ms" Sep 10 00:41:28.524431 kubelet[2160]: E0910 00:41:28.524376 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.625097 kubelet[2160]: E0910 00:41:28.624916 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.648272 kubelet[2160]: E0910 00:41:28.648180 2160 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 10 00:41:28.706733 kubelet[2160]: I0910 00:41:28.706623 2160 policy_none.go:49] "None policy: Start" Sep 10 00:41:28.706733 kubelet[2160]: I0910 00:41:28.706675 2160 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 10 00:41:28.706733 kubelet[2160]: I0910 00:41:28.706697 2160 state_mem.go:35] "Initializing new in-memory state store" Sep 10 00:41:28.725649 kubelet[2160]: E0910 00:41:28.725598 2160 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:41:28.740779 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 10 00:41:28.760826 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 10 00:41:28.764547 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 10 00:41:28.772262 kubelet[2160]: E0910 00:41:28.772189 2160 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 10 00:41:28.772534 kubelet[2160]: I0910 00:41:28.772504 2160 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 10 00:41:28.772596 kubelet[2160]: I0910 00:41:28.772524 2160 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 10 00:41:28.773344 kubelet[2160]: I0910 00:41:28.773094 2160 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 10 00:41:28.773577 kubelet[2160]: E0910 00:41:28.773550 2160 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 10 00:41:28.773630 kubelet[2160]: E0910 00:41:28.773608 2160 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 10 00:41:28.875039 kubelet[2160]: I0910 00:41:28.874955 2160 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:41:28.875700 kubelet[2160]: E0910 00:41:28.875539 2160 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.97:6443/api/v1/nodes\": dial tcp 10.0.0.97:6443: connect: connection refused" node="localhost" Sep 10 00:41:28.924002 kubelet[2160]: E0910 00:41:28.923933 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.97:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 10 00:41:28.958275 kubelet[2160]: E0910 00:41:28.958175 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.97:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 10 00:41:29.029886 kubelet[2160]: E0910 00:41:29.029799 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.97:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 10 00:41:29.077677 kubelet[2160]: I0910 00:41:29.077630 2160 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:41:29.078012 kubelet[2160]: E0910 00:41:29.077971 2160 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.97:6443/api/v1/nodes\": dial tcp 10.0.0.97:6443: connect: connection refused" node="localhost" Sep 10 00:41:29.276382 kubelet[2160]: E0910 00:41:29.276296 2160 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.97:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 10 00:41:29.324823 kubelet[2160]: E0910 00:41:29.324750 2160 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.97:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.97:6443: connect: connection refused" interval="1.6s" Sep 10 00:41:29.479841 kubelet[2160]: I0910 00:41:29.479791 2160 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:41:29.480357 kubelet[2160]: E0910 00:41:29.480309 2160 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.97:6443/api/v1/nodes\": dial tcp 10.0.0.97:6443: connect: connection refused" node="localhost" Sep 10 00:41:29.531083 kubelet[2160]: I0910 00:41:29.530888 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7df7bc200be9e91df7ab6c44d9024ad3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"7df7bc200be9e91df7ab6c44d9024ad3\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:29.531083 kubelet[2160]: I0910 00:41:29.530967 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7df7bc200be9e91df7ab6c44d9024ad3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"7df7bc200be9e91df7ab6c44d9024ad3\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:29.531083 kubelet[2160]: I0910 00:41:29.530994 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7df7bc200be9e91df7ab6c44d9024ad3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"7df7bc200be9e91df7ab6c44d9024ad3\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:29.636067 systemd[1]: Created slice kubepods-burstable-pod7df7bc200be9e91df7ab6c44d9024ad3.slice - libcontainer container kubepods-burstable-pod7df7bc200be9e91df7ab6c44d9024ad3.slice. Sep 10 00:41:29.657995 kubelet[2160]: E0910 00:41:29.657935 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:29.658350 kubelet[2160]: E0910 00:41:29.658328 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:29.658938 containerd[1463]: time="2025-09-10T00:41:29.658897127Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:7df7bc200be9e91df7ab6c44d9024ad3,Namespace:kube-system,Attempt:0,}" Sep 10 00:41:29.660370 systemd[1]: Created slice kubepods-burstable-pod8de7187202bee21b84740a213836f615.slice - libcontainer container kubepods-burstable-pod8de7187202bee21b84740a213836f615.slice. Sep 10 00:41:29.670605 kubelet[2160]: E0910 00:41:29.670560 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:29.673425 systemd[1]: Created slice kubepods-burstable-podd75e6f6978d9f275ea19380916c9cccd.slice - libcontainer container kubepods-burstable-podd75e6f6978d9f275ea19380916c9cccd.slice. Sep 10 00:41:29.675177 kubelet[2160]: E0910 00:41:29.675135 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:29.731385 kubelet[2160]: I0910 00:41:29.731311 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:29.731385 kubelet[2160]: I0910 00:41:29.731380 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:29.731644 kubelet[2160]: I0910 00:41:29.731416 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d75e6f6978d9f275ea19380916c9cccd-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d75e6f6978d9f275ea19380916c9cccd\") " pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:29.731644 kubelet[2160]: I0910 00:41:29.731439 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:29.731644 kubelet[2160]: I0910 00:41:29.731457 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:29.731644 kubelet[2160]: I0910 00:41:29.731479 2160 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:29.934559 kubelet[2160]: E0910 00:41:29.934502 2160 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.97:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.97:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 10 00:41:29.972055 kubelet[2160]: E0910 00:41:29.972021 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:29.972419 containerd[1463]: time="2025-09-10T00:41:29.972386458Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8de7187202bee21b84740a213836f615,Namespace:kube-system,Attempt:0,}" Sep 10 00:41:29.975655 kubelet[2160]: E0910 00:41:29.975616 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:29.975982 containerd[1463]: time="2025-09-10T00:41:29.975949567Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d75e6f6978d9f275ea19380916c9cccd,Namespace:kube-system,Attempt:0,}" Sep 10 00:41:30.139973 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3292163713.mount: Deactivated successfully. Sep 10 00:41:30.146820 containerd[1463]: time="2025-09-10T00:41:30.146763617Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:41:30.147884 containerd[1463]: time="2025-09-10T00:41:30.147800149Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:41:30.148814 containerd[1463]: time="2025-09-10T00:41:30.148766448Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Sep 10 00:41:30.149669 containerd[1463]: time="2025-09-10T00:41:30.149623500Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 10 00:41:30.150594 containerd[1463]: time="2025-09-10T00:41:30.150548340Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:41:30.153415 containerd[1463]: time="2025-09-10T00:41:30.153378897Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:41:30.154398 containerd[1463]: time="2025-09-10T00:41:30.154362609Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 495.384526ms" Sep 10 00:41:30.158141 containerd[1463]: time="2025-09-10T00:41:30.158105513Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 182.10023ms" Sep 10 00:41:30.175114 containerd[1463]: time="2025-09-10T00:41:30.175066103Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:41:30.175976 containerd[1463]: time="2025-09-10T00:41:30.175925769Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 10 00:41:30.189623 containerd[1463]: time="2025-09-10T00:41:30.189492671Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 217.032832ms" Sep 10 00:41:30.283071 kubelet[2160]: I0910 00:41:30.282419 2160 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:41:30.283071 kubelet[2160]: E0910 00:41:30.282724 2160 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.97:6443/api/v1/nodes\": dial tcp 10.0.0.97:6443: connect: connection refused" node="localhost" Sep 10 00:41:30.289619 containerd[1463]: time="2025-09-10T00:41:30.289482416Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:41:30.289619 containerd[1463]: time="2025-09-10T00:41:30.289547800Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:41:30.289619 containerd[1463]: time="2025-09-10T00:41:30.289563200Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:30.289619 containerd[1463]: time="2025-09-10T00:41:30.289535327Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:41:30.289619 containerd[1463]: time="2025-09-10T00:41:30.289603978Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:41:30.289805 containerd[1463]: time="2025-09-10T00:41:30.289624456Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:30.289805 containerd[1463]: time="2025-09-10T00:41:30.289652951Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:30.289805 containerd[1463]: time="2025-09-10T00:41:30.289731099Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:30.290549 containerd[1463]: time="2025-09-10T00:41:30.290417436Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:41:30.290549 containerd[1463]: time="2025-09-10T00:41:30.290465998Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:41:30.290549 containerd[1463]: time="2025-09-10T00:41:30.290478381Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:30.292431 containerd[1463]: time="2025-09-10T00:41:30.292327291Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:30.314432 systemd[1]: Started cri-containerd-854fd75414c5f1d4c16fa8eb8521a2eb30f7f16fd34f2b993b6c039d87e97188.scope - libcontainer container 854fd75414c5f1d4c16fa8eb8521a2eb30f7f16fd34f2b993b6c039d87e97188. Sep 10 00:41:30.320281 systemd[1]: Started cri-containerd-2693071a6e821f6aedfe43977cfcf6e142938a449636027c566f19ca7344434c.scope - libcontainer container 2693071a6e821f6aedfe43977cfcf6e142938a449636027c566f19ca7344434c. Sep 10 00:41:30.323299 systemd[1]: Started cri-containerd-bfeec307d2165a46d4f35fd8b95615dd6c1736aafbd2c15793647c436cc0ec9b.scope - libcontainer container bfeec307d2165a46d4f35fd8b95615dd6c1736aafbd2c15793647c436cc0ec9b. Sep 10 00:41:30.361594 containerd[1463]: time="2025-09-10T00:41:30.361535211Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:7df7bc200be9e91df7ab6c44d9024ad3,Namespace:kube-system,Attempt:0,} returns sandbox id \"854fd75414c5f1d4c16fa8eb8521a2eb30f7f16fd34f2b993b6c039d87e97188\"" Sep 10 00:41:30.364465 kubelet[2160]: E0910 00:41:30.364413 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:30.370559 containerd[1463]: time="2025-09-10T00:41:30.370510303Z" level=info msg="CreateContainer within sandbox \"854fd75414c5f1d4c16fa8eb8521a2eb30f7f16fd34f2b993b6c039d87e97188\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 10 00:41:30.372308 containerd[1463]: time="2025-09-10T00:41:30.371935014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8de7187202bee21b84740a213836f615,Namespace:kube-system,Attempt:0,} returns sandbox id \"bfeec307d2165a46d4f35fd8b95615dd6c1736aafbd2c15793647c436cc0ec9b\"" Sep 10 00:41:30.373291 kubelet[2160]: E0910 00:41:30.373259 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:30.381715 containerd[1463]: time="2025-09-10T00:41:30.381576424Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d75e6f6978d9f275ea19380916c9cccd,Namespace:kube-system,Attempt:0,} returns sandbox id \"2693071a6e821f6aedfe43977cfcf6e142938a449636027c566f19ca7344434c\"" Sep 10 00:41:30.381715 containerd[1463]: time="2025-09-10T00:41:30.381632030Z" level=info msg="CreateContainer within sandbox \"bfeec307d2165a46d4f35fd8b95615dd6c1736aafbd2c15793647c436cc0ec9b\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 10 00:41:30.382125 kubelet[2160]: E0910 00:41:30.382095 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:30.386813 containerd[1463]: time="2025-09-10T00:41:30.386644991Z" level=info msg="CreateContainer within sandbox \"2693071a6e821f6aedfe43977cfcf6e142938a449636027c566f19ca7344434c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 10 00:41:30.389975 containerd[1463]: time="2025-09-10T00:41:30.389947887Z" level=info msg="CreateContainer within sandbox \"854fd75414c5f1d4c16fa8eb8521a2eb30f7f16fd34f2b993b6c039d87e97188\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"1ebf84b87671b15ce8682cc34d52adb86de7e3259c72358358084b59e86d5ed4\"" Sep 10 00:41:30.390567 containerd[1463]: time="2025-09-10T00:41:30.390533892Z" level=info msg="StartContainer for \"1ebf84b87671b15ce8682cc34d52adb86de7e3259c72358358084b59e86d5ed4\"" Sep 10 00:41:30.411981 containerd[1463]: time="2025-09-10T00:41:30.411891211Z" level=info msg="CreateContainer within sandbox \"2693071a6e821f6aedfe43977cfcf6e142938a449636027c566f19ca7344434c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"52bfd581ddf6fbef13398e9254d39293623756d4921c796328f1dfc937414017\"" Sep 10 00:41:30.413825 containerd[1463]: time="2025-09-10T00:41:30.412492255Z" level=info msg="StartContainer for \"52bfd581ddf6fbef13398e9254d39293623756d4921c796328f1dfc937414017\"" Sep 10 00:41:30.414463 containerd[1463]: time="2025-09-10T00:41:30.414416828Z" level=info msg="CreateContainer within sandbox \"bfeec307d2165a46d4f35fd8b95615dd6c1736aafbd2c15793647c436cc0ec9b\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"0e90ae2f4c2ab4d9ee5f437c2d14c3b1e45fb370a06ea537f8165efa67ac3321\"" Sep 10 00:41:30.415077 containerd[1463]: time="2025-09-10T00:41:30.415021098Z" level=info msg="StartContainer for \"0e90ae2f4c2ab4d9ee5f437c2d14c3b1e45fb370a06ea537f8165efa67ac3321\"" Sep 10 00:41:30.420686 systemd[1]: Started cri-containerd-1ebf84b87671b15ce8682cc34d52adb86de7e3259c72358358084b59e86d5ed4.scope - libcontainer container 1ebf84b87671b15ce8682cc34d52adb86de7e3259c72358358084b59e86d5ed4. Sep 10 00:41:30.446777 systemd[1]: Started cri-containerd-0e90ae2f4c2ab4d9ee5f437c2d14c3b1e45fb370a06ea537f8165efa67ac3321.scope - libcontainer container 0e90ae2f4c2ab4d9ee5f437c2d14c3b1e45fb370a06ea537f8165efa67ac3321. Sep 10 00:41:30.455412 systemd[1]: Started cri-containerd-52bfd581ddf6fbef13398e9254d39293623756d4921c796328f1dfc937414017.scope - libcontainer container 52bfd581ddf6fbef13398e9254d39293623756d4921c796328f1dfc937414017. Sep 10 00:41:30.477714 containerd[1463]: time="2025-09-10T00:41:30.477669354Z" level=info msg="StartContainer for \"1ebf84b87671b15ce8682cc34d52adb86de7e3259c72358358084b59e86d5ed4\" returns successfully" Sep 10 00:41:30.495607 containerd[1463]: time="2025-09-10T00:41:30.495548431Z" level=info msg="StartContainer for \"0e90ae2f4c2ab4d9ee5f437c2d14c3b1e45fb370a06ea537f8165efa67ac3321\" returns successfully" Sep 10 00:41:30.518495 containerd[1463]: time="2025-09-10T00:41:30.518428488Z" level=info msg="StartContainer for \"52bfd581ddf6fbef13398e9254d39293623756d4921c796328f1dfc937414017\" returns successfully" Sep 10 00:41:30.956417 kubelet[2160]: E0910 00:41:30.956139 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:30.956417 kubelet[2160]: E0910 00:41:30.956293 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:30.958359 kubelet[2160]: E0910 00:41:30.958332 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:30.958437 kubelet[2160]: E0910 00:41:30.958413 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:30.960627 kubelet[2160]: E0910 00:41:30.960598 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:30.960742 kubelet[2160]: E0910 00:41:30.960716 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:31.887684 kubelet[2160]: I0910 00:41:31.885158 2160 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:41:31.967381 kubelet[2160]: E0910 00:41:31.967327 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:31.967644 kubelet[2160]: E0910 00:41:31.967492 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:31.968170 kubelet[2160]: E0910 00:41:31.968142 2160 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:41:31.968283 kubelet[2160]: E0910 00:41:31.968240 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:32.089034 kubelet[2160]: E0910 00:41:32.088983 2160 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 10 00:41:32.167999 kubelet[2160]: I0910 00:41:32.167927 2160 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 10 00:41:32.168183 kubelet[2160]: E0910 00:41:32.168023 2160 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 10 00:41:32.220335 kubelet[2160]: I0910 00:41:32.220234 2160 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:32.378746 kubelet[2160]: E0910 00:41:32.378484 2160 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:32.378746 kubelet[2160]: I0910 00:41:32.378529 2160 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:32.383005 kubelet[2160]: E0910 00:41:32.380699 2160 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:32.383005 kubelet[2160]: I0910 00:41:32.380721 2160 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:32.383005 kubelet[2160]: E0910 00:41:32.382914 2160 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:32.908285 kubelet[2160]: I0910 00:41:32.908217 2160 apiserver.go:52] "Watching apiserver" Sep 10 00:41:32.920546 kubelet[2160]: I0910 00:41:32.920480 2160 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 10 00:41:32.966711 kubelet[2160]: I0910 00:41:32.966654 2160 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:32.969704 kubelet[2160]: E0910 00:41:32.969633 2160 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:32.969965 kubelet[2160]: E0910 00:41:32.969904 2160 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:33.152346 update_engine[1451]: I20250910 00:41:33.152161 1451 update_attempter.cc:509] Updating boot flags... Sep 10 00:41:33.402715 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (2460) Sep 10 00:41:33.446286 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (2462) Sep 10 00:41:36.918081 systemd[1]: Reloading requested from client PID 2468 ('systemctl') (unit session-7.scope)... Sep 10 00:41:36.918106 systemd[1]: Reloading... Sep 10 00:41:37.014293 zram_generator::config[2507]: No configuration found. Sep 10 00:41:37.160023 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:41:37.279088 systemd[1]: Reloading finished in 360 ms. Sep 10 00:41:37.338502 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:37.352916 systemd[1]: kubelet.service: Deactivated successfully. Sep 10 00:41:37.353331 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:37.353394 systemd[1]: kubelet.service: Consumed 1.652s CPU time, 132.6M memory peak, 0B memory swap peak. Sep 10 00:41:37.364763 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:41:37.560195 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:41:37.573918 (kubelet)[2552]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 10 00:41:37.627723 kubelet[2552]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:41:37.627723 kubelet[2552]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 10 00:41:37.627723 kubelet[2552]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:41:37.628315 kubelet[2552]: I0910 00:41:37.627748 2552 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 10 00:41:37.634838 kubelet[2552]: I0910 00:41:37.634780 2552 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 10 00:41:37.634838 kubelet[2552]: I0910 00:41:37.634813 2552 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 10 00:41:37.635118 kubelet[2552]: I0910 00:41:37.635091 2552 server.go:956] "Client rotation is on, will bootstrap in background" Sep 10 00:41:37.636630 kubelet[2552]: I0910 00:41:37.636600 2552 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 10 00:41:37.639121 kubelet[2552]: I0910 00:41:37.639083 2552 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 10 00:41:37.644068 kubelet[2552]: E0910 00:41:37.644034 2552 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 10 00:41:37.644068 kubelet[2552]: I0910 00:41:37.644069 2552 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 10 00:41:37.649443 kubelet[2552]: I0910 00:41:37.649407 2552 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 10 00:41:37.649675 kubelet[2552]: I0910 00:41:37.649631 2552 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 10 00:41:37.649875 kubelet[2552]: I0910 00:41:37.649658 2552 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 10 00:41:37.649984 kubelet[2552]: I0910 00:41:37.649878 2552 topology_manager.go:138] "Creating topology manager with none policy" Sep 10 00:41:37.649984 kubelet[2552]: I0910 00:41:37.649889 2552 container_manager_linux.go:303] "Creating device plugin manager" Sep 10 00:41:37.650792 kubelet[2552]: I0910 00:41:37.650762 2552 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:41:37.651056 kubelet[2552]: I0910 00:41:37.650996 2552 kubelet.go:480] "Attempting to sync node with API server" Sep 10 00:41:37.651056 kubelet[2552]: I0910 00:41:37.651038 2552 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 10 00:41:37.651149 kubelet[2552]: I0910 00:41:37.651068 2552 kubelet.go:386] "Adding apiserver pod source" Sep 10 00:41:37.651149 kubelet[2552]: I0910 00:41:37.651088 2552 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 10 00:41:37.656571 kubelet[2552]: I0910 00:41:37.656537 2552 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 10 00:41:37.657685 kubelet[2552]: I0910 00:41:37.657646 2552 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 10 00:41:37.661406 kubelet[2552]: I0910 00:41:37.661368 2552 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 10 00:41:37.661496 kubelet[2552]: I0910 00:41:37.661443 2552 server.go:1289] "Started kubelet" Sep 10 00:41:37.662123 kubelet[2552]: I0910 00:41:37.662051 2552 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 10 00:41:37.662531 kubelet[2552]: I0910 00:41:37.662502 2552 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 10 00:41:37.662583 kubelet[2552]: I0910 00:41:37.662556 2552 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 10 00:41:37.665273 kubelet[2552]: I0910 00:41:37.663472 2552 server.go:317] "Adding debug handlers to kubelet server" Sep 10 00:41:37.672996 kubelet[2552]: I0910 00:41:37.672965 2552 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 10 00:41:37.674665 kubelet[2552]: I0910 00:41:37.673853 2552 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 10 00:41:37.678925 kubelet[2552]: I0910 00:41:37.678625 2552 factory.go:223] Registration of the containerd container factory successfully Sep 10 00:41:37.678925 kubelet[2552]: I0910 00:41:37.678645 2552 factory.go:223] Registration of the systemd container factory successfully Sep 10 00:41:37.678925 kubelet[2552]: E0910 00:41:37.678648 2552 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 10 00:41:37.679370 kubelet[2552]: I0910 00:41:37.679343 2552 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 10 00:41:37.679579 kubelet[2552]: I0910 00:41:37.679539 2552 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 10 00:41:37.679740 kubelet[2552]: I0910 00:41:37.679720 2552 reconciler.go:26] "Reconciler: start to sync state" Sep 10 00:41:37.680542 kubelet[2552]: I0910 00:41:37.680376 2552 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 10 00:41:37.696487 kubelet[2552]: I0910 00:41:37.696443 2552 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 10 00:41:37.699760 kubelet[2552]: I0910 00:41:37.699204 2552 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 10 00:41:37.699760 kubelet[2552]: I0910 00:41:37.699221 2552 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 10 00:41:37.699760 kubelet[2552]: I0910 00:41:37.699256 2552 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 10 00:41:37.699760 kubelet[2552]: I0910 00:41:37.699264 2552 kubelet.go:2436] "Starting kubelet main sync loop" Sep 10 00:41:37.699760 kubelet[2552]: E0910 00:41:37.699317 2552 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 10 00:41:37.720655 kubelet[2552]: I0910 00:41:37.720614 2552 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 10 00:41:37.720655 kubelet[2552]: I0910 00:41:37.720636 2552 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 10 00:41:37.720655 kubelet[2552]: I0910 00:41:37.720655 2552 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:41:37.720978 kubelet[2552]: I0910 00:41:37.720807 2552 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 10 00:41:37.720978 kubelet[2552]: I0910 00:41:37.720833 2552 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 10 00:41:37.720978 kubelet[2552]: I0910 00:41:37.720852 2552 policy_none.go:49] "None policy: Start" Sep 10 00:41:37.720978 kubelet[2552]: I0910 00:41:37.720862 2552 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 10 00:41:37.720978 kubelet[2552]: I0910 00:41:37.720874 2552 state_mem.go:35] "Initializing new in-memory state store" Sep 10 00:41:37.720978 kubelet[2552]: I0910 00:41:37.720968 2552 state_mem.go:75] "Updated machine memory state" Sep 10 00:41:37.726649 kubelet[2552]: E0910 00:41:37.726602 2552 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 10 00:41:37.726978 kubelet[2552]: I0910 00:41:37.726832 2552 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 10 00:41:37.726978 kubelet[2552]: I0910 00:41:37.726851 2552 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 10 00:41:37.727189 kubelet[2552]: I0910 00:41:37.727168 2552 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 10 00:41:37.729520 kubelet[2552]: E0910 00:41:37.729489 2552 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 10 00:41:37.801498 kubelet[2552]: I0910 00:41:37.801434 2552 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:37.801498 kubelet[2552]: I0910 00:41:37.801481 2552 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:37.801764 kubelet[2552]: I0910 00:41:37.801572 2552 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:37.835432 kubelet[2552]: I0910 00:41:37.835291 2552 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:41:37.847750 kubelet[2552]: I0910 00:41:37.846784 2552 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 10 00:41:37.847750 kubelet[2552]: I0910 00:41:37.846935 2552 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 10 00:41:37.928181 sudo[2593]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 10 00:41:37.928763 sudo[2593]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 10 00:41:37.980967 kubelet[2552]: I0910 00:41:37.980899 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7df7bc200be9e91df7ab6c44d9024ad3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"7df7bc200be9e91df7ab6c44d9024ad3\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:37.980967 kubelet[2552]: I0910 00:41:37.980961 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:37.981150 kubelet[2552]: I0910 00:41:37.980994 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:37.981150 kubelet[2552]: I0910 00:41:37.981015 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:37.981150 kubelet[2552]: I0910 00:41:37.981036 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7df7bc200be9e91df7ab6c44d9024ad3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"7df7bc200be9e91df7ab6c44d9024ad3\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:37.981150 kubelet[2552]: I0910 00:41:37.981054 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7df7bc200be9e91df7ab6c44d9024ad3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"7df7bc200be9e91df7ab6c44d9024ad3\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:37.981150 kubelet[2552]: I0910 00:41:37.981081 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:37.981284 kubelet[2552]: I0910 00:41:37.981100 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:41:37.981284 kubelet[2552]: I0910 00:41:37.981122 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d75e6f6978d9f275ea19380916c9cccd-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d75e6f6978d9f275ea19380916c9cccd\") " pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:38.108479 kubelet[2552]: E0910 00:41:38.108237 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:38.111199 kubelet[2552]: E0910 00:41:38.111156 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:38.111410 kubelet[2552]: E0910 00:41:38.111339 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:38.456053 sudo[2593]: pam_unix(sudo:session): session closed for user root Sep 10 00:41:38.652400 kubelet[2552]: I0910 00:41:38.652333 2552 apiserver.go:52] "Watching apiserver" Sep 10 00:41:38.679709 kubelet[2552]: I0910 00:41:38.679643 2552 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 10 00:41:38.711550 kubelet[2552]: E0910 00:41:38.710729 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:38.711550 kubelet[2552]: I0910 00:41:38.710847 2552 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:38.711550 kubelet[2552]: I0910 00:41:38.711152 2552 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:38.721821 kubelet[2552]: E0910 00:41:38.720944 2552 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 10 00:41:38.721821 kubelet[2552]: E0910 00:41:38.721169 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:38.721821 kubelet[2552]: E0910 00:41:38.721668 2552 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 10 00:41:38.721821 kubelet[2552]: E0910 00:41:38.721755 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:38.734313 kubelet[2552]: I0910 00:41:38.734212 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.734194713 podStartE2EDuration="1.734194713s" podCreationTimestamp="2025-09-10 00:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:41:38.731193665 +0000 UTC m=+1.150240986" watchObservedRunningTime="2025-09-10 00:41:38.734194713 +0000 UTC m=+1.153242034" Sep 10 00:41:38.750698 kubelet[2552]: I0910 00:41:38.750619 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.7506002010000001 podStartE2EDuration="1.750600201s" podCreationTimestamp="2025-09-10 00:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:41:38.739078923 +0000 UTC m=+1.158126244" watchObservedRunningTime="2025-09-10 00:41:38.750600201 +0000 UTC m=+1.169647522" Sep 10 00:41:38.759912 kubelet[2552]: I0910 00:41:38.759834 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.759816699 podStartE2EDuration="1.759816699s" podCreationTimestamp="2025-09-10 00:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:41:38.750839224 +0000 UTC m=+1.169886555" watchObservedRunningTime="2025-09-10 00:41:38.759816699 +0000 UTC m=+1.178864020" Sep 10 00:41:39.711339 kubelet[2552]: E0910 00:41:39.711303 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:39.711752 kubelet[2552]: E0910 00:41:39.711377 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:40.437360 sudo[1647]: pam_unix(sudo:session): session closed for user root Sep 10 00:41:40.460043 sshd[1644]: pam_unix(sshd:session): session closed for user core Sep 10 00:41:40.464350 systemd[1]: sshd@6-10.0.0.97:22-10.0.0.1:42634.service: Deactivated successfully. Sep 10 00:41:40.466591 systemd[1]: session-7.scope: Deactivated successfully. Sep 10 00:41:40.466847 systemd[1]: session-7.scope: Consumed 8.916s CPU time, 162.8M memory peak, 0B memory swap peak. Sep 10 00:41:40.467347 systemd-logind[1448]: Session 7 logged out. Waiting for processes to exit. Sep 10 00:41:40.468725 systemd-logind[1448]: Removed session 7. Sep 10 00:41:41.858236 kubelet[2552]: I0910 00:41:41.858195 2552 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 10 00:41:41.858821 kubelet[2552]: I0910 00:41:41.858668 2552 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 10 00:41:41.858875 containerd[1463]: time="2025-09-10T00:41:41.858497486Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 10 00:41:43.040543 systemd[1]: Created slice kubepods-besteffort-podf060eaf7_8cca_4250_8ad1_4f2a4237ae9a.slice - libcontainer container kubepods-besteffort-podf060eaf7_8cca_4250_8ad1_4f2a4237ae9a.slice. Sep 10 00:41:43.056871 systemd[1]: Created slice kubepods-burstable-podcd0204fb_1928_4cc6_b8b5_673ee11acce8.slice - libcontainer container kubepods-burstable-podcd0204fb_1928_4cc6_b8b5_673ee11acce8.slice. Sep 10 00:41:43.076098 systemd[1]: Created slice kubepods-besteffort-pod21e19d89_fbf5_4b8f_b135_124fe927b8a1.slice - libcontainer container kubepods-besteffort-pod21e19d89_fbf5_4b8f_b135_124fe927b8a1.slice. Sep 10 00:41:43.115411 kubelet[2552]: I0910 00:41:43.115341 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwrr\" (UniqueName: \"kubernetes.io/projected/f060eaf7-8cca-4250-8ad1-4f2a4237ae9a-kube-api-access-jgwrr\") pod \"kube-proxy-wf8n8\" (UID: \"f060eaf7-8cca-4250-8ad1-4f2a4237ae9a\") " pod="kube-system/kube-proxy-wf8n8" Sep 10 00:41:43.115411 kubelet[2552]: I0910 00:41:43.115390 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-etc-cni-netd\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.115979 kubelet[2552]: I0910 00:41:43.115412 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-xtables-lock\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.115979 kubelet[2552]: I0910 00:41:43.115473 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f060eaf7-8cca-4250-8ad1-4f2a4237ae9a-xtables-lock\") pod \"kube-proxy-wf8n8\" (UID: \"f060eaf7-8cca-4250-8ad1-4f2a4237ae9a\") " pod="kube-system/kube-proxy-wf8n8" Sep 10 00:41:43.115979 kubelet[2552]: I0910 00:41:43.115509 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-kernel\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.115979 kubelet[2552]: I0910 00:41:43.115615 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f060eaf7-8cca-4250-8ad1-4f2a4237ae9a-lib-modules\") pod \"kube-proxy-wf8n8\" (UID: \"f060eaf7-8cca-4250-8ad1-4f2a4237ae9a\") " pod="kube-system/kube-proxy-wf8n8" Sep 10 00:41:43.115979 kubelet[2552]: I0910 00:41:43.115667 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-lib-modules\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.115979 kubelet[2552]: I0910 00:41:43.115701 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-net\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116224 kubelet[2552]: I0910 00:41:43.115732 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/21e19d89-fbf5-4b8f-b135-124fe927b8a1-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-g5m7h\" (UID: \"21e19d89-fbf5-4b8f-b135-124fe927b8a1\") " pod="kube-system/cilium-operator-6c4d7847fc-g5m7h" Sep 10 00:41:43.116224 kubelet[2552]: I0910 00:41:43.115772 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f060eaf7-8cca-4250-8ad1-4f2a4237ae9a-kube-proxy\") pod \"kube-proxy-wf8n8\" (UID: \"f060eaf7-8cca-4250-8ad1-4f2a4237ae9a\") " pod="kube-system/kube-proxy-wf8n8" Sep 10 00:41:43.116224 kubelet[2552]: I0910 00:41:43.115791 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-bpf-maps\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116224 kubelet[2552]: I0910 00:41:43.115812 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ckjl\" (UniqueName: \"kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-kube-api-access-2ckjl\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116224 kubelet[2552]: I0910 00:41:43.115835 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hostproc\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116430 kubelet[2552]: I0910 00:41:43.115855 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/cd0204fb-1928-4cc6-b8b5-673ee11acce8-clustermesh-secrets\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116430 kubelet[2552]: I0910 00:41:43.115875 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-config-path\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116430 kubelet[2552]: I0910 00:41:43.115899 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cni-path\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116430 kubelet[2552]: I0910 00:41:43.115923 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s92fg\" (UniqueName: \"kubernetes.io/projected/21e19d89-fbf5-4b8f-b135-124fe927b8a1-kube-api-access-s92fg\") pod \"cilium-operator-6c4d7847fc-g5m7h\" (UID: \"21e19d89-fbf5-4b8f-b135-124fe927b8a1\") " pod="kube-system/cilium-operator-6c4d7847fc-g5m7h" Sep 10 00:41:43.116430 kubelet[2552]: I0910 00:41:43.115942 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-run\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116581 kubelet[2552]: I0910 00:41:43.115962 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-cgroup\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.116581 kubelet[2552]: I0910 00:41:43.115982 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hubble-tls\") pod \"cilium-cwph7\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " pod="kube-system/cilium-cwph7" Sep 10 00:41:43.649714 kubelet[2552]: E0910 00:41:43.649640 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:43.650556 containerd[1463]: time="2025-09-10T00:41:43.650514609Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wf8n8,Uid:f060eaf7-8cca-4250-8ad1-4f2a4237ae9a,Namespace:kube-system,Attempt:0,}" Sep 10 00:41:43.664948 kubelet[2552]: E0910 00:41:43.664869 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:43.665749 containerd[1463]: time="2025-09-10T00:41:43.665658851Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-cwph7,Uid:cd0204fb-1928-4cc6-b8b5-673ee11acce8,Namespace:kube-system,Attempt:0,}" Sep 10 00:41:43.683572 kubelet[2552]: E0910 00:41:43.683499 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:43.684213 containerd[1463]: time="2025-09-10T00:41:43.684158909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-g5m7h,Uid:21e19d89-fbf5-4b8f-b135-124fe927b8a1,Namespace:kube-system,Attempt:0,}" Sep 10 00:41:44.459283 containerd[1463]: time="2025-09-10T00:41:44.457470600Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:41:44.459283 containerd[1463]: time="2025-09-10T00:41:44.457556222Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:41:44.459283 containerd[1463]: time="2025-09-10T00:41:44.457568747Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:44.459283 containerd[1463]: time="2025-09-10T00:41:44.458032852Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:44.466947 containerd[1463]: time="2025-09-10T00:41:44.466324505Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:41:44.466947 containerd[1463]: time="2025-09-10T00:41:44.466399065Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:41:44.466947 containerd[1463]: time="2025-09-10T00:41:44.466418242Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:44.466947 containerd[1463]: time="2025-09-10T00:41:44.466518181Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:44.480004 containerd[1463]: time="2025-09-10T00:41:44.479753622Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:41:44.480132 containerd[1463]: time="2025-09-10T00:41:44.480093742Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:41:44.480180 containerd[1463]: time="2025-09-10T00:41:44.480151822Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:44.480586 containerd[1463]: time="2025-09-10T00:41:44.480473790Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:41:44.489090 systemd[1]: Started cri-containerd-c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872.scope - libcontainer container c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872. Sep 10 00:41:44.499550 systemd[1]: Started cri-containerd-adbf8b98e595b49a28fdda649c03d21210039d8c3fb38ac01f22a14fdc62c13d.scope - libcontainer container adbf8b98e595b49a28fdda649c03d21210039d8c3fb38ac01f22a14fdc62c13d. Sep 10 00:41:44.504927 systemd[1]: Started cri-containerd-08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e.scope - libcontainer container 08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e. Sep 10 00:41:44.538662 containerd[1463]: time="2025-09-10T00:41:44.538420566Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-cwph7,Uid:cd0204fb-1928-4cc6-b8b5-673ee11acce8,Namespace:kube-system,Attempt:0,} returns sandbox id \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\"" Sep 10 00:41:44.539778 kubelet[2552]: E0910 00:41:44.539448 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:44.544271 containerd[1463]: time="2025-09-10T00:41:44.541934345Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wf8n8,Uid:f060eaf7-8cca-4250-8ad1-4f2a4237ae9a,Namespace:kube-system,Attempt:0,} returns sandbox id \"adbf8b98e595b49a28fdda649c03d21210039d8c3fb38ac01f22a14fdc62c13d\"" Sep 10 00:41:44.544271 containerd[1463]: time="2025-09-10T00:41:44.543415619Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 10 00:41:44.544361 kubelet[2552]: E0910 00:41:44.543796 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:44.556878 containerd[1463]: time="2025-09-10T00:41:44.556818526Z" level=info msg="CreateContainer within sandbox \"adbf8b98e595b49a28fdda649c03d21210039d8c3fb38ac01f22a14fdc62c13d\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 10 00:41:44.569882 containerd[1463]: time="2025-09-10T00:41:44.569828451Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-g5m7h,Uid:21e19d89-fbf5-4b8f-b135-124fe927b8a1,Namespace:kube-system,Attempt:0,} returns sandbox id \"08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e\"" Sep 10 00:41:44.570609 kubelet[2552]: E0910 00:41:44.570575 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:44.583829 containerd[1463]: time="2025-09-10T00:41:44.583738495Z" level=info msg="CreateContainer within sandbox \"adbf8b98e595b49a28fdda649c03d21210039d8c3fb38ac01f22a14fdc62c13d\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"e6ab433fd481e1da67f08eb98918fb304a08c08a12e25fb2ab07285ad3b52fb6\"" Sep 10 00:41:44.584600 containerd[1463]: time="2025-09-10T00:41:44.584562139Z" level=info msg="StartContainer for \"e6ab433fd481e1da67f08eb98918fb304a08c08a12e25fb2ab07285ad3b52fb6\"" Sep 10 00:41:44.624585 systemd[1]: Started cri-containerd-e6ab433fd481e1da67f08eb98918fb304a08c08a12e25fb2ab07285ad3b52fb6.scope - libcontainer container e6ab433fd481e1da67f08eb98918fb304a08c08a12e25fb2ab07285ad3b52fb6. Sep 10 00:41:44.664687 containerd[1463]: time="2025-09-10T00:41:44.664618849Z" level=info msg="StartContainer for \"e6ab433fd481e1da67f08eb98918fb304a08c08a12e25fb2ab07285ad3b52fb6\" returns successfully" Sep 10 00:41:44.724063 kubelet[2552]: E0910 00:41:44.723902 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:46.942553 kubelet[2552]: E0910 00:41:46.942510 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:46.956506 kubelet[2552]: I0910 00:41:46.956011 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-wf8n8" podStartSLOduration=4.955992903 podStartE2EDuration="4.955992903s" podCreationTimestamp="2025-09-10 00:41:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:41:44.736229313 +0000 UTC m=+7.155276634" watchObservedRunningTime="2025-09-10 00:41:46.955992903 +0000 UTC m=+9.375040214" Sep 10 00:41:47.389478 kubelet[2552]: E0910 00:41:47.389305 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:47.729532 kubelet[2552]: E0910 00:41:47.729491 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:47.729752 kubelet[2552]: E0910 00:41:47.729721 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:48.134729 kubelet[2552]: E0910 00:41:48.132672 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:41:48.730948 kubelet[2552]: E0910 00:41:48.730893 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:00.832643 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1016583076.mount: Deactivated successfully. Sep 10 00:42:03.848615 containerd[1463]: time="2025-09-10T00:42:03.848518356Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:42:03.849453 containerd[1463]: time="2025-09-10T00:42:03.849382820Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Sep 10 00:42:03.850724 containerd[1463]: time="2025-09-10T00:42:03.850654850Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:42:03.852518 containerd[1463]: time="2025-09-10T00:42:03.852493063Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 19.309047056s" Sep 10 00:42:03.852572 containerd[1463]: time="2025-09-10T00:42:03.852530143Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 10 00:42:03.857117 containerd[1463]: time="2025-09-10T00:42:03.857070242Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 10 00:42:03.945317 containerd[1463]: time="2025-09-10T00:42:03.945219936Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 10 00:42:03.965506 containerd[1463]: time="2025-09-10T00:42:03.965431888Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3\"" Sep 10 00:42:03.966235 containerd[1463]: time="2025-09-10T00:42:03.966189351Z" level=info msg="StartContainer for \"0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3\"" Sep 10 00:42:04.004722 systemd[1]: run-containerd-runc-k8s.io-0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3-runc.RdPMos.mount: Deactivated successfully. Sep 10 00:42:04.017555 systemd[1]: Started cri-containerd-0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3.scope - libcontainer container 0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3. Sep 10 00:42:04.236809 containerd[1463]: time="2025-09-10T00:42:04.236732324Z" level=info msg="StartContainer for \"0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3\" returns successfully" Sep 10 00:42:04.249127 systemd[1]: cri-containerd-0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3.scope: Deactivated successfully. Sep 10 00:42:04.588451 containerd[1463]: time="2025-09-10T00:42:04.588281106Z" level=info msg="shim disconnected" id=0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3 namespace=k8s.io Sep 10 00:42:04.588451 containerd[1463]: time="2025-09-10T00:42:04.588343573Z" level=warning msg="cleaning up after shim disconnected" id=0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3 namespace=k8s.io Sep 10 00:42:04.588451 containerd[1463]: time="2025-09-10T00:42:04.588354804Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:04.768192 kubelet[2552]: E0910 00:42:04.768119 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:04.777149 containerd[1463]: time="2025-09-10T00:42:04.777092522Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 10 00:42:04.798575 containerd[1463]: time="2025-09-10T00:42:04.798517157Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45\"" Sep 10 00:42:04.799240 containerd[1463]: time="2025-09-10T00:42:04.799182156Z" level=info msg="StartContainer for \"60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45\"" Sep 10 00:42:04.830605 systemd[1]: Started cri-containerd-60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45.scope - libcontainer container 60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45. Sep 10 00:42:04.872356 containerd[1463]: time="2025-09-10T00:42:04.872169520Z" level=info msg="StartContainer for \"60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45\" returns successfully" Sep 10 00:42:04.890110 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 10 00:42:04.890493 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:42:04.890597 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:42:04.898821 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:42:04.899131 systemd[1]: cri-containerd-60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45.scope: Deactivated successfully. Sep 10 00:42:04.959810 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3-rootfs.mount: Deactivated successfully. Sep 10 00:42:05.061096 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:42:05.325681 containerd[1463]: time="2025-09-10T00:42:05.325583069Z" level=info msg="shim disconnected" id=60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45 namespace=k8s.io Sep 10 00:42:05.325681 containerd[1463]: time="2025-09-10T00:42:05.325643643Z" level=warning msg="cleaning up after shim disconnected" id=60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45 namespace=k8s.io Sep 10 00:42:05.325681 containerd[1463]: time="2025-09-10T00:42:05.325655465Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:05.771240 kubelet[2552]: E0910 00:42:05.771182 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:05.820275 containerd[1463]: time="2025-09-10T00:42:05.820154769Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 10 00:42:05.905672 containerd[1463]: time="2025-09-10T00:42:05.905604386Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14\"" Sep 10 00:42:05.906927 containerd[1463]: time="2025-09-10T00:42:05.906425369Z" level=info msg="StartContainer for \"a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14\"" Sep 10 00:42:05.958486 systemd[1]: Started cri-containerd-a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14.scope - libcontainer container a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14. Sep 10 00:42:06.001172 systemd[1]: cri-containerd-a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14.scope: Deactivated successfully. Sep 10 00:42:06.002509 containerd[1463]: time="2025-09-10T00:42:06.002458684Z" level=info msg="StartContainer for \"a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14\" returns successfully" Sep 10 00:42:06.032510 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14-rootfs.mount: Deactivated successfully. Sep 10 00:42:06.041610 containerd[1463]: time="2025-09-10T00:42:06.041533620Z" level=info msg="shim disconnected" id=a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14 namespace=k8s.io Sep 10 00:42:06.041610 containerd[1463]: time="2025-09-10T00:42:06.041601308Z" level=warning msg="cleaning up after shim disconnected" id=a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14 namespace=k8s.io Sep 10 00:42:06.041610 containerd[1463]: time="2025-09-10T00:42:06.041612719Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:06.097794 systemd[1]: Started sshd@7-10.0.0.97:22-10.0.0.1:37374.service - OpenSSH per-connection server daemon (10.0.0.1:37374). Sep 10 00:42:06.141989 sshd[3147]: Accepted publickey for core from 10.0.0.1 port 37374 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:06.144291 sshd[3147]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:06.149644 systemd-logind[1448]: New session 8 of user core. Sep 10 00:42:06.162322 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 10 00:42:06.314843 sshd[3147]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:06.319635 systemd[1]: sshd@7-10.0.0.97:22-10.0.0.1:37374.service: Deactivated successfully. Sep 10 00:42:06.321846 systemd[1]: session-8.scope: Deactivated successfully. Sep 10 00:42:06.322600 systemd-logind[1448]: Session 8 logged out. Waiting for processes to exit. Sep 10 00:42:06.323807 systemd-logind[1448]: Removed session 8. Sep 10 00:42:06.777478 kubelet[2552]: E0910 00:42:06.777439 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:06.784528 containerd[1463]: time="2025-09-10T00:42:06.784276538Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 10 00:42:06.850456 containerd[1463]: time="2025-09-10T00:42:06.850218893Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b\"" Sep 10 00:42:06.852473 containerd[1463]: time="2025-09-10T00:42:06.850893500Z" level=info msg="StartContainer for \"f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b\"" Sep 10 00:42:06.884473 systemd[1]: Started cri-containerd-f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b.scope - libcontainer container f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b. Sep 10 00:42:06.935556 systemd[1]: cri-containerd-f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b.scope: Deactivated successfully. Sep 10 00:42:06.937188 containerd[1463]: time="2025-09-10T00:42:06.937151310Z" level=info msg="StartContainer for \"f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b\" returns successfully" Sep 10 00:42:06.967268 containerd[1463]: time="2025-09-10T00:42:06.967157094Z" level=info msg="shim disconnected" id=f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b namespace=k8s.io Sep 10 00:42:06.967571 containerd[1463]: time="2025-09-10T00:42:06.967239999Z" level=warning msg="cleaning up after shim disconnected" id=f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b namespace=k8s.io Sep 10 00:42:06.967571 containerd[1463]: time="2025-09-10T00:42:06.967314119Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:07.033036 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b-rootfs.mount: Deactivated successfully. Sep 10 00:42:07.287476 containerd[1463]: time="2025-09-10T00:42:07.287330499Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:42:07.288124 containerd[1463]: time="2025-09-10T00:42:07.288051172Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Sep 10 00:42:07.289161 containerd[1463]: time="2025-09-10T00:42:07.289107236Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:42:07.290403 containerd[1463]: time="2025-09-10T00:42:07.290368213Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.433251765s" Sep 10 00:42:07.290403 containerd[1463]: time="2025-09-10T00:42:07.290400173Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 10 00:42:07.480481 containerd[1463]: time="2025-09-10T00:42:07.480411820Z" level=info msg="CreateContainer within sandbox \"08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 10 00:42:07.682470 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount675763117.mount: Deactivated successfully. Sep 10 00:42:07.685437 containerd[1463]: time="2025-09-10T00:42:07.685376507Z" level=info msg="CreateContainer within sandbox \"08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\"" Sep 10 00:42:07.687935 containerd[1463]: time="2025-09-10T00:42:07.687768829Z" level=info msg="StartContainer for \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\"" Sep 10 00:42:07.720463 systemd[1]: Started cri-containerd-375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156.scope - libcontainer container 375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156. Sep 10 00:42:08.035086 containerd[1463]: time="2025-09-10T00:42:08.034919594Z" level=info msg="StartContainer for \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\" returns successfully" Sep 10 00:42:08.046877 kubelet[2552]: E0910 00:42:08.046830 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:08.060811 kubelet[2552]: E0910 00:42:08.060765 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:08.083341 containerd[1463]: time="2025-09-10T00:42:08.080008812Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 10 00:42:08.123192 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1853931479.mount: Deactivated successfully. Sep 10 00:42:08.141085 containerd[1463]: time="2025-09-10T00:42:08.140027973Z" level=info msg="CreateContainer within sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\"" Sep 10 00:42:08.145273 containerd[1463]: time="2025-09-10T00:42:08.142764301Z" level=info msg="StartContainer for \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\"" Sep 10 00:42:08.161390 kubelet[2552]: I0910 00:42:08.161296 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-g5m7h" podStartSLOduration=3.441512791 podStartE2EDuration="26.161274288s" podCreationTimestamp="2025-09-10 00:41:42 +0000 UTC" firstStartedPulling="2025-09-10 00:41:44.571391931 +0000 UTC m=+6.990439252" lastFinishedPulling="2025-09-10 00:42:07.291153428 +0000 UTC m=+29.710200749" observedRunningTime="2025-09-10 00:42:08.096871866 +0000 UTC m=+30.515919197" watchObservedRunningTime="2025-09-10 00:42:08.161274288 +0000 UTC m=+30.580321609" Sep 10 00:42:08.214761 systemd[1]: Started cri-containerd-986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec.scope - libcontainer container 986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec. Sep 10 00:42:08.271451 containerd[1463]: time="2025-09-10T00:42:08.271209761Z" level=info msg="StartContainer for \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\" returns successfully" Sep 10 00:42:08.490184 kubelet[2552]: I0910 00:42:08.490136 2552 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 10 00:42:08.583267 systemd[1]: Created slice kubepods-burstable-pod163a26ab_d9d3_4c8e_a3f7_4b8d3c07fd34.slice - libcontainer container kubepods-burstable-pod163a26ab_d9d3_4c8e_a3f7_4b8d3c07fd34.slice. Sep 10 00:42:08.589857 kubelet[2552]: I0910 00:42:08.588383 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp5qh\" (UniqueName: \"kubernetes.io/projected/163a26ab-d9d3-4c8e-a3f7-4b8d3c07fd34-kube-api-access-qp5qh\") pod \"coredns-674b8bbfcf-54wx5\" (UID: \"163a26ab-d9d3-4c8e-a3f7-4b8d3c07fd34\") " pod="kube-system/coredns-674b8bbfcf-54wx5" Sep 10 00:42:08.589857 kubelet[2552]: I0910 00:42:08.588478 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/163a26ab-d9d3-4c8e-a3f7-4b8d3c07fd34-config-volume\") pod \"coredns-674b8bbfcf-54wx5\" (UID: \"163a26ab-d9d3-4c8e-a3f7-4b8d3c07fd34\") " pod="kube-system/coredns-674b8bbfcf-54wx5" Sep 10 00:42:08.605124 systemd[1]: Created slice kubepods-burstable-pod339b371a_119f_4c79_81ee_716635c09df5.slice - libcontainer container kubepods-burstable-pod339b371a_119f_4c79_81ee_716635c09df5.slice. Sep 10 00:42:08.689726 kubelet[2552]: I0910 00:42:08.689662 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/339b371a-119f-4c79-81ee-716635c09df5-config-volume\") pod \"coredns-674b8bbfcf-5l52m\" (UID: \"339b371a-119f-4c79-81ee-716635c09df5\") " pod="kube-system/coredns-674b8bbfcf-5l52m" Sep 10 00:42:08.689726 kubelet[2552]: I0910 00:42:08.689727 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbsd2\" (UniqueName: \"kubernetes.io/projected/339b371a-119f-4c79-81ee-716635c09df5-kube-api-access-fbsd2\") pod \"coredns-674b8bbfcf-5l52m\" (UID: \"339b371a-119f-4c79-81ee-716635c09df5\") " pod="kube-system/coredns-674b8bbfcf-5l52m" Sep 10 00:42:08.896423 kubelet[2552]: E0910 00:42:08.896217 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:08.897084 containerd[1463]: time="2025-09-10T00:42:08.897018758Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-54wx5,Uid:163a26ab-d9d3-4c8e-a3f7-4b8d3c07fd34,Namespace:kube-system,Attempt:0,}" Sep 10 00:42:08.912267 kubelet[2552]: E0910 00:42:08.911603 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:08.913342 containerd[1463]: time="2025-09-10T00:42:08.913292757Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-5l52m,Uid:339b371a-119f-4c79-81ee-716635c09df5,Namespace:kube-system,Attempt:0,}" Sep 10 00:42:09.071636 kubelet[2552]: E0910 00:42:09.071599 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:09.072338 kubelet[2552]: E0910 00:42:09.072298 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:10.073195 kubelet[2552]: E0910 00:42:10.073149 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:11.075378 kubelet[2552]: E0910 00:42:11.075332 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:11.330416 systemd[1]: Started sshd@8-10.0.0.97:22-10.0.0.1:33810.service - OpenSSH per-connection server daemon (10.0.0.1:33810). Sep 10 00:42:11.374649 sshd[3415]: Accepted publickey for core from 10.0.0.1 port 33810 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:11.376926 sshd[3415]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:11.382045 systemd-logind[1448]: New session 9 of user core. Sep 10 00:42:11.392454 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 10 00:42:11.521456 sshd[3415]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:11.527042 systemd[1]: sshd@8-10.0.0.97:22-10.0.0.1:33810.service: Deactivated successfully. Sep 10 00:42:11.529353 systemd[1]: session-9.scope: Deactivated successfully. Sep 10 00:42:11.530190 systemd-logind[1448]: Session 9 logged out. Waiting for processes to exit. Sep 10 00:42:11.531573 systemd-logind[1448]: Removed session 9. Sep 10 00:42:11.598579 systemd-networkd[1391]: cilium_host: Link UP Sep 10 00:42:11.598758 systemd-networkd[1391]: cilium_net: Link UP Sep 10 00:42:11.598763 systemd-networkd[1391]: cilium_net: Gained carrier Sep 10 00:42:11.598944 systemd-networkd[1391]: cilium_host: Gained carrier Sep 10 00:42:11.599189 systemd-networkd[1391]: cilium_host: Gained IPv6LL Sep 10 00:42:11.725312 systemd-networkd[1391]: cilium_vxlan: Link UP Sep 10 00:42:11.725329 systemd-networkd[1391]: cilium_vxlan: Gained carrier Sep 10 00:42:11.875492 systemd-networkd[1391]: cilium_net: Gained IPv6LL Sep 10 00:42:11.969323 kernel: NET: Registered PF_ALG protocol family Sep 10 00:42:12.787549 systemd-networkd[1391]: cilium_vxlan: Gained IPv6LL Sep 10 00:42:12.798501 systemd-networkd[1391]: lxc_health: Link UP Sep 10 00:42:12.811661 systemd-networkd[1391]: lxc_health: Gained carrier Sep 10 00:42:13.008085 systemd-networkd[1391]: lxc08b2004ff39d: Link UP Sep 10 00:42:13.022812 systemd-networkd[1391]: lxc2c5f5aed319a: Link UP Sep 10 00:42:13.034281 kernel: eth0: renamed from tmpa783c Sep 10 00:42:13.041397 kernel: eth0: renamed from tmpc846c Sep 10 00:42:13.050621 systemd-networkd[1391]: lxc08b2004ff39d: Gained carrier Sep 10 00:42:13.054041 systemd-networkd[1391]: lxc2c5f5aed319a: Gained carrier Sep 10 00:42:13.666857 kubelet[2552]: E0910 00:42:13.666808 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:13.690304 kubelet[2552]: I0910 00:42:13.688787 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-cwph7" podStartSLOduration=12.374693273 podStartE2EDuration="31.688768821s" podCreationTimestamp="2025-09-10 00:41:42 +0000 UTC" firstStartedPulling="2025-09-10 00:41:44.542752659 +0000 UTC m=+6.961799980" lastFinishedPulling="2025-09-10 00:42:03.856828207 +0000 UTC m=+26.275875528" observedRunningTime="2025-09-10 00:42:09.099804506 +0000 UTC m=+31.518851827" watchObservedRunningTime="2025-09-10 00:42:13.688768821 +0000 UTC m=+36.107816142" Sep 10 00:42:14.082381 kubelet[2552]: E0910 00:42:14.080895 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:14.646342 systemd-networkd[1391]: lxc08b2004ff39d: Gained IPv6LL Sep 10 00:42:14.771735 systemd-networkd[1391]: lxc2c5f5aed319a: Gained IPv6LL Sep 10 00:42:14.835531 systemd-networkd[1391]: lxc_health: Gained IPv6LL Sep 10 00:42:15.082053 kubelet[2552]: E0910 00:42:15.082017 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:16.532783 systemd[1]: Started sshd@9-10.0.0.97:22-10.0.0.1:33820.service - OpenSSH per-connection server daemon (10.0.0.1:33820). Sep 10 00:42:16.597079 sshd[3815]: Accepted publickey for core from 10.0.0.1 port 33820 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:16.598214 sshd[3815]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:16.605564 systemd-logind[1448]: New session 10 of user core. Sep 10 00:42:16.612455 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 10 00:42:16.783677 sshd[3815]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:16.787796 systemd[1]: sshd@9-10.0.0.97:22-10.0.0.1:33820.service: Deactivated successfully. Sep 10 00:42:16.790151 systemd[1]: session-10.scope: Deactivated successfully. Sep 10 00:42:16.790852 systemd-logind[1448]: Session 10 logged out. Waiting for processes to exit. Sep 10 00:42:16.792082 systemd-logind[1448]: Removed session 10. Sep 10 00:42:16.804041 containerd[1463]: time="2025-09-10T00:42:16.803910185Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:42:16.804041 containerd[1463]: time="2025-09-10T00:42:16.803994062Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:42:16.804041 containerd[1463]: time="2025-09-10T00:42:16.804007798Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:42:16.804519 containerd[1463]: time="2025-09-10T00:42:16.804109949Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:42:16.828423 systemd[1]: Started cri-containerd-a783c72ab1ded692d18b97c2afefe0daaf573b51e6bfd193ed306efb04ea3275.scope - libcontainer container a783c72ab1ded692d18b97c2afefe0daaf573b51e6bfd193ed306efb04ea3275. Sep 10 00:42:16.842377 systemd-resolved[1328]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 10 00:42:16.868868 containerd[1463]: time="2025-09-10T00:42:16.868630641Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:42:16.868868 containerd[1463]: time="2025-09-10T00:42:16.868708047Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:42:16.868868 containerd[1463]: time="2025-09-10T00:42:16.868723355Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:42:16.869206 containerd[1463]: time="2025-09-10T00:42:16.868811230Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:42:16.881081 containerd[1463]: time="2025-09-10T00:42:16.881028619Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-54wx5,Uid:163a26ab-d9d3-4c8e-a3f7-4b8d3c07fd34,Namespace:kube-system,Attempt:0,} returns sandbox id \"a783c72ab1ded692d18b97c2afefe0daaf573b51e6bfd193ed306efb04ea3275\"" Sep 10 00:42:16.882166 kubelet[2552]: E0910 00:42:16.882125 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:16.890879 systemd[1]: run-containerd-runc-k8s.io-c846c8fd4d2db9e39af6bb8fb6fdc9a012f3859818a0280303d0a48b43ca8307-runc.ZARk0A.mount: Deactivated successfully. Sep 10 00:42:16.907562 systemd[1]: Started cri-containerd-c846c8fd4d2db9e39af6bb8fb6fdc9a012f3859818a0280303d0a48b43ca8307.scope - libcontainer container c846c8fd4d2db9e39af6bb8fb6fdc9a012f3859818a0280303d0a48b43ca8307. Sep 10 00:42:16.923386 systemd-resolved[1328]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 10 00:42:16.948517 containerd[1463]: time="2025-09-10T00:42:16.948452830Z" level=info msg="CreateContainer within sandbox \"a783c72ab1ded692d18b97c2afefe0daaf573b51e6bfd193ed306efb04ea3275\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 10 00:42:16.954934 containerd[1463]: time="2025-09-10T00:42:16.954887797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-5l52m,Uid:339b371a-119f-4c79-81ee-716635c09df5,Namespace:kube-system,Attempt:0,} returns sandbox id \"c846c8fd4d2db9e39af6bb8fb6fdc9a012f3859818a0280303d0a48b43ca8307\"" Sep 10 00:42:16.955818 kubelet[2552]: E0910 00:42:16.955776 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:17.086673 containerd[1463]: time="2025-09-10T00:42:17.086429075Z" level=info msg="CreateContainer within sandbox \"c846c8fd4d2db9e39af6bb8fb6fdc9a012f3859818a0280303d0a48b43ca8307\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 10 00:42:17.548427 containerd[1463]: time="2025-09-10T00:42:17.548356725Z" level=info msg="CreateContainer within sandbox \"a783c72ab1ded692d18b97c2afefe0daaf573b51e6bfd193ed306efb04ea3275\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a5398ee33dec5f647a9568aad570667b918c8ad6f606b05079d02f822568f54f\"" Sep 10 00:42:17.549048 containerd[1463]: time="2025-09-10T00:42:17.548994562Z" level=info msg="StartContainer for \"a5398ee33dec5f647a9568aad570667b918c8ad6f606b05079d02f822568f54f\"" Sep 10 00:42:17.550629 containerd[1463]: time="2025-09-10T00:42:17.550594885Z" level=info msg="CreateContainer within sandbox \"c846c8fd4d2db9e39af6bb8fb6fdc9a012f3859818a0280303d0a48b43ca8307\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"bdb4d509ccd566b0d71119ef1773378c54a3efe89f34fbcabebaa7458581a2da\"" Sep 10 00:42:17.551198 containerd[1463]: time="2025-09-10T00:42:17.551156239Z" level=info msg="StartContainer for \"bdb4d509ccd566b0d71119ef1773378c54a3efe89f34fbcabebaa7458581a2da\"" Sep 10 00:42:17.585484 systemd[1]: Started cri-containerd-bdb4d509ccd566b0d71119ef1773378c54a3efe89f34fbcabebaa7458581a2da.scope - libcontainer container bdb4d509ccd566b0d71119ef1773378c54a3efe89f34fbcabebaa7458581a2da. Sep 10 00:42:17.589204 systemd[1]: Started cri-containerd-a5398ee33dec5f647a9568aad570667b918c8ad6f606b05079d02f822568f54f.scope - libcontainer container a5398ee33dec5f647a9568aad570667b918c8ad6f606b05079d02f822568f54f. Sep 10 00:42:17.626589 containerd[1463]: time="2025-09-10T00:42:17.626535453Z" level=info msg="StartContainer for \"a5398ee33dec5f647a9568aad570667b918c8ad6f606b05079d02f822568f54f\" returns successfully" Sep 10 00:42:17.626724 containerd[1463]: time="2025-09-10T00:42:17.626591588Z" level=info msg="StartContainer for \"bdb4d509ccd566b0d71119ef1773378c54a3efe89f34fbcabebaa7458581a2da\" returns successfully" Sep 10 00:42:18.093317 kubelet[2552]: E0910 00:42:18.093168 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:18.096356 kubelet[2552]: E0910 00:42:18.096280 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:18.408262 kubelet[2552]: I0910 00:42:18.407733 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-5l52m" podStartSLOduration=36.407708736000004 podStartE2EDuration="36.407708736s" podCreationTimestamp="2025-09-10 00:41:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:42:18.388819128 +0000 UTC m=+40.807866449" watchObservedRunningTime="2025-09-10 00:42:18.407708736 +0000 UTC m=+40.826756057" Sep 10 00:42:18.421955 kubelet[2552]: I0910 00:42:18.421849 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-54wx5" podStartSLOduration=35.421825428 podStartE2EDuration="35.421825428s" podCreationTimestamp="2025-09-10 00:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:42:18.407675845 +0000 UTC m=+40.826723166" watchObservedRunningTime="2025-09-10 00:42:18.421825428 +0000 UTC m=+40.840872759" Sep 10 00:42:19.097993 kubelet[2552]: E0910 00:42:19.097939 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:19.097993 kubelet[2552]: E0910 00:42:19.098002 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:20.105071 kubelet[2552]: E0910 00:42:20.105026 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:20.105625 kubelet[2552]: E0910 00:42:20.105218 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:21.800628 systemd[1]: Started sshd@10-10.0.0.97:22-10.0.0.1:58832.service - OpenSSH per-connection server daemon (10.0.0.1:58832). Sep 10 00:42:21.853048 sshd[4004]: Accepted publickey for core from 10.0.0.1 port 58832 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:21.856119 sshd[4004]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:21.861448 systemd-logind[1448]: New session 11 of user core. Sep 10 00:42:21.871535 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 10 00:42:22.029427 sshd[4004]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:22.034016 systemd[1]: sshd@10-10.0.0.97:22-10.0.0.1:58832.service: Deactivated successfully. Sep 10 00:42:22.036001 systemd[1]: session-11.scope: Deactivated successfully. Sep 10 00:42:22.036634 systemd-logind[1448]: Session 11 logged out. Waiting for processes to exit. Sep 10 00:42:22.037742 systemd-logind[1448]: Removed session 11. Sep 10 00:42:27.045116 systemd[1]: Started sshd@11-10.0.0.97:22-10.0.0.1:58834.service - OpenSSH per-connection server daemon (10.0.0.1:58834). Sep 10 00:42:27.092390 sshd[4019]: Accepted publickey for core from 10.0.0.1 port 58834 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:27.094868 sshd[4019]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:27.099414 systemd-logind[1448]: New session 12 of user core. Sep 10 00:42:27.114561 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 10 00:42:27.242396 sshd[4019]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:27.254306 systemd[1]: sshd@11-10.0.0.97:22-10.0.0.1:58834.service: Deactivated successfully. Sep 10 00:42:27.256358 systemd[1]: session-12.scope: Deactivated successfully. Sep 10 00:42:27.258077 systemd-logind[1448]: Session 12 logged out. Waiting for processes to exit. Sep 10 00:42:27.269500 systemd[1]: Started sshd@12-10.0.0.97:22-10.0.0.1:58846.service - OpenSSH per-connection server daemon (10.0.0.1:58846). Sep 10 00:42:27.270777 systemd-logind[1448]: Removed session 12. Sep 10 00:42:27.306298 sshd[4034]: Accepted publickey for core from 10.0.0.1 port 58846 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:27.308396 sshd[4034]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:27.313039 systemd-logind[1448]: New session 13 of user core. Sep 10 00:42:27.320374 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 10 00:42:27.564775 sshd[4034]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:27.580318 systemd[1]: sshd@12-10.0.0.97:22-10.0.0.1:58846.service: Deactivated successfully. Sep 10 00:42:27.583354 systemd[1]: session-13.scope: Deactivated successfully. Sep 10 00:42:27.585811 systemd-logind[1448]: Session 13 logged out. Waiting for processes to exit. Sep 10 00:42:27.594698 systemd[1]: Started sshd@13-10.0.0.97:22-10.0.0.1:58854.service - OpenSSH per-connection server daemon (10.0.0.1:58854). Sep 10 00:42:27.596187 systemd-logind[1448]: Removed session 13. Sep 10 00:42:27.630098 sshd[4046]: Accepted publickey for core from 10.0.0.1 port 58854 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:27.632149 sshd[4046]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:27.637405 systemd-logind[1448]: New session 14 of user core. Sep 10 00:42:27.644389 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 10 00:42:27.762908 sshd[4046]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:27.768222 systemd[1]: sshd@13-10.0.0.97:22-10.0.0.1:58854.service: Deactivated successfully. Sep 10 00:42:27.770460 systemd[1]: session-14.scope: Deactivated successfully. Sep 10 00:42:27.771331 systemd-logind[1448]: Session 14 logged out. Waiting for processes to exit. Sep 10 00:42:27.772742 systemd-logind[1448]: Removed session 14. Sep 10 00:42:32.783890 systemd[1]: Started sshd@14-10.0.0.97:22-10.0.0.1:59494.service - OpenSSH per-connection server daemon (10.0.0.1:59494). Sep 10 00:42:32.822809 sshd[4061]: Accepted publickey for core from 10.0.0.1 port 59494 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:32.824642 sshd[4061]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:32.829010 systemd-logind[1448]: New session 15 of user core. Sep 10 00:42:32.839405 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 10 00:42:32.957531 sshd[4061]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:32.962859 systemd[1]: sshd@14-10.0.0.97:22-10.0.0.1:59494.service: Deactivated successfully. Sep 10 00:42:32.964955 systemd[1]: session-15.scope: Deactivated successfully. Sep 10 00:42:32.965718 systemd-logind[1448]: Session 15 logged out. Waiting for processes to exit. Sep 10 00:42:32.966805 systemd-logind[1448]: Removed session 15. Sep 10 00:42:37.970778 systemd[1]: Started sshd@15-10.0.0.97:22-10.0.0.1:59502.service - OpenSSH per-connection server daemon (10.0.0.1:59502). Sep 10 00:42:38.024456 sshd[4077]: Accepted publickey for core from 10.0.0.1 port 59502 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:38.026915 sshd[4077]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:38.033315 systemd-logind[1448]: New session 16 of user core. Sep 10 00:42:38.040606 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 10 00:42:38.182158 sshd[4077]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:38.197459 systemd[1]: sshd@15-10.0.0.97:22-10.0.0.1:59502.service: Deactivated successfully. Sep 10 00:42:38.200905 systemd[1]: session-16.scope: Deactivated successfully. Sep 10 00:42:38.203120 systemd-logind[1448]: Session 16 logged out. Waiting for processes to exit. Sep 10 00:42:38.214899 systemd[1]: Started sshd@16-10.0.0.97:22-10.0.0.1:59518.service - OpenSSH per-connection server daemon (10.0.0.1:59518). Sep 10 00:42:38.217059 systemd-logind[1448]: Removed session 16. Sep 10 00:42:38.254609 sshd[4092]: Accepted publickey for core from 10.0.0.1 port 59518 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:38.256507 sshd[4092]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:38.263150 systemd-logind[1448]: New session 17 of user core. Sep 10 00:42:38.270488 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 10 00:42:38.669201 sshd[4092]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:38.685406 systemd[1]: sshd@16-10.0.0.97:22-10.0.0.1:59518.service: Deactivated successfully. Sep 10 00:42:38.687320 systemd[1]: session-17.scope: Deactivated successfully. Sep 10 00:42:38.688962 systemd-logind[1448]: Session 17 logged out. Waiting for processes to exit. Sep 10 00:42:38.690901 systemd[1]: Started sshd@17-10.0.0.97:22-10.0.0.1:59520.service - OpenSSH per-connection server daemon (10.0.0.1:59520). Sep 10 00:42:38.692439 systemd-logind[1448]: Removed session 17. Sep 10 00:42:38.730203 sshd[4104]: Accepted publickey for core from 10.0.0.1 port 59520 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:38.732088 sshd[4104]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:38.737184 systemd-logind[1448]: New session 18 of user core. Sep 10 00:42:38.746435 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 10 00:42:39.513221 sshd[4104]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:39.522880 systemd[1]: sshd@17-10.0.0.97:22-10.0.0.1:59520.service: Deactivated successfully. Sep 10 00:42:39.528031 systemd[1]: session-18.scope: Deactivated successfully. Sep 10 00:42:39.532917 systemd-logind[1448]: Session 18 logged out. Waiting for processes to exit. Sep 10 00:42:39.539571 systemd[1]: Started sshd@18-10.0.0.97:22-10.0.0.1:59530.service - OpenSSH per-connection server daemon (10.0.0.1:59530). Sep 10 00:42:39.541353 systemd-logind[1448]: Removed session 18. Sep 10 00:42:39.578960 sshd[4124]: Accepted publickey for core from 10.0.0.1 port 59530 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:39.580709 sshd[4124]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:39.585238 systemd-logind[1448]: New session 19 of user core. Sep 10 00:42:39.594411 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 10 00:42:39.839411 sshd[4124]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:39.846493 systemd[1]: sshd@18-10.0.0.97:22-10.0.0.1:59530.service: Deactivated successfully. Sep 10 00:42:39.848514 systemd[1]: session-19.scope: Deactivated successfully. Sep 10 00:42:39.851743 systemd-logind[1448]: Session 19 logged out. Waiting for processes to exit. Sep 10 00:42:39.872733 systemd[1]: Started sshd@19-10.0.0.97:22-10.0.0.1:59540.service - OpenSSH per-connection server daemon (10.0.0.1:59540). Sep 10 00:42:39.873807 systemd-logind[1448]: Removed session 19. Sep 10 00:42:39.907153 sshd[4137]: Accepted publickey for core from 10.0.0.1 port 59540 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:39.909498 sshd[4137]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:39.914515 systemd-logind[1448]: New session 20 of user core. Sep 10 00:42:39.924402 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 10 00:42:40.045079 sshd[4137]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:40.049300 systemd[1]: sshd@19-10.0.0.97:22-10.0.0.1:59540.service: Deactivated successfully. Sep 10 00:42:40.051453 systemd[1]: session-20.scope: Deactivated successfully. Sep 10 00:42:40.052153 systemd-logind[1448]: Session 20 logged out. Waiting for processes to exit. Sep 10 00:42:40.053267 systemd-logind[1448]: Removed session 20. Sep 10 00:42:45.060731 systemd[1]: Started sshd@20-10.0.0.97:22-10.0.0.1:34390.service - OpenSSH per-connection server daemon (10.0.0.1:34390). Sep 10 00:42:45.102102 sshd[4153]: Accepted publickey for core from 10.0.0.1 port 34390 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:45.104401 sshd[4153]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:45.109048 systemd-logind[1448]: New session 21 of user core. Sep 10 00:42:45.116386 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 10 00:42:45.227581 sshd[4153]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:45.231385 systemd[1]: sshd@20-10.0.0.97:22-10.0.0.1:34390.service: Deactivated successfully. Sep 10 00:42:45.233502 systemd[1]: session-21.scope: Deactivated successfully. Sep 10 00:42:45.234174 systemd-logind[1448]: Session 21 logged out. Waiting for processes to exit. Sep 10 00:42:45.235097 systemd-logind[1448]: Removed session 21. Sep 10 00:42:50.239646 systemd[1]: Started sshd@21-10.0.0.97:22-10.0.0.1:38222.service - OpenSSH per-connection server daemon (10.0.0.1:38222). Sep 10 00:42:50.281146 sshd[4169]: Accepted publickey for core from 10.0.0.1 port 38222 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:50.283308 sshd[4169]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:50.287509 systemd-logind[1448]: New session 22 of user core. Sep 10 00:42:50.298413 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 10 00:42:50.424353 sshd[4169]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:50.429107 systemd[1]: sshd@21-10.0.0.97:22-10.0.0.1:38222.service: Deactivated successfully. Sep 10 00:42:50.431418 systemd[1]: session-22.scope: Deactivated successfully. Sep 10 00:42:50.432164 systemd-logind[1448]: Session 22 logged out. Waiting for processes to exit. Sep 10 00:42:50.433335 systemd-logind[1448]: Removed session 22. Sep 10 00:42:51.703044 kubelet[2552]: E0910 00:42:51.703003 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:52.700819 kubelet[2552]: E0910 00:42:52.700769 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:42:55.439194 systemd[1]: Started sshd@22-10.0.0.97:22-10.0.0.1:38230.service - OpenSSH per-connection server daemon (10.0.0.1:38230). Sep 10 00:42:55.476614 sshd[4183]: Accepted publickey for core from 10.0.0.1 port 38230 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:55.478197 sshd[4183]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:55.482372 systemd-logind[1448]: New session 23 of user core. Sep 10 00:42:55.489383 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 10 00:42:55.596475 sshd[4183]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:55.607609 systemd[1]: sshd@22-10.0.0.97:22-10.0.0.1:38230.service: Deactivated successfully. Sep 10 00:42:55.609681 systemd[1]: session-23.scope: Deactivated successfully. Sep 10 00:42:55.611701 systemd-logind[1448]: Session 23 logged out. Waiting for processes to exit. Sep 10 00:42:55.616576 systemd[1]: Started sshd@23-10.0.0.97:22-10.0.0.1:38234.service - OpenSSH per-connection server daemon (10.0.0.1:38234). Sep 10 00:42:55.617556 systemd-logind[1448]: Removed session 23. Sep 10 00:42:55.652230 sshd[4197]: Accepted publickey for core from 10.0.0.1 port 38234 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:55.654120 sshd[4197]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:55.658529 systemd-logind[1448]: New session 24 of user core. Sep 10 00:42:55.667571 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 10 00:42:57.025194 containerd[1463]: time="2025-09-10T00:42:57.025102028Z" level=info msg="StopContainer for \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\" with timeout 30 (s)" Sep 10 00:42:57.026075 containerd[1463]: time="2025-09-10T00:42:57.025805296Z" level=info msg="Stop container \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\" with signal terminated" Sep 10 00:42:57.060984 systemd[1]: cri-containerd-375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156.scope: Deactivated successfully. Sep 10 00:42:57.072039 systemd[1]: run-containerd-runc-k8s.io-986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec-runc.nSVzXY.mount: Deactivated successfully. Sep 10 00:42:57.096004 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156-rootfs.mount: Deactivated successfully. Sep 10 00:42:57.097886 containerd[1463]: time="2025-09-10T00:42:57.096832419Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 10 00:42:57.099908 containerd[1463]: time="2025-09-10T00:42:57.099876755Z" level=info msg="StopContainer for \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\" with timeout 2 (s)" Sep 10 00:42:57.100145 containerd[1463]: time="2025-09-10T00:42:57.100120309Z" level=info msg="Stop container \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\" with signal terminated" Sep 10 00:42:57.104473 containerd[1463]: time="2025-09-10T00:42:57.104412609Z" level=info msg="shim disconnected" id=375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156 namespace=k8s.io Sep 10 00:42:57.104473 containerd[1463]: time="2025-09-10T00:42:57.104467774Z" level=warning msg="cleaning up after shim disconnected" id=375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156 namespace=k8s.io Sep 10 00:42:57.104473 containerd[1463]: time="2025-09-10T00:42:57.104479476Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:57.108759 systemd-networkd[1391]: lxc_health: Link DOWN Sep 10 00:42:57.108769 systemd-networkd[1391]: lxc_health: Lost carrier Sep 10 00:42:57.136074 systemd[1]: cri-containerd-986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec.scope: Deactivated successfully. Sep 10 00:42:57.136502 systemd[1]: cri-containerd-986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec.scope: Consumed 7.646s CPU time. Sep 10 00:42:57.138013 containerd[1463]: time="2025-09-10T00:42:57.137955524Z" level=info msg="StopContainer for \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\" returns successfully" Sep 10 00:42:57.138803 containerd[1463]: time="2025-09-10T00:42:57.138764995Z" level=info msg="StopPodSandbox for \"08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e\"" Sep 10 00:42:57.138884 containerd[1463]: time="2025-09-10T00:42:57.138812005Z" level=info msg="Container to stop \"375a34828bfaaee05a0d1548c32e708de0c149139e74fb7511e69128e9d71156\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:42:57.143635 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e-shm.mount: Deactivated successfully. Sep 10 00:42:57.158952 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec-rootfs.mount: Deactivated successfully. Sep 10 00:42:57.161147 systemd[1]: cri-containerd-08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e.scope: Deactivated successfully. Sep 10 00:42:57.176089 containerd[1463]: time="2025-09-10T00:42:57.175845275Z" level=info msg="shim disconnected" id=986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec namespace=k8s.io Sep 10 00:42:57.176089 containerd[1463]: time="2025-09-10T00:42:57.175911892Z" level=warning msg="cleaning up after shim disconnected" id=986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec namespace=k8s.io Sep 10 00:42:57.176089 containerd[1463]: time="2025-09-10T00:42:57.175923834Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:57.188513 containerd[1463]: time="2025-09-10T00:42:57.188305925Z" level=info msg="shim disconnected" id=08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e namespace=k8s.io Sep 10 00:42:57.188513 containerd[1463]: time="2025-09-10T00:42:57.188367593Z" level=warning msg="cleaning up after shim disconnected" id=08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e namespace=k8s.io Sep 10 00:42:57.188513 containerd[1463]: time="2025-09-10T00:42:57.188376109Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:57.204614 containerd[1463]: time="2025-09-10T00:42:57.204548687Z" level=info msg="StopContainer for \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\" returns successfully" Sep 10 00:42:57.205081 containerd[1463]: time="2025-09-10T00:42:57.205056714Z" level=info msg="StopPodSandbox for \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\"" Sep 10 00:42:57.205152 containerd[1463]: time="2025-09-10T00:42:57.205091511Z" level=info msg="Container to stop \"60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:42:57.205152 containerd[1463]: time="2025-09-10T00:42:57.205103112Z" level=info msg="Container to stop \"f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:42:57.205152 containerd[1463]: time="2025-09-10T00:42:57.205111608Z" level=info msg="Container to stop \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:42:57.205152 containerd[1463]: time="2025-09-10T00:42:57.205120215Z" level=info msg="Container to stop \"0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:42:57.205152 containerd[1463]: time="2025-09-10T00:42:57.205131977Z" level=info msg="Container to stop \"a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:42:57.212023 systemd[1]: cri-containerd-c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872.scope: Deactivated successfully. Sep 10 00:42:57.221160 kubelet[2552]: I0910 00:42:57.221071 2552 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e" Sep 10 00:42:57.222778 containerd[1463]: time="2025-09-10T00:42:57.222708438Z" level=info msg="TearDown network for sandbox \"08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e\" successfully" Sep 10 00:42:57.223181 containerd[1463]: time="2025-09-10T00:42:57.222831732Z" level=info msg="StopPodSandbox for \"08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e\" returns successfully" Sep 10 00:42:57.244588 containerd[1463]: time="2025-09-10T00:42:57.244426461Z" level=info msg="shim disconnected" id=c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872 namespace=k8s.io Sep 10 00:42:57.244588 containerd[1463]: time="2025-09-10T00:42:57.244508938Z" level=warning msg="cleaning up after shim disconnected" id=c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872 namespace=k8s.io Sep 10 00:42:57.244588 containerd[1463]: time="2025-09-10T00:42:57.244520401Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:42:57.265839 containerd[1463]: time="2025-09-10T00:42:57.265778990Z" level=info msg="TearDown network for sandbox \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" successfully" Sep 10 00:42:57.265839 containerd[1463]: time="2025-09-10T00:42:57.265825097Z" level=info msg="StopPodSandbox for \"c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872\" returns successfully" Sep 10 00:42:57.290476 kubelet[2552]: I0910 00:42:57.289736 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s92fg\" (UniqueName: \"kubernetes.io/projected/21e19d89-fbf5-4b8f-b135-124fe927b8a1-kube-api-access-s92fg\") pod \"21e19d89-fbf5-4b8f-b135-124fe927b8a1\" (UID: \"21e19d89-fbf5-4b8f-b135-124fe927b8a1\") " Sep 10 00:42:57.290476 kubelet[2552]: I0910 00:42:57.289805 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/21e19d89-fbf5-4b8f-b135-124fe927b8a1-cilium-config-path\") pod \"21e19d89-fbf5-4b8f-b135-124fe927b8a1\" (UID: \"21e19d89-fbf5-4b8f-b135-124fe927b8a1\") " Sep 10 00:42:57.293779 kubelet[2552]: I0910 00:42:57.293744 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21e19d89-fbf5-4b8f-b135-124fe927b8a1-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "21e19d89-fbf5-4b8f-b135-124fe927b8a1" (UID: "21e19d89-fbf5-4b8f-b135-124fe927b8a1"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 10 00:42:57.294126 kubelet[2552]: I0910 00:42:57.294063 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e19d89-fbf5-4b8f-b135-124fe927b8a1-kube-api-access-s92fg" (OuterVolumeSpecName: "kube-api-access-s92fg") pod "21e19d89-fbf5-4b8f-b135-124fe927b8a1" (UID: "21e19d89-fbf5-4b8f-b135-124fe927b8a1"). InnerVolumeSpecName "kube-api-access-s92fg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 10 00:42:57.390560 kubelet[2552]: I0910 00:42:57.390494 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-xtables-lock\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.390560 kubelet[2552]: I0910 00:42:57.390551 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-net\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.390560 kubelet[2552]: I0910 00:42:57.390573 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hostproc\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.390836 kubelet[2552]: I0910 00:42:57.390606 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hubble-tls\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.390836 kubelet[2552]: I0910 00:42:57.390626 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-bpf-maps\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.390836 kubelet[2552]: I0910 00:42:57.390612 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.390836 kubelet[2552]: I0910 00:42:57.390643 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cni-path\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.390836 kubelet[2552]: I0910 00:42:57.390671 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hostproc" (OuterVolumeSpecName: "hostproc") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391004 kubelet[2552]: I0910 00:42:57.390671 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cni-path" (OuterVolumeSpecName: "cni-path") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391004 kubelet[2552]: I0910 00:42:57.390715 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-run\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391004 kubelet[2552]: I0910 00:42:57.390721 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391004 kubelet[2552]: I0910 00:42:57.390738 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391004 kubelet[2552]: I0910 00:42:57.390754 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391180 kubelet[2552]: I0910 00:42:57.390760 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ckjl\" (UniqueName: \"kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-kube-api-access-2ckjl\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391180 kubelet[2552]: I0910 00:42:57.390797 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-cgroup\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391180 kubelet[2552]: I0910 00:42:57.390814 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-lib-modules\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391180 kubelet[2552]: I0910 00:42:57.390830 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-kernel\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391180 kubelet[2552]: I0910 00:42:57.390852 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-config-path\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391180 kubelet[2552]: I0910 00:42:57.390874 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-etc-cni-netd\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390891 2552 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/cd0204fb-1928-4cc6-b8b5-673ee11acce8-clustermesh-secrets\") pod \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\" (UID: \"cd0204fb-1928-4cc6-b8b5-673ee11acce8\") " Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390923 2552 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-s92fg\" (UniqueName: \"kubernetes.io/projected/21e19d89-fbf5-4b8f-b135-124fe927b8a1-kube-api-access-s92fg\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390933 2552 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390942 2552 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390950 2552 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390966 2552 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/21e19d89-fbf5-4b8f-b135-124fe927b8a1-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390974 2552 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391415 kubelet[2552]: I0910 00:42:57.390984 2552 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391692 kubelet[2552]: I0910 00:42:57.390992 2552 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.391692 kubelet[2552]: I0910 00:42:57.391286 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391692 kubelet[2552]: I0910 00:42:57.391316 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.391692 kubelet[2552]: I0910 00:42:57.391333 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.394411 kubelet[2552]: I0910 00:42:57.394307 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-kube-api-access-2ckjl" (OuterVolumeSpecName: "kube-api-access-2ckjl") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "kube-api-access-2ckjl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 10 00:42:57.394411 kubelet[2552]: I0910 00:42:57.394353 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:42:57.394780 kubelet[2552]: I0910 00:42:57.394755 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 10 00:42:57.394904 kubelet[2552]: I0910 00:42:57.394867 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0204fb-1928-4cc6-b8b5-673ee11acce8-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 10 00:42:57.394943 kubelet[2552]: I0910 00:42:57.394916 2552 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "cd0204fb-1928-4cc6-b8b5-673ee11acce8" (UID: "cd0204fb-1928-4cc6-b8b5-673ee11acce8"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 10 00:42:57.491461 kubelet[2552]: I0910 00:42:57.491398 2552 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2ckjl\" (UniqueName: \"kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-kube-api-access-2ckjl\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491461 kubelet[2552]: I0910 00:42:57.491446 2552 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491461 kubelet[2552]: I0910 00:42:57.491463 2552 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491461 kubelet[2552]: I0910 00:42:57.491474 2552 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491461 kubelet[2552]: I0910 00:42:57.491485 2552 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd0204fb-1928-4cc6-b8b5-673ee11acce8-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491808 kubelet[2552]: I0910 00:42:57.491494 2552 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd0204fb-1928-4cc6-b8b5-673ee11acce8-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491808 kubelet[2552]: I0910 00:42:57.491502 2552 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/cd0204fb-1928-4cc6-b8b5-673ee11acce8-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.491808 kubelet[2552]: I0910 00:42:57.491509 2552 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/cd0204fb-1928-4cc6-b8b5-673ee11acce8-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 10 00:42:57.708900 systemd[1]: Removed slice kubepods-besteffort-pod21e19d89_fbf5_4b8f_b135_124fe927b8a1.slice - libcontainer container kubepods-besteffort-pod21e19d89_fbf5_4b8f_b135_124fe927b8a1.slice. Sep 10 00:42:57.710101 systemd[1]: Removed slice kubepods-burstable-podcd0204fb_1928_4cc6_b8b5_673ee11acce8.slice - libcontainer container kubepods-burstable-podcd0204fb_1928_4cc6_b8b5_673ee11acce8.slice. Sep 10 00:42:57.710198 systemd[1]: kubepods-burstable-podcd0204fb_1928_4cc6_b8b5_673ee11acce8.slice: Consumed 7.781s CPU time. Sep 10 00:42:57.751804 kubelet[2552]: E0910 00:42:57.751747 2552 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 10 00:42:58.063902 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-08f2271dc92b61ac074f55bdb4d673920f9019b3f96e02dd6725a57af595755e-rootfs.mount: Deactivated successfully. Sep 10 00:42:58.064034 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872-rootfs.mount: Deactivated successfully. Sep 10 00:42:58.064114 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c5b891f1196bc97567c3236cf5ceabaab760649cca912fb1f7ea77566f74f872-shm.mount: Deactivated successfully. Sep 10 00:42:58.064192 systemd[1]: var-lib-kubelet-pods-21e19d89\x2dfbf5\x2d4b8f\x2db135\x2d124fe927b8a1-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ds92fg.mount: Deactivated successfully. Sep 10 00:42:58.064308 systemd[1]: var-lib-kubelet-pods-cd0204fb\x2d1928\x2d4cc6\x2db8b5\x2d673ee11acce8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d2ckjl.mount: Deactivated successfully. Sep 10 00:42:58.064392 systemd[1]: var-lib-kubelet-pods-cd0204fb\x2d1928\x2d4cc6\x2db8b5\x2d673ee11acce8-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 10 00:42:58.064474 systemd[1]: var-lib-kubelet-pods-cd0204fb\x2d1928\x2d4cc6\x2db8b5\x2d673ee11acce8-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 10 00:42:58.229011 kubelet[2552]: I0910 00:42:58.228968 2552 scope.go:117] "RemoveContainer" containerID="986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec" Sep 10 00:42:58.231156 containerd[1463]: time="2025-09-10T00:42:58.231110457Z" level=info msg="RemoveContainer for \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\"" Sep 10 00:42:58.409164 containerd[1463]: time="2025-09-10T00:42:58.409113762Z" level=info msg="RemoveContainer for \"986827347a10d732b44b6b96bff4abdc3365d402f9dbdd27991e624c03f944ec\" returns successfully" Sep 10 00:42:58.409488 kubelet[2552]: I0910 00:42:58.409440 2552 scope.go:117] "RemoveContainer" containerID="f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b" Sep 10 00:42:58.410874 containerd[1463]: time="2025-09-10T00:42:58.410837703Z" level=info msg="RemoveContainer for \"f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b\"" Sep 10 00:42:58.584229 containerd[1463]: time="2025-09-10T00:42:58.583996841Z" level=info msg="RemoveContainer for \"f05639bd0bb06afbc853ebc10a6ec35f7899ad803a71fc76426de6ab7702777b\" returns successfully" Sep 10 00:42:58.584420 kubelet[2552]: I0910 00:42:58.584226 2552 scope.go:117] "RemoveContainer" containerID="a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14" Sep 10 00:42:58.586113 containerd[1463]: time="2025-09-10T00:42:58.586085556Z" level=info msg="RemoveContainer for \"a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14\"" Sep 10 00:42:58.601437 containerd[1463]: time="2025-09-10T00:42:58.601370453Z" level=info msg="RemoveContainer for \"a7e5b61ee4816a2624fda07d500b28ae036314c6aa88101faf3c4ceec159fc14\" returns successfully" Sep 10 00:42:58.601674 kubelet[2552]: I0910 00:42:58.601640 2552 scope.go:117] "RemoveContainer" containerID="60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45" Sep 10 00:42:58.630532 containerd[1463]: time="2025-09-10T00:42:58.630473297Z" level=info msg="RemoveContainer for \"60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45\"" Sep 10 00:42:58.634035 containerd[1463]: time="2025-09-10T00:42:58.634004998Z" level=info msg="RemoveContainer for \"60775f7ae9b2ce1b5a60a047bd7ad16c0c3a18e75494cfa2b8655d43e4f39e45\" returns successfully" Sep 10 00:42:58.634188 kubelet[2552]: I0910 00:42:58.634159 2552 scope.go:117] "RemoveContainer" containerID="0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3" Sep 10 00:42:58.635318 containerd[1463]: time="2025-09-10T00:42:58.635266419Z" level=info msg="RemoveContainer for \"0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3\"" Sep 10 00:42:58.638313 containerd[1463]: time="2025-09-10T00:42:58.638291986Z" level=info msg="RemoveContainer for \"0cc5f7e6d6ca0ba4a28c6c07fc53d544c01a546540b0cd6b423900fb4c87cee3\" returns successfully" Sep 10 00:42:58.991851 sshd[4197]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:59.000049 systemd[1]: sshd@23-10.0.0.97:22-10.0.0.1:38234.service: Deactivated successfully. Sep 10 00:42:59.002686 systemd[1]: session-24.scope: Deactivated successfully. Sep 10 00:42:59.004723 systemd-logind[1448]: Session 24 logged out. Waiting for processes to exit. Sep 10 00:42:59.015715 systemd[1]: Started sshd@24-10.0.0.97:22-10.0.0.1:38250.service - OpenSSH per-connection server daemon (10.0.0.1:38250). Sep 10 00:42:59.016919 systemd-logind[1448]: Removed session 24. Sep 10 00:42:59.055060 sshd[4360]: Accepted publickey for core from 10.0.0.1 port 38250 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:59.057302 sshd[4360]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:59.061996 systemd-logind[1448]: New session 25 of user core. Sep 10 00:42:59.080553 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 10 00:42:59.671943 sshd[4360]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:59.687115 systemd[1]: sshd@24-10.0.0.97:22-10.0.0.1:38250.service: Deactivated successfully. Sep 10 00:42:59.694850 systemd[1]: session-25.scope: Deactivated successfully. Sep 10 00:42:59.700316 systemd-logind[1448]: Session 25 logged out. Waiting for processes to exit. Sep 10 00:42:59.704640 kubelet[2552]: I0910 00:42:59.704555 2552 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e19d89-fbf5-4b8f-b135-124fe927b8a1" path="/var/lib/kubelet/pods/21e19d89-fbf5-4b8f-b135-124fe927b8a1/volumes" Sep 10 00:42:59.715277 kubelet[2552]: I0910 00:42:59.707972 2552 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd0204fb-1928-4cc6-b8b5-673ee11acce8" path="/var/lib/kubelet/pods/cd0204fb-1928-4cc6-b8b5-673ee11acce8/volumes" Sep 10 00:42:59.711130 systemd[1]: Started sshd@25-10.0.0.97:22-10.0.0.1:38266.service - OpenSSH per-connection server daemon (10.0.0.1:38266). Sep 10 00:42:59.712521 systemd-logind[1448]: Removed session 25. Sep 10 00:42:59.730711 systemd[1]: Created slice kubepods-burstable-podaed0f297_2dd3_4b82_afaf_8dc13013880f.slice - libcontainer container kubepods-burstable-podaed0f297_2dd3_4b82_afaf_8dc13013880f.slice. Sep 10 00:42:59.777961 sshd[4373]: Accepted publickey for core from 10.0.0.1 port 38266 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:59.779828 sshd[4373]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:59.784112 systemd-logind[1448]: New session 26 of user core. Sep 10 00:42:59.785535 kubelet[2552]: I0910 00:42:59.785485 2552 setters.go:618] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-10T00:42:59Z","lastTransitionTime":"2025-09-10T00:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 10 00:42:59.793561 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 10 00:42:59.804536 kubelet[2552]: I0910 00:42:59.804485 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-etc-cni-netd\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804536 kubelet[2552]: I0910 00:42:59.804522 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-xtables-lock\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804536 kubelet[2552]: I0910 00:42:59.804543 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aed0f297-2dd3-4b82-afaf-8dc13013880f-clustermesh-secrets\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804749 kubelet[2552]: I0910 00:42:59.804632 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-host-proc-sys-kernel\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804749 kubelet[2552]: I0910 00:42:59.804686 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-bpf-maps\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804749 kubelet[2552]: I0910 00:42:59.804711 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-cilium-run\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804749 kubelet[2552]: I0910 00:42:59.804726 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/aed0f297-2dd3-4b82-afaf-8dc13013880f-cilium-ipsec-secrets\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804749 kubelet[2552]: I0910 00:42:59.804741 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-host-proc-sys-net\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804864 kubelet[2552]: I0910 00:42:59.804756 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aed0f297-2dd3-4b82-afaf-8dc13013880f-cilium-config-path\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804864 kubelet[2552]: I0910 00:42:59.804813 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aed0f297-2dd3-4b82-afaf-8dc13013880f-hubble-tls\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804913 kubelet[2552]: I0910 00:42:59.804865 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-cni-path\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804913 kubelet[2552]: I0910 00:42:59.804881 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-lib-modules\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804913 kubelet[2552]: I0910 00:42:59.804904 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-cilium-cgroup\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804983 kubelet[2552]: I0910 00:42:59.804922 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aed0f297-2dd3-4b82-afaf-8dc13013880f-hostproc\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.804983 kubelet[2552]: I0910 00:42:59.804940 2552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdhnd\" (UniqueName: \"kubernetes.io/projected/aed0f297-2dd3-4b82-afaf-8dc13013880f-kube-api-access-cdhnd\") pod \"cilium-jsjnz\" (UID: \"aed0f297-2dd3-4b82-afaf-8dc13013880f\") " pod="kube-system/cilium-jsjnz" Sep 10 00:42:59.847896 sshd[4373]: pam_unix(sshd:session): session closed for user core Sep 10 00:42:59.856280 systemd[1]: sshd@25-10.0.0.97:22-10.0.0.1:38266.service: Deactivated successfully. Sep 10 00:42:59.858271 systemd[1]: session-26.scope: Deactivated successfully. Sep 10 00:42:59.860155 systemd-logind[1448]: Session 26 logged out. Waiting for processes to exit. Sep 10 00:42:59.866553 systemd[1]: Started sshd@26-10.0.0.97:22-10.0.0.1:38282.service - OpenSSH per-connection server daemon (10.0.0.1:38282). Sep 10 00:42:59.867592 systemd-logind[1448]: Removed session 26. Sep 10 00:42:59.901166 sshd[4381]: Accepted publickey for core from 10.0.0.1 port 38282 ssh2: RSA SHA256:8lYmw5fyCyWfPmBBOTh1KYsG06iZ45OCbq9sG6CkCSY Sep 10 00:42:59.903036 sshd[4381]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:42:59.910322 systemd-logind[1448]: New session 27 of user core. Sep 10 00:42:59.945462 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 10 00:43:00.050029 kubelet[2552]: E0910 00:43:00.049985 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:00.050713 containerd[1463]: time="2025-09-10T00:43:00.050676294Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jsjnz,Uid:aed0f297-2dd3-4b82-afaf-8dc13013880f,Namespace:kube-system,Attempt:0,}" Sep 10 00:43:00.083419 containerd[1463]: time="2025-09-10T00:43:00.082393894Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:43:00.083419 containerd[1463]: time="2025-09-10T00:43:00.083341316Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:43:00.083419 containerd[1463]: time="2025-09-10T00:43:00.083357217Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:43:00.083642 containerd[1463]: time="2025-09-10T00:43:00.083472566Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:43:00.113442 systemd[1]: Started cri-containerd-9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc.scope - libcontainer container 9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc. Sep 10 00:43:00.144868 containerd[1463]: time="2025-09-10T00:43:00.144663388Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jsjnz,Uid:aed0f297-2dd3-4b82-afaf-8dc13013880f,Namespace:kube-system,Attempt:0,} returns sandbox id \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\"" Sep 10 00:43:00.146601 kubelet[2552]: E0910 00:43:00.145644 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:00.168897 containerd[1463]: time="2025-09-10T00:43:00.168741149Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 10 00:43:00.182506 containerd[1463]: time="2025-09-10T00:43:00.182445531Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830\"" Sep 10 00:43:00.182972 containerd[1463]: time="2025-09-10T00:43:00.182942888Z" level=info msg="StartContainer for \"afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830\"" Sep 10 00:43:00.215457 systemd[1]: Started cri-containerd-afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830.scope - libcontainer container afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830. Sep 10 00:43:00.247904 containerd[1463]: time="2025-09-10T00:43:00.247853593Z" level=info msg="StartContainer for \"afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830\" returns successfully" Sep 10 00:43:00.260558 systemd[1]: cri-containerd-afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830.scope: Deactivated successfully. Sep 10 00:43:00.291918 containerd[1463]: time="2025-09-10T00:43:00.291852160Z" level=info msg="shim disconnected" id=afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830 namespace=k8s.io Sep 10 00:43:00.291918 containerd[1463]: time="2025-09-10T00:43:00.291910621Z" level=warning msg="cleaning up after shim disconnected" id=afbb2faa77272eb85c1254cbce4f6066089cace3f2d344be402e5c9fb9d5e830 namespace=k8s.io Sep 10 00:43:00.291918 containerd[1463]: time="2025-09-10T00:43:00.291919267Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:43:01.241472 kubelet[2552]: E0910 00:43:01.241436 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:01.247989 containerd[1463]: time="2025-09-10T00:43:01.247928859Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 10 00:43:01.269159 containerd[1463]: time="2025-09-10T00:43:01.269102263Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7\"" Sep 10 00:43:01.269693 containerd[1463]: time="2025-09-10T00:43:01.269666296Z" level=info msg="StartContainer for \"775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7\"" Sep 10 00:43:01.305458 systemd[1]: Started cri-containerd-775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7.scope - libcontainer container 775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7. Sep 10 00:43:01.335636 containerd[1463]: time="2025-09-10T00:43:01.335552719Z" level=info msg="StartContainer for \"775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7\" returns successfully" Sep 10 00:43:01.344613 systemd[1]: cri-containerd-775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7.scope: Deactivated successfully. Sep 10 00:43:01.371920 containerd[1463]: time="2025-09-10T00:43:01.371837513Z" level=info msg="shim disconnected" id=775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7 namespace=k8s.io Sep 10 00:43:01.371920 containerd[1463]: time="2025-09-10T00:43:01.371908117Z" level=warning msg="cleaning up after shim disconnected" id=775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7 namespace=k8s.io Sep 10 00:43:01.371920 containerd[1463]: time="2025-09-10T00:43:01.371921562Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:43:01.911510 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-775d2dac7ca8b4b1e9ab3217aec5672a46561ec5ddf2be7256d4ba6f0591e7d7-rootfs.mount: Deactivated successfully. Sep 10 00:43:02.244670 kubelet[2552]: E0910 00:43:02.244498 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:02.249827 containerd[1463]: time="2025-09-10T00:43:02.249779379Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 10 00:43:02.272820 containerd[1463]: time="2025-09-10T00:43:02.272767590Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319\"" Sep 10 00:43:02.273335 containerd[1463]: time="2025-09-10T00:43:02.273305633Z" level=info msg="StartContainer for \"51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319\"" Sep 10 00:43:02.309445 systemd[1]: Started cri-containerd-51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319.scope - libcontainer container 51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319. Sep 10 00:43:02.342185 containerd[1463]: time="2025-09-10T00:43:02.342134570Z" level=info msg="StartContainer for \"51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319\" returns successfully" Sep 10 00:43:02.344032 systemd[1]: cri-containerd-51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319.scope: Deactivated successfully. Sep 10 00:43:02.374628 containerd[1463]: time="2025-09-10T00:43:02.374559498Z" level=info msg="shim disconnected" id=51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319 namespace=k8s.io Sep 10 00:43:02.374628 containerd[1463]: time="2025-09-10T00:43:02.374620223Z" level=warning msg="cleaning up after shim disconnected" id=51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319 namespace=k8s.io Sep 10 00:43:02.374628 containerd[1463]: time="2025-09-10T00:43:02.374631905Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:43:02.753657 kubelet[2552]: E0910 00:43:02.753604 2552 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 10 00:43:02.911627 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-51ac183b8c9f8d19e3193e7d3e4860a20723d322d928fda26a3ad97ecb272319-rootfs.mount: Deactivated successfully. Sep 10 00:43:03.247968 kubelet[2552]: E0910 00:43:03.247916 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:03.252756 containerd[1463]: time="2025-09-10T00:43:03.252709747Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 10 00:43:03.299146 containerd[1463]: time="2025-09-10T00:43:03.299100272Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9\"" Sep 10 00:43:03.299651 containerd[1463]: time="2025-09-10T00:43:03.299621692Z" level=info msg="StartContainer for \"50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9\"" Sep 10 00:43:03.332467 systemd[1]: Started cri-containerd-50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9.scope - libcontainer container 50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9. Sep 10 00:43:03.360328 systemd[1]: cri-containerd-50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9.scope: Deactivated successfully. Sep 10 00:43:03.362294 containerd[1463]: time="2025-09-10T00:43:03.362216858Z" level=info msg="StartContainer for \"50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9\" returns successfully" Sep 10 00:43:03.389572 containerd[1463]: time="2025-09-10T00:43:03.389502382Z" level=info msg="shim disconnected" id=50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9 namespace=k8s.io Sep 10 00:43:03.389572 containerd[1463]: time="2025-09-10T00:43:03.389563729Z" level=warning msg="cleaning up after shim disconnected" id=50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9 namespace=k8s.io Sep 10 00:43:03.389572 containerd[1463]: time="2025-09-10T00:43:03.389574719Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:43:03.912167 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-50f8d26db2a8183889f2f97b7e0261efb588fb481ed7d70a515204406da3a9a9-rootfs.mount: Deactivated successfully. Sep 10 00:43:04.255450 kubelet[2552]: E0910 00:43:04.255306 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:04.261622 containerd[1463]: time="2025-09-10T00:43:04.261548122Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 10 00:43:04.286426 containerd[1463]: time="2025-09-10T00:43:04.286375482Z" level=info msg="CreateContainer within sandbox \"9ff2b52d24b70d67e6f71702ffa0d006479e148a53f53590429636a2b42b9efc\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb\"" Sep 10 00:43:04.287325 containerd[1463]: time="2025-09-10T00:43:04.287295910Z" level=info msg="StartContainer for \"b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb\"" Sep 10 00:43:04.327515 systemd[1]: Started cri-containerd-b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb.scope - libcontainer container b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb. Sep 10 00:43:04.363463 containerd[1463]: time="2025-09-10T00:43:04.363399011Z" level=info msg="StartContainer for \"b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb\" returns successfully" Sep 10 00:43:04.700260 kubelet[2552]: E0910 00:43:04.700191 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:04.797280 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 10 00:43:04.911639 systemd[1]: run-containerd-runc-k8s.io-b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb-runc.8GltIz.mount: Deactivated successfully. Sep 10 00:43:05.259878 kubelet[2552]: E0910 00:43:05.259827 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:05.273535 kubelet[2552]: I0910 00:43:05.273441 2552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-jsjnz" podStartSLOduration=6.27342342 podStartE2EDuration="6.27342342s" podCreationTimestamp="2025-09-10 00:42:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:43:05.273274326 +0000 UTC m=+87.692321658" watchObservedRunningTime="2025-09-10 00:43:05.27342342 +0000 UTC m=+87.692470751" Sep 10 00:43:06.262429 kubelet[2552]: E0910 00:43:06.262361 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:07.264347 kubelet[2552]: E0910 00:43:07.264295 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:08.010921 systemd-networkd[1391]: lxc_health: Link UP Sep 10 00:43:08.018648 systemd-networkd[1391]: lxc_health: Gained carrier Sep 10 00:43:08.265895 kubelet[2552]: E0910 00:43:08.265702 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:08.554696 systemd[1]: run-containerd-runc-k8s.io-b861fdcf2d8718acda58528370c24b599d215b6cc783b9af2f51393062782fbb-runc.d0Rtor.mount: Deactivated successfully. Sep 10 00:43:09.268673 kubelet[2552]: E0910 00:43:09.268210 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:09.686339 systemd-networkd[1391]: lxc_health: Gained IPv6LL Sep 10 00:43:10.270270 kubelet[2552]: E0910 00:43:10.270216 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:13.700809 kubelet[2552]: E0910 00:43:13.700737 2552 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:43:14.972450 sshd[4381]: pam_unix(sshd:session): session closed for user core Sep 10 00:43:14.976538 systemd[1]: sshd@26-10.0.0.97:22-10.0.0.1:38282.service: Deactivated successfully. Sep 10 00:43:14.978483 systemd[1]: session-27.scope: Deactivated successfully. Sep 10 00:43:14.979130 systemd-logind[1448]: Session 27 logged out. Waiting for processes to exit. Sep 10 00:43:14.980183 systemd-logind[1448]: Removed session 27.