Sep 10 00:32:17.905121 kernel: Linux version 6.6.104-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Sep 9 22:56:44 -00 2025 Sep 10 00:32:17.905144 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a3dbdfb50e14c8de85dda26f853cdd6055239b4b8b15c08fb0eb00b67ce87a58 Sep 10 00:32:17.905156 kernel: BIOS-provided physical RAM map: Sep 10 00:32:17.905162 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 10 00:32:17.905168 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 10 00:32:17.905175 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 10 00:32:17.905184 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Sep 10 00:32:17.905191 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Sep 10 00:32:17.905198 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Sep 10 00:32:17.905209 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Sep 10 00:32:17.905215 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 10 00:32:17.905221 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 10 00:32:17.905231 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 10 00:32:17.905238 kernel: NX (Execute Disable) protection: active Sep 10 00:32:17.905245 kernel: APIC: Static calls initialized Sep 10 00:32:17.905258 kernel: SMBIOS 2.8 present. Sep 10 00:32:17.905265 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Sep 10 00:32:17.905279 kernel: Hypervisor detected: KVM Sep 10 00:32:17.905289 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 10 00:32:17.905303 kernel: kvm-clock: using sched offset of 2721447127 cycles Sep 10 00:32:17.905317 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 10 00:32:17.905331 kernel: tsc: Detected 2794.748 MHz processor Sep 10 00:32:17.905345 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 10 00:32:17.905356 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 10 00:32:17.905370 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Sep 10 00:32:17.905404 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Sep 10 00:32:17.905418 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 10 00:32:17.905428 kernel: Using GB pages for direct mapping Sep 10 00:32:17.905442 kernel: ACPI: Early table checksum verification disabled Sep 10 00:32:17.905455 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Sep 10 00:32:17.905465 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905479 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905486 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905496 kernel: ACPI: FACS 0x000000009CFE0000 000040 Sep 10 00:32:17.905503 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905509 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905516 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905523 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 10 00:32:17.905530 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Sep 10 00:32:17.905537 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Sep 10 00:32:17.905548 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Sep 10 00:32:17.905558 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Sep 10 00:32:17.905565 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Sep 10 00:32:17.905572 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Sep 10 00:32:17.905579 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Sep 10 00:32:17.905590 kernel: No NUMA configuration found Sep 10 00:32:17.905597 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Sep 10 00:32:17.905604 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Sep 10 00:32:17.905614 kernel: Zone ranges: Sep 10 00:32:17.905622 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 10 00:32:17.905629 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Sep 10 00:32:17.905636 kernel: Normal empty Sep 10 00:32:17.905643 kernel: Movable zone start for each node Sep 10 00:32:17.905650 kernel: Early memory node ranges Sep 10 00:32:17.905657 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 10 00:32:17.905664 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Sep 10 00:32:17.905672 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Sep 10 00:32:17.905690 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 10 00:32:17.905699 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 10 00:32:17.905707 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Sep 10 00:32:17.905714 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 10 00:32:17.905721 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 10 00:32:17.905729 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 10 00:32:17.905736 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 10 00:32:17.905744 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 10 00:32:17.905751 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 10 00:32:17.905770 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 10 00:32:17.905780 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 10 00:32:17.905787 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 10 00:32:17.905803 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 10 00:32:17.905811 kernel: TSC deadline timer available Sep 10 00:32:17.905818 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Sep 10 00:32:17.905825 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 10 00:32:17.905832 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 10 00:32:17.905842 kernel: kvm-guest: setup PV sched yield Sep 10 00:32:17.905853 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Sep 10 00:32:17.905861 kernel: Booting paravirtualized kernel on KVM Sep 10 00:32:17.905868 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 10 00:32:17.905875 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 10 00:32:17.905883 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u524288 Sep 10 00:32:17.905890 kernel: pcpu-alloc: s197160 r8192 d32216 u524288 alloc=1*2097152 Sep 10 00:32:17.905897 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 10 00:32:17.905904 kernel: kvm-guest: PV spinlocks enabled Sep 10 00:32:17.905911 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 10 00:32:17.905922 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a3dbdfb50e14c8de85dda26f853cdd6055239b4b8b15c08fb0eb00b67ce87a58 Sep 10 00:32:17.905930 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 10 00:32:17.905937 kernel: random: crng init done Sep 10 00:32:17.905945 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 10 00:32:17.905952 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 10 00:32:17.905959 kernel: Fallback order for Node 0: 0 Sep 10 00:32:17.905966 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Sep 10 00:32:17.905974 kernel: Policy zone: DMA32 Sep 10 00:32:17.905984 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 10 00:32:17.905991 kernel: Memory: 2434592K/2571752K available (12288K kernel code, 2293K rwdata, 22744K rodata, 42884K init, 2312K bss, 136900K reserved, 0K cma-reserved) Sep 10 00:32:17.905999 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 10 00:32:17.906007 kernel: ftrace: allocating 37969 entries in 149 pages Sep 10 00:32:17.906017 kernel: ftrace: allocated 149 pages with 4 groups Sep 10 00:32:17.906025 kernel: Dynamic Preempt: voluntary Sep 10 00:32:17.906034 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 10 00:32:17.906043 kernel: rcu: RCU event tracing is enabled. Sep 10 00:32:17.906052 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 10 00:32:17.906064 kernel: Trampoline variant of Tasks RCU enabled. Sep 10 00:32:17.906073 kernel: Rude variant of Tasks RCU enabled. Sep 10 00:32:17.906081 kernel: Tracing variant of Tasks RCU enabled. Sep 10 00:32:17.906090 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 10 00:32:17.906101 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 10 00:32:17.906110 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 10 00:32:17.906120 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 10 00:32:17.906130 kernel: Console: colour VGA+ 80x25 Sep 10 00:32:17.906139 kernel: printk: console [ttyS0] enabled Sep 10 00:32:17.906149 kernel: ACPI: Core revision 20230628 Sep 10 00:32:17.906164 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 10 00:32:17.906174 kernel: APIC: Switch to symmetric I/O mode setup Sep 10 00:32:17.906183 kernel: x2apic enabled Sep 10 00:32:17.906193 kernel: APIC: Switched APIC routing to: physical x2apic Sep 10 00:32:17.906203 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 10 00:32:17.906212 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 10 00:32:17.906219 kernel: kvm-guest: setup PV IPIs Sep 10 00:32:17.906239 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 10 00:32:17.906246 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Sep 10 00:32:17.906254 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Sep 10 00:32:17.906261 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 10 00:32:17.906272 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 10 00:32:17.906280 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 10 00:32:17.906287 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 10 00:32:17.906295 kernel: Spectre V2 : Mitigation: Retpolines Sep 10 00:32:17.906302 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 10 00:32:17.906313 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 10 00:32:17.906320 kernel: active return thunk: retbleed_return_thunk Sep 10 00:32:17.906331 kernel: RETBleed: Mitigation: untrained return thunk Sep 10 00:32:17.906339 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 10 00:32:17.906346 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 10 00:32:17.906354 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 10 00:32:17.906362 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 10 00:32:17.906370 kernel: active return thunk: srso_return_thunk Sep 10 00:32:17.906380 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 10 00:32:17.906401 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 10 00:32:17.906408 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 10 00:32:17.906416 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 10 00:32:17.906423 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 10 00:32:17.906431 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 10 00:32:17.906438 kernel: Freeing SMP alternatives memory: 32K Sep 10 00:32:17.906446 kernel: pid_max: default: 32768 minimum: 301 Sep 10 00:32:17.906453 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 10 00:32:17.906464 kernel: landlock: Up and running. Sep 10 00:32:17.906472 kernel: SELinux: Initializing. Sep 10 00:32:17.906479 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 10 00:32:17.906487 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 10 00:32:17.906495 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 10 00:32:17.906502 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 10 00:32:17.906510 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 10 00:32:17.906518 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 10 00:32:17.906528 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 10 00:32:17.906539 kernel: ... version: 0 Sep 10 00:32:17.906546 kernel: ... bit width: 48 Sep 10 00:32:17.906554 kernel: ... generic registers: 6 Sep 10 00:32:17.906561 kernel: ... value mask: 0000ffffffffffff Sep 10 00:32:17.906569 kernel: ... max period: 00007fffffffffff Sep 10 00:32:17.906576 kernel: ... fixed-purpose events: 0 Sep 10 00:32:17.906584 kernel: ... event mask: 000000000000003f Sep 10 00:32:17.906591 kernel: signal: max sigframe size: 1776 Sep 10 00:32:17.906598 kernel: rcu: Hierarchical SRCU implementation. Sep 10 00:32:17.906609 kernel: rcu: Max phase no-delay instances is 400. Sep 10 00:32:17.906617 kernel: smp: Bringing up secondary CPUs ... Sep 10 00:32:17.906624 kernel: smpboot: x86: Booting SMP configuration: Sep 10 00:32:17.906631 kernel: .... node #0, CPUs: #1 #2 #3 Sep 10 00:32:17.906639 kernel: smp: Brought up 1 node, 4 CPUs Sep 10 00:32:17.906646 kernel: smpboot: Max logical packages: 1 Sep 10 00:32:17.906654 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Sep 10 00:32:17.906661 kernel: devtmpfs: initialized Sep 10 00:32:17.906669 kernel: x86/mm: Memory block size: 128MB Sep 10 00:32:17.906679 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 10 00:32:17.906701 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 10 00:32:17.906710 kernel: pinctrl core: initialized pinctrl subsystem Sep 10 00:32:17.906719 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 10 00:32:17.906728 kernel: audit: initializing netlink subsys (disabled) Sep 10 00:32:17.906737 kernel: audit: type=2000 audit(1757464337.061:1): state=initialized audit_enabled=0 res=1 Sep 10 00:32:17.906747 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 10 00:32:17.906758 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 10 00:32:17.906768 kernel: cpuidle: using governor menu Sep 10 00:32:17.906782 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 10 00:32:17.906793 kernel: dca service started, version 1.12.1 Sep 10 00:32:17.906803 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Sep 10 00:32:17.906814 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Sep 10 00:32:17.906824 kernel: PCI: Using configuration type 1 for base access Sep 10 00:32:17.906835 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 10 00:32:17.906845 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 10 00:32:17.906856 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 10 00:32:17.906866 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 10 00:32:17.906881 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 10 00:32:17.906891 kernel: ACPI: Added _OSI(Module Device) Sep 10 00:32:17.906902 kernel: ACPI: Added _OSI(Processor Device) Sep 10 00:32:17.906912 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 10 00:32:17.906923 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 10 00:32:17.906933 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 10 00:32:17.906943 kernel: ACPI: Interpreter enabled Sep 10 00:32:17.906953 kernel: ACPI: PM: (supports S0 S3 S5) Sep 10 00:32:17.906962 kernel: ACPI: Using IOAPIC for interrupt routing Sep 10 00:32:17.906973 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 10 00:32:17.906981 kernel: PCI: Using E820 reservations for host bridge windows Sep 10 00:32:17.906990 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 10 00:32:17.907000 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 10 00:32:17.907361 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 10 00:32:17.907534 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 10 00:32:17.907663 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 10 00:32:17.907673 kernel: PCI host bridge to bus 0000:00 Sep 10 00:32:17.907829 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 10 00:32:17.907948 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 10 00:32:17.908099 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 10 00:32:17.908253 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Sep 10 00:32:17.908370 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Sep 10 00:32:17.908512 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Sep 10 00:32:17.908641 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 10 00:32:17.908808 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Sep 10 00:32:17.908953 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Sep 10 00:32:17.909088 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Sep 10 00:32:17.909233 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Sep 10 00:32:17.909413 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Sep 10 00:32:17.909554 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 10 00:32:17.909726 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Sep 10 00:32:17.909854 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Sep 10 00:32:17.909980 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Sep 10 00:32:17.910106 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Sep 10 00:32:17.910248 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Sep 10 00:32:17.910523 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Sep 10 00:32:17.910654 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Sep 10 00:32:17.910803 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Sep 10 00:32:17.910952 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Sep 10 00:32:17.911082 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Sep 10 00:32:17.911248 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Sep 10 00:32:17.911509 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Sep 10 00:32:17.911639 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Sep 10 00:32:17.911809 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Sep 10 00:32:17.911963 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 10 00:32:17.912114 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Sep 10 00:32:17.912260 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Sep 10 00:32:17.912484 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Sep 10 00:32:17.912663 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Sep 10 00:32:17.912801 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Sep 10 00:32:17.912818 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 10 00:32:17.912826 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 10 00:32:17.912834 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 10 00:32:17.912841 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 10 00:32:17.912849 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 10 00:32:17.912856 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 10 00:32:17.912864 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 10 00:32:17.912872 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 10 00:32:17.912879 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 10 00:32:17.912890 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 10 00:32:17.912897 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 10 00:32:17.912905 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 10 00:32:17.912913 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 10 00:32:17.912920 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 10 00:32:17.912927 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 10 00:32:17.912935 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 10 00:32:17.912942 kernel: iommu: Default domain type: Translated Sep 10 00:32:17.912950 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 10 00:32:17.912960 kernel: PCI: Using ACPI for IRQ routing Sep 10 00:32:17.912968 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 10 00:32:17.912975 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 10 00:32:17.912983 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Sep 10 00:32:17.913106 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 10 00:32:17.913230 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 10 00:32:17.913418 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 10 00:32:17.913437 kernel: vgaarb: loaded Sep 10 00:32:17.913456 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 10 00:32:17.913467 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 10 00:32:17.913478 kernel: clocksource: Switched to clocksource kvm-clock Sep 10 00:32:17.913488 kernel: VFS: Disk quotas dquot_6.6.0 Sep 10 00:32:17.913496 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 10 00:32:17.913504 kernel: pnp: PnP ACPI init Sep 10 00:32:17.913704 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Sep 10 00:32:17.913723 kernel: pnp: PnP ACPI: found 6 devices Sep 10 00:32:17.913735 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 10 00:32:17.913752 kernel: NET: Registered PF_INET protocol family Sep 10 00:32:17.913762 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 10 00:32:17.913773 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 10 00:32:17.913783 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 10 00:32:17.913791 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 10 00:32:17.913799 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 10 00:32:17.913807 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 10 00:32:17.913815 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 10 00:32:17.913826 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 10 00:32:17.913834 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 10 00:32:17.913842 kernel: NET: Registered PF_XDP protocol family Sep 10 00:32:17.913978 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 10 00:32:17.914116 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 10 00:32:17.914238 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 10 00:32:17.914358 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Sep 10 00:32:17.914570 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Sep 10 00:32:17.914736 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Sep 10 00:32:17.914758 kernel: PCI: CLS 0 bytes, default 64 Sep 10 00:32:17.914769 kernel: Initialise system trusted keyrings Sep 10 00:32:17.914780 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 10 00:32:17.914788 kernel: Key type asymmetric registered Sep 10 00:32:17.914796 kernel: Asymmetric key parser 'x509' registered Sep 10 00:32:17.914803 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 10 00:32:17.914811 kernel: io scheduler mq-deadline registered Sep 10 00:32:17.914818 kernel: io scheduler kyber registered Sep 10 00:32:17.914826 kernel: io scheduler bfq registered Sep 10 00:32:17.914837 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 10 00:32:17.914845 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 10 00:32:17.914852 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 10 00:32:17.914860 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 10 00:32:17.914867 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 10 00:32:17.914875 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 10 00:32:17.914883 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 10 00:32:17.914890 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 10 00:32:17.914898 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 10 00:32:17.915055 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 10 00:32:17.915067 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 10 00:32:17.915189 kernel: rtc_cmos 00:04: registered as rtc0 Sep 10 00:32:17.915308 kernel: rtc_cmos 00:04: setting system clock to 2025-09-10T00:32:17 UTC (1757464337) Sep 10 00:32:17.915456 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Sep 10 00:32:17.915468 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 10 00:32:17.915476 kernel: NET: Registered PF_INET6 protocol family Sep 10 00:32:17.915484 kernel: Segment Routing with IPv6 Sep 10 00:32:17.915497 kernel: In-situ OAM (IOAM) with IPv6 Sep 10 00:32:17.915505 kernel: NET: Registered PF_PACKET protocol family Sep 10 00:32:17.915512 kernel: Key type dns_resolver registered Sep 10 00:32:17.915520 kernel: IPI shorthand broadcast: enabled Sep 10 00:32:17.915528 kernel: sched_clock: Marking stable (773002819, 102139128)->(892685943, -17543996) Sep 10 00:32:17.915536 kernel: registered taskstats version 1 Sep 10 00:32:17.915544 kernel: Loading compiled-in X.509 certificates Sep 10 00:32:17.915552 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.104-flatcar: a614f1c62f27a560d677bbf0283703118c9005ec' Sep 10 00:32:17.915560 kernel: Key type .fscrypt registered Sep 10 00:32:17.915570 kernel: Key type fscrypt-provisioning registered Sep 10 00:32:17.915578 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 10 00:32:17.915586 kernel: ima: Allocated hash algorithm: sha1 Sep 10 00:32:17.915593 kernel: ima: No architecture policies found Sep 10 00:32:17.915601 kernel: clk: Disabling unused clocks Sep 10 00:32:17.915609 kernel: Freeing unused kernel image (initmem) memory: 42884K Sep 10 00:32:17.915617 kernel: Write protecting the kernel read-only data: 36864k Sep 10 00:32:17.915624 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Sep 10 00:32:17.915635 kernel: Run /init as init process Sep 10 00:32:17.915642 kernel: with arguments: Sep 10 00:32:17.915650 kernel: /init Sep 10 00:32:17.915657 kernel: with environment: Sep 10 00:32:17.915665 kernel: HOME=/ Sep 10 00:32:17.915672 kernel: TERM=linux Sep 10 00:32:17.915679 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 10 00:32:17.915697 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 10 00:32:17.915710 systemd[1]: Detected virtualization kvm. Sep 10 00:32:17.915719 systemd[1]: Detected architecture x86-64. Sep 10 00:32:17.915727 systemd[1]: Running in initrd. Sep 10 00:32:17.915735 systemd[1]: No hostname configured, using default hostname. Sep 10 00:32:17.915744 systemd[1]: Hostname set to . Sep 10 00:32:17.915753 systemd[1]: Initializing machine ID from VM UUID. Sep 10 00:32:17.915761 systemd[1]: Queued start job for default target initrd.target. Sep 10 00:32:17.915769 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 10 00:32:17.915781 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 10 00:32:17.915790 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 10 00:32:17.915813 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 10 00:32:17.915824 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 10 00:32:17.915833 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 10 00:32:17.915846 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 10 00:32:17.915855 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 10 00:32:17.915863 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 10 00:32:17.915872 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 10 00:32:17.915880 systemd[1]: Reached target paths.target - Path Units. Sep 10 00:32:17.915889 systemd[1]: Reached target slices.target - Slice Units. Sep 10 00:32:17.915897 systemd[1]: Reached target swap.target - Swaps. Sep 10 00:32:17.915906 systemd[1]: Reached target timers.target - Timer Units. Sep 10 00:32:17.915917 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 10 00:32:17.915926 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 10 00:32:17.915934 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 10 00:32:17.915943 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 10 00:32:17.915952 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 10 00:32:17.915960 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 10 00:32:17.915969 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 10 00:32:17.915977 systemd[1]: Reached target sockets.target - Socket Units. Sep 10 00:32:17.915986 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 10 00:32:17.915997 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 10 00:32:17.916006 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 10 00:32:17.916014 systemd[1]: Starting systemd-fsck-usr.service... Sep 10 00:32:17.916023 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 10 00:32:17.916033 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 10 00:32:17.916044 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:32:17.916055 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 10 00:32:17.916066 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 10 00:32:17.916077 systemd[1]: Finished systemd-fsck-usr.service. Sep 10 00:32:17.916094 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 10 00:32:17.916133 systemd-journald[193]: Collecting audit messages is disabled. Sep 10 00:32:17.916172 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 10 00:32:17.916185 systemd-journald[193]: Journal started Sep 10 00:32:17.916216 systemd-journald[193]: Runtime Journal (/run/log/journal/a9167803e28142719c6011975151400a) is 6.0M, max 48.4M, 42.3M free. Sep 10 00:32:17.905824 systemd-modules-load[194]: Inserted module 'overlay' Sep 10 00:32:17.948467 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 10 00:32:17.948495 kernel: Bridge firewalling registered Sep 10 00:32:17.934973 systemd-modules-load[194]: Inserted module 'br_netfilter' Sep 10 00:32:17.951550 systemd[1]: Started systemd-journald.service - Journal Service. Sep 10 00:32:17.952117 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 10 00:32:17.954740 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:32:17.970627 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 10 00:32:17.973821 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:32:17.976429 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 10 00:32:17.979572 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 10 00:32:17.989674 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 10 00:32:17.991975 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:32:17.992649 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:32:18.002525 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 10 00:32:18.003722 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 10 00:32:18.007639 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 10 00:32:18.016369 dracut-cmdline[227]: dracut-dracut-053 Sep 10 00:32:18.019495 dracut-cmdline[227]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a3dbdfb50e14c8de85dda26f853cdd6055239b4b8b15c08fb0eb00b67ce87a58 Sep 10 00:32:18.044955 systemd-resolved[231]: Positive Trust Anchors: Sep 10 00:32:18.044976 systemd-resolved[231]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 10 00:32:18.045008 systemd-resolved[231]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 10 00:32:18.047784 systemd-resolved[231]: Defaulting to hostname 'linux'. Sep 10 00:32:18.049071 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 10 00:32:18.054660 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 10 00:32:18.109440 kernel: SCSI subsystem initialized Sep 10 00:32:18.118411 kernel: Loading iSCSI transport class v2.0-870. Sep 10 00:32:18.129427 kernel: iscsi: registered transport (tcp) Sep 10 00:32:18.150432 kernel: iscsi: registered transport (qla4xxx) Sep 10 00:32:18.150503 kernel: QLogic iSCSI HBA Driver Sep 10 00:32:18.206852 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 10 00:32:18.216521 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 10 00:32:18.244199 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 10 00:32:18.244291 kernel: device-mapper: uevent: version 1.0.3 Sep 10 00:32:18.244304 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 10 00:32:18.287432 kernel: raid6: avx2x4 gen() 30580 MB/s Sep 10 00:32:18.326431 kernel: raid6: avx2x2 gen() 31391 MB/s Sep 10 00:32:18.343506 kernel: raid6: avx2x1 gen() 25409 MB/s Sep 10 00:32:18.343605 kernel: raid6: using algorithm avx2x2 gen() 31391 MB/s Sep 10 00:32:18.361493 kernel: raid6: .... xor() 19578 MB/s, rmw enabled Sep 10 00:32:18.361601 kernel: raid6: using avx2x2 recovery algorithm Sep 10 00:32:18.384447 kernel: xor: automatically using best checksumming function avx Sep 10 00:32:18.545434 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 10 00:32:18.560976 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 10 00:32:18.568673 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 10 00:32:18.581784 systemd-udevd[414]: Using default interface naming scheme 'v255'. Sep 10 00:32:18.586572 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 10 00:32:18.596538 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 10 00:32:18.611656 dracut-pre-trigger[424]: rd.md=0: removing MD RAID activation Sep 10 00:32:18.648637 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 10 00:32:18.661552 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 10 00:32:18.728053 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 10 00:32:18.741134 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 10 00:32:18.752832 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 10 00:32:18.755523 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 10 00:32:18.757002 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 10 00:32:18.758216 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 10 00:32:18.769537 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 10 00:32:18.773655 kernel: cryptd: max_cpu_qlen set to 1000 Sep 10 00:32:18.782583 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 10 00:32:18.785431 kernel: AVX2 version of gcm_enc/dec engaged. Sep 10 00:32:18.785456 kernel: AES CTR mode by8 optimization enabled Sep 10 00:32:18.787292 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 10 00:32:18.794544 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 10 00:32:18.797531 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 10 00:32:18.797554 kernel: GPT:9289727 != 19775487 Sep 10 00:32:18.797564 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 10 00:32:18.798634 kernel: GPT:9289727 != 19775487 Sep 10 00:32:18.798670 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 10 00:32:18.799635 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:32:18.803460 kernel: libata version 3.00 loaded. Sep 10 00:32:18.811692 kernel: ahci 0000:00:1f.2: version 3.0 Sep 10 00:32:18.811892 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 10 00:32:18.812280 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 10 00:32:18.819870 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Sep 10 00:32:18.820077 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 10 00:32:18.813783 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:32:18.814650 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 10 00:32:18.814940 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 10 00:32:18.815055 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:32:18.815283 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:32:18.830695 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:32:18.837414 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (463) Sep 10 00:32:18.842416 kernel: BTRFS: device fsid 47ffa5df-7ab2-4f1a-b68f-595717991426 devid 1 transid 34 /dev/vda3 scanned by (udev-worker) (475) Sep 10 00:32:18.845409 kernel: scsi host0: ahci Sep 10 00:32:18.853178 kernel: scsi host1: ahci Sep 10 00:32:18.853443 kernel: scsi host2: ahci Sep 10 00:32:18.853608 kernel: scsi host3: ahci Sep 10 00:32:18.853768 kernel: scsi host4: ahci Sep 10 00:32:18.858566 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 10 00:32:18.918047 kernel: scsi host5: ahci Sep 10 00:32:18.918259 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Sep 10 00:32:18.918272 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Sep 10 00:32:18.918293 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Sep 10 00:32:18.918303 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Sep 10 00:32:18.918314 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Sep 10 00:32:18.918324 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Sep 10 00:32:18.891992 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 10 00:32:18.922721 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:32:18.928442 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 10 00:32:18.929663 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 10 00:32:18.936650 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 10 00:32:18.948750 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 10 00:32:18.951959 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 10 00:32:18.958694 disk-uuid[567]: Primary Header is updated. Sep 10 00:32:18.958694 disk-uuid[567]: Secondary Entries is updated. Sep 10 00:32:18.958694 disk-uuid[567]: Secondary Header is updated. Sep 10 00:32:18.962841 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:32:18.966402 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:32:18.974709 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:32:19.211354 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 10 00:32:19.211479 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 10 00:32:19.211492 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 10 00:32:19.211503 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 10 00:32:19.212420 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 10 00:32:19.213420 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 10 00:32:19.213440 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 10 00:32:19.214510 kernel: ata3.00: applying bridge limits Sep 10 00:32:19.215409 kernel: ata3.00: configured for UDMA/100 Sep 10 00:32:19.216412 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 10 00:32:19.256444 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 10 00:32:19.256863 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 10 00:32:19.270428 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 10 00:32:19.968307 disk-uuid[570]: The operation has completed successfully. Sep 10 00:32:19.969522 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 10 00:32:20.005749 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 10 00:32:20.005877 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 10 00:32:20.021513 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 10 00:32:20.025405 sh[593]: Success Sep 10 00:32:20.039433 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Sep 10 00:32:20.073681 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 10 00:32:20.090175 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 10 00:32:20.092882 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 10 00:32:20.105179 kernel: BTRFS info (device dm-0): first mount of filesystem 47ffa5df-7ab2-4f1a-b68f-595717991426 Sep 10 00:32:20.105212 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:32:20.105223 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 10 00:32:20.106140 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 10 00:32:20.106842 kernel: BTRFS info (device dm-0): using free space tree Sep 10 00:32:20.111294 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 10 00:32:20.112940 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 10 00:32:20.120530 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 10 00:32:20.122312 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 10 00:32:20.131417 kernel: BTRFS info (device vda6): first mount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:32:20.131449 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:32:20.131460 kernel: BTRFS info (device vda6): using free space tree Sep 10 00:32:20.134400 kernel: BTRFS info (device vda6): auto enabling async discard Sep 10 00:32:20.145320 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 10 00:32:20.146946 kernel: BTRFS info (device vda6): last unmount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:32:20.165452 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 10 00:32:20.172570 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 10 00:32:20.361819 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 10 00:32:20.368939 ignition[695]: Ignition 2.19.0 Sep 10 00:32:20.368957 ignition[695]: Stage: fetch-offline Sep 10 00:32:20.373612 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 10 00:32:20.369041 ignition[695]: no configs at "/usr/lib/ignition/base.d" Sep 10 00:32:20.369056 ignition[695]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:32:20.369171 ignition[695]: parsed url from cmdline: "" Sep 10 00:32:20.369176 ignition[695]: no config URL provided Sep 10 00:32:20.369183 ignition[695]: reading system config file "/usr/lib/ignition/user.ign" Sep 10 00:32:20.369195 ignition[695]: no config at "/usr/lib/ignition/user.ign" Sep 10 00:32:20.369235 ignition[695]: op(1): [started] loading QEMU firmware config module Sep 10 00:32:20.369241 ignition[695]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 10 00:32:20.378414 ignition[695]: op(1): [finished] loading QEMU firmware config module Sep 10 00:32:20.414770 systemd-networkd[780]: lo: Link UP Sep 10 00:32:20.414779 systemd-networkd[780]: lo: Gained carrier Sep 10 00:32:20.417870 systemd-networkd[780]: Enumeration completed Sep 10 00:32:20.418810 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 10 00:32:20.420136 systemd[1]: Reached target network.target - Network. Sep 10 00:32:20.423247 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:32:20.423256 systemd-networkd[780]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 10 00:32:20.427235 ignition[695]: parsing config with SHA512: 7ddf5ec1f7cfd19936b4ff9ae2f919fb6df566033b3572faee19d9a4e6dedb330f591e1075f62bc32e1f78fe8ab6b81038b462438946a7d7b69e1146461a12ae Sep 10 00:32:20.428267 systemd-networkd[780]: eth0: Link UP Sep 10 00:32:20.428275 systemd-networkd[780]: eth0: Gained carrier Sep 10 00:32:20.428284 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:32:20.435129 unknown[695]: fetched base config from "system" Sep 10 00:32:20.435140 unknown[695]: fetched user config from "qemu" Sep 10 00:32:20.436978 ignition[695]: fetch-offline: fetch-offline passed Sep 10 00:32:20.437782 ignition[695]: Ignition finished successfully Sep 10 00:32:20.440625 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 10 00:32:20.441456 systemd-networkd[780]: eth0: DHCPv4 address 10.0.0.29/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 10 00:32:20.442968 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 10 00:32:20.447586 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 10 00:32:20.463501 ignition[786]: Ignition 2.19.0 Sep 10 00:32:20.463514 ignition[786]: Stage: kargs Sep 10 00:32:20.463715 ignition[786]: no configs at "/usr/lib/ignition/base.d" Sep 10 00:32:20.463727 ignition[786]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:32:20.464648 ignition[786]: kargs: kargs passed Sep 10 00:32:20.464697 ignition[786]: Ignition finished successfully Sep 10 00:32:20.470578 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 10 00:32:20.480564 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 10 00:32:20.494624 ignition[795]: Ignition 2.19.0 Sep 10 00:32:20.494635 ignition[795]: Stage: disks Sep 10 00:32:20.494787 ignition[795]: no configs at "/usr/lib/ignition/base.d" Sep 10 00:32:20.494798 ignition[795]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:32:20.495627 ignition[795]: disks: disks passed Sep 10 00:32:20.495678 ignition[795]: Ignition finished successfully Sep 10 00:32:20.500844 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 10 00:32:20.502811 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 10 00:32:20.503056 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 10 00:32:20.503358 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 10 00:32:20.507296 systemd[1]: Reached target sysinit.target - System Initialization. Sep 10 00:32:20.509040 systemd[1]: Reached target basic.target - Basic System. Sep 10 00:32:20.520652 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 10 00:32:20.531070 systemd-resolved[231]: Detected conflict on linux IN A 10.0.0.29 Sep 10 00:32:20.531085 systemd-resolved[231]: Hostname conflict, changing published hostname from 'linux' to 'linux7'. Sep 10 00:32:20.533290 systemd-fsck[805]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 10 00:32:20.539722 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 10 00:32:20.554525 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 10 00:32:20.653409 kernel: EXT4-fs (vda9): mounted filesystem 0a9bf3c7-f8cd-4d40-b949-283957ba2f96 r/w with ordered data mode. Quota mode: none. Sep 10 00:32:20.653697 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 10 00:32:20.654686 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 10 00:32:20.662499 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 10 00:32:20.664364 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 10 00:32:20.665184 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 10 00:32:20.670495 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (813) Sep 10 00:32:20.665236 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 10 00:32:20.676025 kernel: BTRFS info (device vda6): first mount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:32:20.676047 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:32:20.676062 kernel: BTRFS info (device vda6): using free space tree Sep 10 00:32:20.676076 kernel: BTRFS info (device vda6): auto enabling async discard Sep 10 00:32:20.665264 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 10 00:32:20.678113 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 10 00:32:20.695534 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 10 00:32:20.696846 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 10 00:32:20.731194 initrd-setup-root[837]: cut: /sysroot/etc/passwd: No such file or directory Sep 10 00:32:20.736265 initrd-setup-root[844]: cut: /sysroot/etc/group: No such file or directory Sep 10 00:32:20.740921 initrd-setup-root[851]: cut: /sysroot/etc/shadow: No such file or directory Sep 10 00:32:20.745380 initrd-setup-root[858]: cut: /sysroot/etc/gshadow: No such file or directory Sep 10 00:32:20.829102 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 10 00:32:20.833514 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 10 00:32:20.835426 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 10 00:32:20.845410 kernel: BTRFS info (device vda6): last unmount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:32:20.861290 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 10 00:32:21.020035 ignition[930]: INFO : Ignition 2.19.0 Sep 10 00:32:21.020035 ignition[930]: INFO : Stage: mount Sep 10 00:32:21.021858 ignition[930]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 10 00:32:21.021858 ignition[930]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:32:21.021858 ignition[930]: INFO : mount: mount passed Sep 10 00:32:21.021858 ignition[930]: INFO : Ignition finished successfully Sep 10 00:32:21.023890 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 10 00:32:21.028584 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 10 00:32:21.104672 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 10 00:32:21.116528 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 10 00:32:21.124411 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (939) Sep 10 00:32:21.126969 kernel: BTRFS info (device vda6): first mount of filesystem 81146077-6e72-4c2f-a205-63f64096a038 Sep 10 00:32:21.126991 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 10 00:32:21.127001 kernel: BTRFS info (device vda6): using free space tree Sep 10 00:32:21.129425 kernel: BTRFS info (device vda6): auto enabling async discard Sep 10 00:32:21.130900 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 10 00:32:21.158640 ignition[956]: INFO : Ignition 2.19.0 Sep 10 00:32:21.158640 ignition[956]: INFO : Stage: files Sep 10 00:32:21.160609 ignition[956]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 10 00:32:21.160609 ignition[956]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:32:21.160609 ignition[956]: DEBUG : files: compiled without relabeling support, skipping Sep 10 00:32:21.163681 ignition[956]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 10 00:32:21.163681 ignition[956]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 10 00:32:21.167945 ignition[956]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 10 00:32:21.169390 ignition[956]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 10 00:32:21.171071 unknown[956]: wrote ssh authorized keys file for user: core Sep 10 00:32:21.172377 ignition[956]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 10 00:32:21.173890 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 10 00:32:21.175976 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Sep 10 00:32:21.310507 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 10 00:32:21.677755 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 10 00:32:21.679983 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 10 00:32:21.682263 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 10 00:32:21.776756 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 10 00:32:22.122422 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 10 00:32:22.122422 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:32:22.126482 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Sep 10 00:32:22.283771 systemd-networkd[780]: eth0: Gained IPv6LL Sep 10 00:32:22.509429 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 10 00:32:23.232026 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 10 00:32:23.232026 ignition[956]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Sep 10 00:32:23.236213 ignition[956]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Sep 10 00:32:23.345596 ignition[956]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 10 00:32:23.352475 ignition[956]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 10 00:32:23.354002 ignition[956]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Sep 10 00:32:23.354002 ignition[956]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Sep 10 00:32:23.354002 ignition[956]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Sep 10 00:32:23.354002 ignition[956]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 10 00:32:23.354002 ignition[956]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 10 00:32:23.354002 ignition[956]: INFO : files: files passed Sep 10 00:32:23.354002 ignition[956]: INFO : Ignition finished successfully Sep 10 00:32:23.356137 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 10 00:32:23.365669 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 10 00:32:23.368276 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 10 00:32:23.369962 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 10 00:32:23.370084 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 10 00:32:23.378806 initrd-setup-root-after-ignition[985]: grep: /sysroot/oem/oem-release: No such file or directory Sep 10 00:32:23.381744 initrd-setup-root-after-ignition[987]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 10 00:32:23.383327 initrd-setup-root-after-ignition[987]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 10 00:32:23.384825 initrd-setup-root-after-ignition[991]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 10 00:32:23.384522 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 10 00:32:23.386702 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 10 00:32:23.395550 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 10 00:32:23.424337 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 10 00:32:23.424525 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 10 00:32:23.427503 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 10 00:32:23.428831 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 10 00:32:23.430956 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 10 00:32:23.432149 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 10 00:32:23.453168 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 10 00:32:23.465963 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 10 00:32:23.479762 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 10 00:32:23.482112 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 10 00:32:23.482905 systemd[1]: Stopped target timers.target - Timer Units. Sep 10 00:32:23.483476 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 10 00:32:23.483645 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 10 00:32:23.488439 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 10 00:32:23.489052 systemd[1]: Stopped target basic.target - Basic System. Sep 10 00:32:23.489397 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 10 00:32:23.489888 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 10 00:32:23.490204 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 10 00:32:23.490701 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 10 00:32:23.491011 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 10 00:32:23.491339 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 10 00:32:23.491875 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 10 00:32:23.492148 systemd[1]: Stopped target swap.target - Swaps. Sep 10 00:32:23.492490 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 10 00:32:23.492628 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 10 00:32:23.493268 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 10 00:32:23.493778 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 10 00:32:23.494062 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 10 00:32:23.494262 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 10 00:32:23.514895 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 10 00:32:23.515066 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 10 00:32:23.519824 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 10 00:32:23.519951 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 10 00:32:23.520442 systemd[1]: Stopped target paths.target - Path Units. Sep 10 00:32:23.520815 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 10 00:32:23.526516 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 10 00:32:23.527087 systemd[1]: Stopped target slices.target - Slice Units. Sep 10 00:32:23.527657 systemd[1]: Stopped target sockets.target - Socket Units. Sep 10 00:32:23.528001 systemd[1]: iscsid.socket: Deactivated successfully. Sep 10 00:32:23.528108 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 10 00:32:23.532651 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 10 00:32:23.532743 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 10 00:32:23.534278 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 10 00:32:23.534418 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 10 00:32:23.536066 systemd[1]: ignition-files.service: Deactivated successfully. Sep 10 00:32:23.536175 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 10 00:32:23.549613 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 10 00:32:23.551680 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 10 00:32:23.552559 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 10 00:32:23.552688 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 10 00:32:23.553213 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 10 00:32:23.553319 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 10 00:32:23.561463 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 10 00:32:23.561612 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 10 00:32:23.566294 ignition[1011]: INFO : Ignition 2.19.0 Sep 10 00:32:23.566294 ignition[1011]: INFO : Stage: umount Sep 10 00:32:23.582771 ignition[1011]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 10 00:32:23.582771 ignition[1011]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 10 00:32:23.582771 ignition[1011]: INFO : umount: umount passed Sep 10 00:32:23.582771 ignition[1011]: INFO : Ignition finished successfully Sep 10 00:32:23.580000 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 10 00:32:23.580160 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 10 00:32:23.583770 systemd[1]: Stopped target network.target - Network. Sep 10 00:32:23.586443 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 10 00:32:23.586515 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 10 00:32:23.587040 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 10 00:32:23.587102 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 10 00:32:23.587691 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 10 00:32:23.587765 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 10 00:32:23.588088 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 10 00:32:23.588155 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 10 00:32:23.602769 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 10 00:32:23.604906 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 10 00:32:23.611431 systemd-networkd[780]: eth0: DHCPv6 lease lost Sep 10 00:32:23.614864 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 10 00:32:23.615049 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 10 00:32:23.615892 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 10 00:32:23.615948 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 10 00:32:23.645690 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 10 00:32:23.646214 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 10 00:32:23.646321 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 10 00:32:23.649007 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 10 00:32:23.650780 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 10 00:32:23.651443 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 10 00:32:23.651613 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 10 00:32:23.659059 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 10 00:32:23.659195 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:32:23.660810 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 10 00:32:23.660866 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 10 00:32:23.661136 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 10 00:32:23.661183 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 10 00:32:23.667862 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 10 00:32:23.667992 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 10 00:32:23.677355 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 10 00:32:23.677582 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 10 00:32:23.678130 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 10 00:32:23.678183 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 10 00:32:23.681825 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 10 00:32:23.681869 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 10 00:32:23.682160 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 10 00:32:23.682209 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 10 00:32:23.683156 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 10 00:32:23.683206 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 10 00:32:23.684093 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 10 00:32:23.684140 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 10 00:32:23.693452 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 10 00:32:23.693954 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 10 00:32:23.694012 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 10 00:32:23.694306 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 10 00:32:23.694353 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 10 00:32:23.694847 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 10 00:32:23.694894 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 10 00:32:23.695204 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 10 00:32:23.695249 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:32:23.714402 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 10 00:32:23.714551 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 10 00:32:24.269477 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 10 00:32:24.269642 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 10 00:32:24.271988 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 10 00:32:24.273933 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 10 00:32:24.274011 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 10 00:32:24.286586 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 10 00:32:24.293818 systemd[1]: Switching root. Sep 10 00:32:24.325455 systemd-journald[193]: Journal stopped Sep 10 00:32:25.805585 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). Sep 10 00:32:25.805672 kernel: SELinux: policy capability network_peer_controls=1 Sep 10 00:32:25.805696 kernel: SELinux: policy capability open_perms=1 Sep 10 00:32:25.805708 kernel: SELinux: policy capability extended_socket_class=1 Sep 10 00:32:25.805719 kernel: SELinux: policy capability always_check_network=0 Sep 10 00:32:25.805731 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 10 00:32:25.805749 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 10 00:32:25.805761 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 10 00:32:25.805778 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 10 00:32:25.805790 kernel: audit: type=1403 audit(1757464344.857:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 10 00:32:25.805803 systemd[1]: Successfully loaded SELinux policy in 47.273ms. Sep 10 00:32:25.805826 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.945ms. Sep 10 00:32:25.805839 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 10 00:32:25.805852 systemd[1]: Detected virtualization kvm. Sep 10 00:32:25.805870 systemd[1]: Detected architecture x86-64. Sep 10 00:32:25.805886 systemd[1]: Detected first boot. Sep 10 00:32:25.805902 systemd[1]: Initializing machine ID from VM UUID. Sep 10 00:32:25.805919 zram_generator::config[1056]: No configuration found. Sep 10 00:32:25.805933 systemd[1]: Populated /etc with preset unit settings. Sep 10 00:32:25.805948 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 10 00:32:25.805965 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 10 00:32:25.805983 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 10 00:32:25.805996 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 10 00:32:25.806009 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 10 00:32:25.806021 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 10 00:32:25.806033 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 10 00:32:25.806046 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 10 00:32:25.806058 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 10 00:32:25.806074 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 10 00:32:25.806086 systemd[1]: Created slice user.slice - User and Session Slice. Sep 10 00:32:25.806098 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 10 00:32:25.806110 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 10 00:32:25.806123 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 10 00:32:25.806140 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 10 00:32:25.806158 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 10 00:32:25.806176 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 10 00:32:25.806193 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 10 00:32:25.806214 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 10 00:32:25.806233 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 10 00:32:25.806250 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 10 00:32:25.806268 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 10 00:32:25.806285 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 10 00:32:25.806302 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 10 00:32:25.806320 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 10 00:32:25.806342 systemd[1]: Reached target slices.target - Slice Units. Sep 10 00:32:25.806360 systemd[1]: Reached target swap.target - Swaps. Sep 10 00:32:25.806377 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 10 00:32:25.806414 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 10 00:32:25.806441 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 10 00:32:25.806459 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 10 00:32:25.806487 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 10 00:32:25.806504 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 10 00:32:25.806521 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 10 00:32:25.806539 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 10 00:32:25.806563 systemd[1]: Mounting media.mount - External Media Directory... Sep 10 00:32:25.806580 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:25.806597 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 10 00:32:25.806615 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 10 00:32:25.806631 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 10 00:32:25.806649 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 10 00:32:25.806666 systemd[1]: Reached target machines.target - Containers. Sep 10 00:32:25.806682 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 10 00:32:25.806704 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:32:25.806722 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 10 00:32:25.806744 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 10 00:32:25.806761 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:32:25.806778 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 10 00:32:25.806795 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:32:25.806811 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 10 00:32:25.806828 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:32:25.806914 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 10 00:32:25.806933 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 10 00:32:25.806949 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 10 00:32:25.806965 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 10 00:32:25.806982 systemd[1]: Stopped systemd-fsck-usr.service. Sep 10 00:32:25.806999 kernel: fuse: init (API version 7.39) Sep 10 00:32:25.807015 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 10 00:32:25.807032 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 10 00:32:25.807048 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 10 00:32:25.807069 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 10 00:32:25.807086 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 10 00:32:25.808401 kernel: loop: module loaded Sep 10 00:32:25.808427 systemd[1]: verity-setup.service: Deactivated successfully. Sep 10 00:32:25.808445 systemd[1]: Stopped verity-setup.service. Sep 10 00:32:25.808462 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:25.808490 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 10 00:32:25.808507 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 10 00:32:25.808526 systemd[1]: Mounted media.mount - External Media Directory. Sep 10 00:32:25.808549 kernel: ACPI: bus type drm_connector registered Sep 10 00:32:25.808565 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 10 00:32:25.808582 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 10 00:32:25.808612 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 10 00:32:25.808629 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 10 00:32:25.808650 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 10 00:32:25.808692 systemd-journald[1126]: Collecting audit messages is disabled. Sep 10 00:32:25.808724 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 10 00:32:25.808742 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:32:25.808759 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:32:25.808776 systemd-journald[1126]: Journal started Sep 10 00:32:25.808810 systemd-journald[1126]: Runtime Journal (/run/log/journal/a9167803e28142719c6011975151400a) is 6.0M, max 48.4M, 42.3M free. Sep 10 00:32:25.547051 systemd[1]: Queued start job for default target multi-user.target. Sep 10 00:32:25.566880 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 10 00:32:25.567444 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 10 00:32:25.810868 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 10 00:32:25.812971 systemd[1]: Started systemd-journald.service - Journal Service. Sep 10 00:32:25.813885 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 10 00:32:25.814079 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 10 00:32:25.815593 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:32:25.815779 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:32:25.817316 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 10 00:32:25.817519 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 10 00:32:25.818887 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:32:25.819074 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:32:25.820569 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 10 00:32:25.821990 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 10 00:32:25.823746 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 10 00:32:25.841623 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 10 00:32:25.858608 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 10 00:32:25.861289 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 10 00:32:25.862455 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 10 00:32:25.862502 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 10 00:32:25.864559 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 10 00:32:25.866951 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 10 00:32:25.870548 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 10 00:32:25.886329 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:32:25.888861 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 10 00:32:25.891191 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 10 00:32:25.892474 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 10 00:32:25.894751 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 10 00:32:25.896005 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 10 00:32:25.899175 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:32:25.901915 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 10 00:32:25.919047 systemd-journald[1126]: Time spent on flushing to /var/log/journal/a9167803e28142719c6011975151400a is 39.134ms for 957 entries. Sep 10 00:32:25.919047 systemd-journald[1126]: System Journal (/var/log/journal/a9167803e28142719c6011975151400a) is 8.0M, max 195.6M, 187.6M free. Sep 10 00:32:25.970365 systemd-journald[1126]: Received client request to flush runtime journal. Sep 10 00:32:25.970420 kernel: loop0: detected capacity change from 0 to 140768 Sep 10 00:32:25.907588 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 10 00:32:25.910871 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 10 00:32:25.914010 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 10 00:32:25.916208 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 10 00:32:25.917762 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 10 00:32:25.920780 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 10 00:32:25.956950 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 10 00:32:25.984592 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 10 00:32:25.984933 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 10 00:32:25.988612 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 10 00:32:25.991963 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 10 00:32:26.003588 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:32:26.018292 udevadm[1183]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 10 00:32:26.032579 kernel: loop1: detected capacity change from 0 to 142488 Sep 10 00:32:26.033199 systemd-tmpfiles[1171]: ACLs are not supported, ignoring. Sep 10 00:32:26.033226 systemd-tmpfiles[1171]: ACLs are not supported, ignoring. Sep 10 00:32:26.044449 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 10 00:32:26.046876 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 10 00:32:26.047567 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 10 00:32:26.057649 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 10 00:32:26.119363 kernel: loop2: detected capacity change from 0 to 229808 Sep 10 00:32:26.125728 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 10 00:32:26.135617 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 10 00:32:26.197417 kernel: loop3: detected capacity change from 0 to 140768 Sep 10 00:32:26.211529 kernel: loop4: detected capacity change from 0 to 142488 Sep 10 00:32:26.218298 systemd-tmpfiles[1194]: ACLs are not supported, ignoring. Sep 10 00:32:26.218438 systemd-tmpfiles[1194]: ACLs are not supported, ignoring. Sep 10 00:32:26.227055 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 10 00:32:26.249713 kernel: loop5: detected capacity change from 0 to 229808 Sep 10 00:32:26.258031 (sd-merge)[1196]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 10 00:32:26.258788 (sd-merge)[1196]: Merged extensions into '/usr'. Sep 10 00:32:26.263477 systemd[1]: Reloading requested from client PID 1170 ('systemd-sysext') (unit systemd-sysext.service)... Sep 10 00:32:26.263498 systemd[1]: Reloading... Sep 10 00:32:26.343425 zram_generator::config[1224]: No configuration found. Sep 10 00:32:26.442836 ldconfig[1165]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 10 00:32:26.505021 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:32:26.565671 systemd[1]: Reloading finished in 301 ms. Sep 10 00:32:26.613606 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 10 00:32:26.615662 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 10 00:32:26.632945 systemd[1]: Starting ensure-sysext.service... Sep 10 00:32:26.636553 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 10 00:32:26.666547 systemd[1]: Reloading requested from client PID 1260 ('systemctl') (unit ensure-sysext.service)... Sep 10 00:32:26.666576 systemd[1]: Reloading... Sep 10 00:32:26.693224 systemd-tmpfiles[1261]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 10 00:32:26.693672 systemd-tmpfiles[1261]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 10 00:32:26.694740 systemd-tmpfiles[1261]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 10 00:32:26.695091 systemd-tmpfiles[1261]: ACLs are not supported, ignoring. Sep 10 00:32:26.695192 systemd-tmpfiles[1261]: ACLs are not supported, ignoring. Sep 10 00:32:26.700270 systemd-tmpfiles[1261]: Detected autofs mount point /boot during canonicalization of boot. Sep 10 00:32:26.700284 systemd-tmpfiles[1261]: Skipping /boot Sep 10 00:32:26.724422 systemd-tmpfiles[1261]: Detected autofs mount point /boot during canonicalization of boot. Sep 10 00:32:26.724450 systemd-tmpfiles[1261]: Skipping /boot Sep 10 00:32:26.758915 zram_generator::config[1290]: No configuration found. Sep 10 00:32:26.871337 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:32:26.921773 systemd[1]: Reloading finished in 254 ms. Sep 10 00:32:26.943424 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 10 00:32:26.953873 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 10 00:32:26.960555 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 10 00:32:26.962918 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 10 00:32:26.965289 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 10 00:32:26.970597 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 10 00:32:26.974595 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 10 00:32:26.978568 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 10 00:32:26.986298 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 10 00:32:26.990063 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:26.990472 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:32:26.992349 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:32:26.995706 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:32:26.998955 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:32:27.000310 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:32:27.000602 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:27.004067 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:32:27.004272 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:32:27.010108 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 10 00:32:27.016224 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:27.016763 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:32:27.026042 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:32:27.027346 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:32:27.027520 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:27.028543 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 10 00:32:27.030681 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:32:27.030886 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:32:27.038295 systemd-udevd[1332]: Using default interface naming scheme 'v255'. Sep 10 00:32:27.044320 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 10 00:32:27.046199 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:32:27.046535 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:32:27.048619 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:32:27.048857 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:32:27.052263 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 10 00:32:27.055074 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 10 00:32:27.055364 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 10 00:32:27.065952 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:27.066189 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 10 00:32:27.067634 augenrules[1363]: No rules Sep 10 00:32:27.073184 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 10 00:32:27.095772 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 10 00:32:27.100654 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 10 00:32:27.104681 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 10 00:32:27.105793 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 10 00:32:27.105916 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 10 00:32:27.106787 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 10 00:32:27.115828 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 10 00:32:27.118180 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 10 00:32:27.122223 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 10 00:32:27.124982 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 10 00:32:27.125182 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 10 00:32:27.129021 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 10 00:32:27.129209 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 10 00:32:27.132249 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (1374) Sep 10 00:32:27.132982 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 10 00:32:27.134613 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 10 00:32:27.147567 systemd[1]: Finished ensure-sysext.service. Sep 10 00:32:27.150026 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 10 00:32:27.152473 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 10 00:32:27.178371 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 10 00:32:27.183551 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 10 00:32:27.198619 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 10 00:32:27.202616 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 10 00:32:27.203148 systemd-resolved[1330]: Positive Trust Anchors: Sep 10 00:32:27.203158 systemd-resolved[1330]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 10 00:32:27.203191 systemd-resolved[1330]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 10 00:32:27.205412 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 10 00:32:27.206167 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 10 00:32:27.206240 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 10 00:32:27.210544 kernel: ACPI: button: Power Button [PWRF] Sep 10 00:32:27.211727 systemd-resolved[1330]: Defaulting to hostname 'linux'. Sep 10 00:32:27.214597 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 10 00:32:27.214876 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 10 00:32:27.215129 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 10 00:32:27.217595 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 10 00:32:27.224343 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 10 00:32:27.251419 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Sep 10 00:32:27.259373 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 10 00:32:27.260059 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Sep 10 00:32:27.260451 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 10 00:32:27.320977 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 10 00:32:27.322615 systemd[1]: Reached target time-set.target - System Time Set. Sep 10 00:32:27.324236 systemd-networkd[1408]: lo: Link UP Sep 10 00:32:27.324254 systemd-networkd[1408]: lo: Gained carrier Sep 10 00:32:27.326158 systemd-networkd[1408]: Enumeration completed Sep 10 00:32:27.326269 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 10 00:32:27.326675 systemd-networkd[1408]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:32:27.326687 systemd-networkd[1408]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 10 00:32:27.328521 systemd[1]: Reached target network.target - Network. Sep 10 00:32:27.328894 systemd-networkd[1408]: eth0: Link UP Sep 10 00:32:27.328906 systemd-networkd[1408]: eth0: Gained carrier Sep 10 00:32:27.328919 systemd-networkd[1408]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 10 00:32:27.340612 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 10 00:32:27.368138 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 10 00:32:27.384087 systemd-networkd[1408]: eth0: DHCPv4 address 10.0.0.29/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 10 00:32:27.386140 systemd-timesyncd[1409]: Network configuration changed, trying to establish connection. Sep 10 00:32:27.387375 systemd-timesyncd[1409]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 10 00:32:27.387583 systemd-timesyncd[1409]: Initial clock synchronization to Wed 2025-09-10 00:32:27.682171 UTC. Sep 10 00:32:27.396447 kernel: mousedev: PS/2 mouse device common for all mice Sep 10 00:32:27.409724 kernel: kvm_amd: TSC scaling supported Sep 10 00:32:27.409765 kernel: kvm_amd: Nested Virtualization enabled Sep 10 00:32:27.409779 kernel: kvm_amd: Nested Paging enabled Sep 10 00:32:27.410857 kernel: kvm_amd: LBR virtualization supported Sep 10 00:32:27.410886 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 10 00:32:27.412078 kernel: kvm_amd: Virtual GIF supported Sep 10 00:32:27.436726 kernel: EDAC MC: Ver: 3.0.0 Sep 10 00:32:27.484868 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 10 00:32:27.495802 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 10 00:32:27.509566 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 10 00:32:27.518183 lvm[1426]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 10 00:32:27.551672 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 10 00:32:27.553404 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 10 00:32:27.554677 systemd[1]: Reached target sysinit.target - System Initialization. Sep 10 00:32:27.555994 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 10 00:32:27.557425 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 10 00:32:27.559055 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 10 00:32:27.560589 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 10 00:32:27.562073 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 10 00:32:27.563501 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 10 00:32:27.563529 systemd[1]: Reached target paths.target - Path Units. Sep 10 00:32:27.564570 systemd[1]: Reached target timers.target - Timer Units. Sep 10 00:32:27.566518 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 10 00:32:27.569480 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 10 00:32:27.584008 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 10 00:32:27.586568 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 10 00:32:27.588245 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 10 00:32:27.589517 systemd[1]: Reached target sockets.target - Socket Units. Sep 10 00:32:27.590476 systemd[1]: Reached target basic.target - Basic System. Sep 10 00:32:27.591409 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 10 00:32:27.591455 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 10 00:32:27.592486 systemd[1]: Starting containerd.service - containerd container runtime... Sep 10 00:32:27.594590 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 10 00:32:27.598502 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 10 00:32:27.601845 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 10 00:32:27.603049 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 10 00:32:27.605615 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 10 00:32:27.611512 jq[1433]: false Sep 10 00:32:27.612708 lvm[1430]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 10 00:32:27.610582 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 10 00:32:27.615555 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 10 00:32:27.619494 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 10 00:32:27.627210 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 10 00:32:27.629278 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 10 00:32:27.629866 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 10 00:32:27.630863 systemd[1]: Starting update-engine.service - Update Engine... Sep 10 00:32:27.636124 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 10 00:32:27.638469 extend-filesystems[1434]: Found loop3 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found loop4 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found loop5 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found sr0 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda1 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda2 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda3 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found usr Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda4 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda6 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda7 Sep 10 00:32:27.638469 extend-filesystems[1434]: Found vda9 Sep 10 00:32:27.638469 extend-filesystems[1434]: Checking size of /dev/vda9 Sep 10 00:32:27.685945 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (1374) Sep 10 00:32:27.685984 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 10 00:32:27.686000 extend-filesystems[1434]: Resized partition /dev/vda9 Sep 10 00:32:27.639577 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 10 00:32:27.648642 dbus-daemon[1432]: [system] SELinux support is enabled Sep 10 00:32:27.687706 update_engine[1445]: I20250910 00:32:27.655438 1445 main.cc:92] Flatcar Update Engine starting Sep 10 00:32:27.687706 update_engine[1445]: I20250910 00:32:27.675184 1445 update_check_scheduler.cc:74] Next update check in 7m55s Sep 10 00:32:27.687968 extend-filesystems[1461]: resize2fs 1.47.1 (20-May-2024) Sep 10 00:32:27.642552 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 10 00:32:27.689709 jq[1448]: true Sep 10 00:32:27.642810 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 10 00:32:27.643149 systemd[1]: motdgen.service: Deactivated successfully. Sep 10 00:32:27.644224 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 10 00:32:27.647756 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 10 00:32:27.691241 jq[1456]: true Sep 10 00:32:27.647981 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 10 00:32:27.651412 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 10 00:32:27.682749 (ntainerd)[1458]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 10 00:32:27.694400 tar[1453]: linux-amd64/LICENSE Sep 10 00:32:27.695178 tar[1453]: linux-amd64/helm Sep 10 00:32:27.736010 systemd[1]: Started update-engine.service - Update Engine. Sep 10 00:32:27.738964 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 10 00:32:27.738999 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 10 00:32:27.740515 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 10 00:32:27.740544 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 10 00:32:27.744410 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 10 00:32:27.746096 systemd-logind[1442]: Watching system buttons on /dev/input/event1 (Power Button) Sep 10 00:32:27.746116 systemd-logind[1442]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 10 00:32:27.748271 systemd-logind[1442]: New seat seat0. Sep 10 00:32:27.753150 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 10 00:32:27.754728 systemd[1]: Started systemd-logind.service - User Login Management. Sep 10 00:32:27.774787 extend-filesystems[1461]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 10 00:32:27.774787 extend-filesystems[1461]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 10 00:32:27.774787 extend-filesystems[1461]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 10 00:32:27.786374 extend-filesystems[1434]: Resized filesystem in /dev/vda9 Sep 10 00:32:27.777408 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 10 00:32:27.777659 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 10 00:32:27.803991 bash[1486]: Updated "/home/core/.ssh/authorized_keys" Sep 10 00:32:27.806095 locksmithd[1474]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 10 00:32:27.806744 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 10 00:32:27.810186 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 10 00:32:27.882104 sshd_keygen[1457]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 10 00:32:27.911991 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 10 00:32:27.925436 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 10 00:32:27.934303 systemd[1]: issuegen.service: Deactivated successfully. Sep 10 00:32:27.934626 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 10 00:32:27.941685 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 10 00:32:27.995056 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 10 00:32:28.007690 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 10 00:32:28.011370 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 10 00:32:28.013641 systemd[1]: Reached target getty.target - Login Prompts. Sep 10 00:32:28.105098 containerd[1458]: time="2025-09-10T00:32:28.104953557Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Sep 10 00:32:28.133481 containerd[1458]: time="2025-09-10T00:32:28.133181733Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.135682 containerd[1458]: time="2025-09-10T00:32:28.135630913Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.104-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:32:28.135682 containerd[1458]: time="2025-09-10T00:32:28.135678305Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 10 00:32:28.135774 containerd[1458]: time="2025-09-10T00:32:28.135702623Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 10 00:32:28.136455 containerd[1458]: time="2025-09-10T00:32:28.136103257Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 10 00:32:28.136455 containerd[1458]: time="2025-09-10T00:32:28.136129456Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136455 containerd[1458]: time="2025-09-10T00:32:28.136205997Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136455 containerd[1458]: time="2025-09-10T00:32:28.136219980Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136646 containerd[1458]: time="2025-09-10T00:32:28.136469537Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136646 containerd[1458]: time="2025-09-10T00:32:28.136487093Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136646 containerd[1458]: time="2025-09-10T00:32:28.136503226Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136646 containerd[1458]: time="2025-09-10T00:32:28.136514300Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136646 containerd[1458]: time="2025-09-10T00:32:28.136616177Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.136929 containerd[1458]: time="2025-09-10T00:32:28.136897991Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 10 00:32:28.137069 containerd[1458]: time="2025-09-10T00:32:28.137040195Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 10 00:32:28.137069 containerd[1458]: time="2025-09-10T00:32:28.137059664Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 10 00:32:28.137182 containerd[1458]: time="2025-09-10T00:32:28.137163619Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 10 00:32:28.137244 containerd[1458]: time="2025-09-10T00:32:28.137227610Z" level=info msg="metadata content store policy set" policy=shared Sep 10 00:32:28.143557 containerd[1458]: time="2025-09-10T00:32:28.143519971Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 10 00:32:28.143602 containerd[1458]: time="2025-09-10T00:32:28.143568942Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 10 00:32:28.143602 containerd[1458]: time="2025-09-10T00:32:28.143586975Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 10 00:32:28.143664 containerd[1458]: time="2025-09-10T00:32:28.143603472Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 10 00:32:28.143664 containerd[1458]: time="2025-09-10T00:32:28.143617839Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 10 00:32:28.143803 containerd[1458]: time="2025-09-10T00:32:28.143773591Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 10 00:32:28.144127 containerd[1458]: time="2025-09-10T00:32:28.144097300Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 10 00:32:28.144247 containerd[1458]: time="2025-09-10T00:32:28.144226758Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 10 00:32:28.144269 containerd[1458]: time="2025-09-10T00:32:28.144248761Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 10 00:32:28.144269 containerd[1458]: time="2025-09-10T00:32:28.144263678Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 10 00:32:28.144323 containerd[1458]: time="2025-09-10T00:32:28.144277806Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144323 containerd[1458]: time="2025-09-10T00:32:28.144292349Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144323 containerd[1458]: time="2025-09-10T00:32:28.144305325Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144323 containerd[1458]: time="2025-09-10T00:32:28.144319431Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144393 containerd[1458]: time="2025-09-10T00:32:28.144335471Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144393 containerd[1458]: time="2025-09-10T00:32:28.144349330Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144393 containerd[1458]: time="2025-09-10T00:32:28.144362606Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144393 containerd[1458]: time="2025-09-10T00:32:28.144374386Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144411565Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144428186Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144441328Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144454843Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144467870Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144482839Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144493 containerd[1458]: time="2025-09-10T00:32:28.144496469Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144511054Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144535414Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144551495Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144566299Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144581819Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144597993Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144614677Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144634238Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144647 containerd[1458]: time="2025-09-10T00:32:28.144646195Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144817 containerd[1458]: time="2025-09-10T00:32:28.144658224Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 10 00:32:28.144817 containerd[1458]: time="2025-09-10T00:32:28.144726403Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 10 00:32:28.144817 containerd[1458]: time="2025-09-10T00:32:28.144746359Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 10 00:32:28.144817 containerd[1458]: time="2025-09-10T00:32:28.144757952Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 10 00:32:28.144899 containerd[1458]: time="2025-09-10T00:32:28.144849878Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 10 00:32:28.144899 containerd[1458]: time="2025-09-10T00:32:28.144864493Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.144899 containerd[1458]: time="2025-09-10T00:32:28.144887016Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 10 00:32:28.144955 containerd[1458]: time="2025-09-10T00:32:28.144901300Z" level=info msg="NRI interface is disabled by configuration." Sep 10 00:32:28.144955 containerd[1458]: time="2025-09-10T00:32:28.144913880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 10 00:32:28.145244 containerd[1458]: time="2025-09-10T00:32:28.145182510Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 10 00:32:28.145244 containerd[1458]: time="2025-09-10T00:32:28.145239614Z" level=info msg="Connect containerd service" Sep 10 00:32:28.145244 containerd[1458]: time="2025-09-10T00:32:28.145271911Z" level=info msg="using legacy CRI server" Sep 10 00:32:28.145244 containerd[1458]: time="2025-09-10T00:32:28.145282663Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 10 00:32:28.145660 containerd[1458]: time="2025-09-10T00:32:28.145367504Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 10 00:32:28.146210 containerd[1458]: time="2025-09-10T00:32:28.146172314Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 10 00:32:28.146625 containerd[1458]: time="2025-09-10T00:32:28.146460059Z" level=info msg="Start subscribing containerd event" Sep 10 00:32:28.146625 containerd[1458]: time="2025-09-10T00:32:28.146522597Z" level=info msg="Start recovering state" Sep 10 00:32:28.146625 containerd[1458]: time="2025-09-10T00:32:28.146572731Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 10 00:32:28.146711 containerd[1458]: time="2025-09-10T00:32:28.146646186Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 10 00:32:28.147513 containerd[1458]: time="2025-09-10T00:32:28.147490129Z" level=info msg="Start event monitor" Sep 10 00:32:28.147550 containerd[1458]: time="2025-09-10T00:32:28.147523236Z" level=info msg="Start snapshots syncer" Sep 10 00:32:28.147550 containerd[1458]: time="2025-09-10T00:32:28.147536055Z" level=info msg="Start cni network conf syncer for default" Sep 10 00:32:28.147550 containerd[1458]: time="2025-09-10T00:32:28.147545176Z" level=info msg="Start streaming server" Sep 10 00:32:28.147629 containerd[1458]: time="2025-09-10T00:32:28.147621737Z" level=info msg="containerd successfully booted in 0.044022s" Sep 10 00:32:28.147889 systemd[1]: Started containerd.service - containerd container runtime. Sep 10 00:32:28.262237 tar[1453]: linux-amd64/README.md Sep 10 00:32:28.287371 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 10 00:32:29.389760 systemd-networkd[1408]: eth0: Gained IPv6LL Sep 10 00:32:29.395407 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 10 00:32:29.397626 systemd[1]: Reached target network-online.target - Network is Online. Sep 10 00:32:29.417019 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 10 00:32:29.420401 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:32:29.422990 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 10 00:32:29.447702 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 10 00:32:29.447996 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 10 00:32:29.449719 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 10 00:32:29.450209 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 10 00:32:30.897191 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:32:30.899002 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 10 00:32:30.904471 systemd[1]: Startup finished in 918ms (kernel) + 7.138s (initrd) + 6.093s (userspace) = 14.150s. Sep 10 00:32:30.913437 (kubelet)[1544]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:32:31.178047 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 10 00:32:31.190673 systemd[1]: Started sshd@0-10.0.0.29:22-10.0.0.1:43260.service - OpenSSH per-connection server daemon (10.0.0.1:43260). Sep 10 00:32:31.241486 sshd[1555]: Accepted publickey for core from 10.0.0.1 port 43260 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:31.243771 sshd[1555]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:31.255289 systemd-logind[1442]: New session 1 of user core. Sep 10 00:32:31.256967 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 10 00:32:31.272708 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 10 00:32:31.285739 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 10 00:32:31.296793 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 10 00:32:31.302198 (systemd)[1559]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 10 00:32:31.427085 systemd[1559]: Queued start job for default target default.target. Sep 10 00:32:31.439841 systemd[1559]: Created slice app.slice - User Application Slice. Sep 10 00:32:31.439868 systemd[1559]: Reached target paths.target - Paths. Sep 10 00:32:31.439883 systemd[1559]: Reached target timers.target - Timers. Sep 10 00:32:31.441779 systemd[1559]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 10 00:32:31.465654 systemd[1559]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 10 00:32:31.465796 systemd[1559]: Reached target sockets.target - Sockets. Sep 10 00:32:31.465812 systemd[1559]: Reached target basic.target - Basic System. Sep 10 00:32:31.465868 systemd[1559]: Reached target default.target - Main User Target. Sep 10 00:32:31.465908 systemd[1559]: Startup finished in 153ms. Sep 10 00:32:31.466080 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 10 00:32:31.476557 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 10 00:32:31.545731 systemd[1]: Started sshd@1-10.0.0.29:22-10.0.0.1:43270.service - OpenSSH per-connection server daemon (10.0.0.1:43270). Sep 10 00:32:31.586372 sshd[1570]: Accepted publickey for core from 10.0.0.1 port 43270 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:31.588306 sshd[1570]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:31.592992 systemd-logind[1442]: New session 2 of user core. Sep 10 00:32:31.607559 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 10 00:32:31.729753 sshd[1570]: pam_unix(sshd:session): session closed for user core Sep 10 00:32:31.740325 systemd[1]: sshd@1-10.0.0.29:22-10.0.0.1:43270.service: Deactivated successfully. Sep 10 00:32:31.742264 systemd[1]: session-2.scope: Deactivated successfully. Sep 10 00:32:31.743996 systemd-logind[1442]: Session 2 logged out. Waiting for processes to exit. Sep 10 00:32:31.754675 systemd[1]: Started sshd@2-10.0.0.29:22-10.0.0.1:43278.service - OpenSSH per-connection server daemon (10.0.0.1:43278). Sep 10 00:32:31.755658 systemd-logind[1442]: Removed session 2. Sep 10 00:32:31.790644 sshd[1578]: Accepted publickey for core from 10.0.0.1 port 43278 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:31.792391 sshd[1578]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:31.797005 systemd-logind[1442]: New session 3 of user core. Sep 10 00:32:31.804618 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 10 00:32:31.839277 kubelet[1544]: E0910 00:32:31.839219 1544 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:32:31.844295 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:32:31.844548 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:32:31.844926 systemd[1]: kubelet.service: Consumed 1.886s CPU time. Sep 10 00:32:31.855582 sshd[1578]: pam_unix(sshd:session): session closed for user core Sep 10 00:32:31.875871 systemd[1]: sshd@2-10.0.0.29:22-10.0.0.1:43278.service: Deactivated successfully. Sep 10 00:32:31.877765 systemd[1]: session-3.scope: Deactivated successfully. Sep 10 00:32:31.880338 systemd-logind[1442]: Session 3 logged out. Waiting for processes to exit. Sep 10 00:32:31.888815 systemd[1]: Started sshd@3-10.0.0.29:22-10.0.0.1:43284.service - OpenSSH per-connection server daemon (10.0.0.1:43284). Sep 10 00:32:31.891037 systemd-logind[1442]: Removed session 3. Sep 10 00:32:31.926784 sshd[1586]: Accepted publickey for core from 10.0.0.1 port 43284 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:31.928492 sshd[1586]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:31.933145 systemd-logind[1442]: New session 4 of user core. Sep 10 00:32:31.942723 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 10 00:32:32.000911 sshd[1586]: pam_unix(sshd:session): session closed for user core Sep 10 00:32:32.018741 systemd[1]: sshd@3-10.0.0.29:22-10.0.0.1:43284.service: Deactivated successfully. Sep 10 00:32:32.020953 systemd[1]: session-4.scope: Deactivated successfully. Sep 10 00:32:32.023029 systemd-logind[1442]: Session 4 logged out. Waiting for processes to exit. Sep 10 00:32:32.024852 systemd[1]: Started sshd@4-10.0.0.29:22-10.0.0.1:43300.service - OpenSSH per-connection server daemon (10.0.0.1:43300). Sep 10 00:32:32.025838 systemd-logind[1442]: Removed session 4. Sep 10 00:32:32.066962 sshd[1593]: Accepted publickey for core from 10.0.0.1 port 43300 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:32.068883 sshd[1593]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:32.073356 systemd-logind[1442]: New session 5 of user core. Sep 10 00:32:32.086603 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 10 00:32:32.148858 sudo[1596]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 10 00:32:32.149316 sudo[1596]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:32:32.170327 sudo[1596]: pam_unix(sudo:session): session closed for user root Sep 10 00:32:32.172689 sshd[1593]: pam_unix(sshd:session): session closed for user core Sep 10 00:32:32.190378 systemd[1]: sshd@4-10.0.0.29:22-10.0.0.1:43300.service: Deactivated successfully. Sep 10 00:32:32.192728 systemd[1]: session-5.scope: Deactivated successfully. Sep 10 00:32:32.194400 systemd-logind[1442]: Session 5 logged out. Waiting for processes to exit. Sep 10 00:32:32.206022 systemd[1]: Started sshd@5-10.0.0.29:22-10.0.0.1:43316.service - OpenSSH per-connection server daemon (10.0.0.1:43316). Sep 10 00:32:32.207286 systemd-logind[1442]: Removed session 5. Sep 10 00:32:32.244113 sshd[1601]: Accepted publickey for core from 10.0.0.1 port 43316 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:32.245834 sshd[1601]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:32.250210 systemd-logind[1442]: New session 6 of user core. Sep 10 00:32:32.266593 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 10 00:32:32.322881 sudo[1605]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 10 00:32:32.323250 sudo[1605]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:32:32.327193 sudo[1605]: pam_unix(sudo:session): session closed for user root Sep 10 00:32:32.334180 sudo[1604]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 10 00:32:32.334637 sudo[1604]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:32:32.354619 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 10 00:32:32.356441 auditctl[1608]: No rules Sep 10 00:32:32.357833 systemd[1]: audit-rules.service: Deactivated successfully. Sep 10 00:32:32.358103 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 10 00:32:32.360023 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 10 00:32:32.397244 augenrules[1626]: No rules Sep 10 00:32:32.399462 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 10 00:32:32.401081 sudo[1604]: pam_unix(sudo:session): session closed for user root Sep 10 00:32:32.403067 sshd[1601]: pam_unix(sshd:session): session closed for user core Sep 10 00:32:32.414923 systemd[1]: sshd@5-10.0.0.29:22-10.0.0.1:43316.service: Deactivated successfully. Sep 10 00:32:32.417256 systemd[1]: session-6.scope: Deactivated successfully. Sep 10 00:32:32.419465 systemd-logind[1442]: Session 6 logged out. Waiting for processes to exit. Sep 10 00:32:32.429836 systemd[1]: Started sshd@6-10.0.0.29:22-10.0.0.1:43326.service - OpenSSH per-connection server daemon (10.0.0.1:43326). Sep 10 00:32:32.431072 systemd-logind[1442]: Removed session 6. Sep 10 00:32:32.468274 sshd[1634]: Accepted publickey for core from 10.0.0.1 port 43326 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:32:32.470033 sshd[1634]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:32:32.474168 systemd-logind[1442]: New session 7 of user core. Sep 10 00:32:32.485536 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 10 00:32:32.540615 sudo[1637]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 10 00:32:32.540967 sudo[1637]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 10 00:32:33.066623 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 10 00:32:33.066843 (dockerd)[1656]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 10 00:32:34.317243 dockerd[1656]: time="2025-09-10T00:32:34.317157428Z" level=info msg="Starting up" Sep 10 00:32:35.072963 systemd[1]: var-lib-docker-metacopy\x2dcheck3556883088-merged.mount: Deactivated successfully. Sep 10 00:32:35.104701 dockerd[1656]: time="2025-09-10T00:32:35.104628368Z" level=info msg="Loading containers: start." Sep 10 00:32:35.321446 kernel: Initializing XFRM netlink socket Sep 10 00:32:35.407067 systemd-networkd[1408]: docker0: Link UP Sep 10 00:32:35.429415 dockerd[1656]: time="2025-09-10T00:32:35.429325587Z" level=info msg="Loading containers: done." Sep 10 00:32:35.446876 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3657601120-merged.mount: Deactivated successfully. Sep 10 00:32:35.448907 dockerd[1656]: time="2025-09-10T00:32:35.448854647Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 10 00:32:35.448995 dockerd[1656]: time="2025-09-10T00:32:35.448976258Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 10 00:32:35.449160 dockerd[1656]: time="2025-09-10T00:32:35.449134083Z" level=info msg="Daemon has completed initialization" Sep 10 00:32:35.602842 dockerd[1656]: time="2025-09-10T00:32:35.602743536Z" level=info msg="API listen on /run/docker.sock" Sep 10 00:32:35.603150 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 10 00:32:36.735883 containerd[1458]: time="2025-09-10T00:32:36.735819271Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.4\"" Sep 10 00:32:37.380751 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3832690982.mount: Deactivated successfully. Sep 10 00:32:38.663266 containerd[1458]: time="2025-09-10T00:32:38.663141772Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:38.664155 containerd[1458]: time="2025-09-10T00:32:38.664072777Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.4: active requests=0, bytes read=30078664" Sep 10 00:32:38.665473 containerd[1458]: time="2025-09-10T00:32:38.665433915Z" level=info msg="ImageCreate event name:\"sha256:1f41885d0a91155d5a5e670b2862eed338c7f12b0e8a5bbc88b1ab4a2d505ae8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:38.669012 containerd[1458]: time="2025-09-10T00:32:38.668958296Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:0d441d0d347145b3f02f20cb313239cdae86067643d7f70803fab8bac2d28876\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:38.670012 containerd[1458]: time="2025-09-10T00:32:38.669973112Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.4\" with image id \"sha256:1f41885d0a91155d5a5e670b2862eed338c7f12b0e8a5bbc88b1ab4a2d505ae8\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:0d441d0d347145b3f02f20cb313239cdae86067643d7f70803fab8bac2d28876\", size \"30075464\" in 1.934094553s" Sep 10 00:32:38.670074 containerd[1458]: time="2025-09-10T00:32:38.670013576Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.4\" returns image reference \"sha256:1f41885d0a91155d5a5e670b2862eed338c7f12b0e8a5bbc88b1ab4a2d505ae8\"" Sep 10 00:32:38.670772 containerd[1458]: time="2025-09-10T00:32:38.670742848Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.4\"" Sep 10 00:32:40.314425 containerd[1458]: time="2025-09-10T00:32:40.314328396Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:40.316221 containerd[1458]: time="2025-09-10T00:32:40.316145114Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.4: active requests=0, bytes read=26018066" Sep 10 00:32:40.318091 containerd[1458]: time="2025-09-10T00:32:40.318044253Z" level=info msg="ImageCreate event name:\"sha256:358ab71c1a1ea4846ad0b3dff0d9db6b124236b64bc8a6b79dc874f65dc0d492\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:40.321743 containerd[1458]: time="2025-09-10T00:32:40.321699852Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:bd22c2af2f30a8f818568b4d5fe131098fdd38267e9e07872cfc33e8f5876bc3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:40.323589 containerd[1458]: time="2025-09-10T00:32:40.323509918Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.4\" with image id \"sha256:358ab71c1a1ea4846ad0b3dff0d9db6b124236b64bc8a6b79dc874f65dc0d492\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:bd22c2af2f30a8f818568b4d5fe131098fdd38267e9e07872cfc33e8f5876bc3\", size \"27646961\" in 1.652725469s" Sep 10 00:32:40.323655 containerd[1458]: time="2025-09-10T00:32:40.323593299Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.4\" returns image reference \"sha256:358ab71c1a1ea4846ad0b3dff0d9db6b124236b64bc8a6b79dc874f65dc0d492\"" Sep 10 00:32:40.324426 containerd[1458]: time="2025-09-10T00:32:40.324190362Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.4\"" Sep 10 00:32:42.034089 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 10 00:32:42.044767 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:32:42.047593 containerd[1458]: time="2025-09-10T00:32:42.047481307Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:42.048968 containerd[1458]: time="2025-09-10T00:32:42.048928913Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.4: active requests=0, bytes read=20153911" Sep 10 00:32:42.050540 containerd[1458]: time="2025-09-10T00:32:42.050482526Z" level=info msg="ImageCreate event name:\"sha256:ab4ad8a84c3c69c18494ef32fa087b32f7c44d71e6acba463d2c7dda798c3d66\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:42.053798 containerd[1458]: time="2025-09-10T00:32:42.053770017Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:71533e5a960e2955a54164905e92dac516ec874a23e0bf31304db82650101a4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:42.054796 containerd[1458]: time="2025-09-10T00:32:42.054741826Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.4\" with image id \"sha256:ab4ad8a84c3c69c18494ef32fa087b32f7c44d71e6acba463d2c7dda798c3d66\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:71533e5a960e2955a54164905e92dac516ec874a23e0bf31304db82650101a4a\", size \"21782824\" in 1.730517619s" Sep 10 00:32:42.054796 containerd[1458]: time="2025-09-10T00:32:42.054790481Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.4\" returns image reference \"sha256:ab4ad8a84c3c69c18494ef32fa087b32f7c44d71e6acba463d2c7dda798c3d66\"" Sep 10 00:32:42.056105 containerd[1458]: time="2025-09-10T00:32:42.056069822Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.4\"" Sep 10 00:32:42.264086 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:32:42.269092 (kubelet)[1876]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:32:42.654975 kubelet[1876]: E0910 00:32:42.654891 1876 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:32:42.662944 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:32:42.663202 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:32:43.684575 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3149254155.mount: Deactivated successfully. Sep 10 00:32:44.254605 containerd[1458]: time="2025-09-10T00:32:44.254516235Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:44.255596 containerd[1458]: time="2025-09-10T00:32:44.255532771Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.4: active requests=0, bytes read=31899626" Sep 10 00:32:44.256903 containerd[1458]: time="2025-09-10T00:32:44.256830087Z" level=info msg="ImageCreate event name:\"sha256:1b2ea5e018dbbbd2efb8e5c540a6d3c463d77f250d3904429402ee057f09c64e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:44.259837 containerd[1458]: time="2025-09-10T00:32:44.259762718Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bb04e9247da3aaeb96406b4d530a79fc865695b6807353dd1a28871df0d7f837\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:44.260407 containerd[1458]: time="2025-09-10T00:32:44.260314441Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.4\" with image id \"sha256:1b2ea5e018dbbbd2efb8e5c540a6d3c463d77f250d3904429402ee057f09c64e\", repo tag \"registry.k8s.io/kube-proxy:v1.33.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:bb04e9247da3aaeb96406b4d530a79fc865695b6807353dd1a28871df0d7f837\", size \"31898645\" in 2.20421049s" Sep 10 00:32:44.260407 containerd[1458]: time="2025-09-10T00:32:44.260358333Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.4\" returns image reference \"sha256:1b2ea5e018dbbbd2efb8e5c540a6d3c463d77f250d3904429402ee057f09c64e\"" Sep 10 00:32:44.261068 containerd[1458]: time="2025-09-10T00:32:44.261037979Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 10 00:32:45.016454 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2645355976.mount: Deactivated successfully. Sep 10 00:32:47.625682 containerd[1458]: time="2025-09-10T00:32:47.625605991Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:47.626891 containerd[1458]: time="2025-09-10T00:32:47.626855833Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Sep 10 00:32:47.629171 containerd[1458]: time="2025-09-10T00:32:47.629131418Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:47.632249 containerd[1458]: time="2025-09-10T00:32:47.632222751Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:47.633502 containerd[1458]: time="2025-09-10T00:32:47.633461550Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 3.372384595s" Sep 10 00:32:47.633565 containerd[1458]: time="2025-09-10T00:32:47.633503701Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Sep 10 00:32:47.634100 containerd[1458]: time="2025-09-10T00:32:47.634072229Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 10 00:32:48.195329 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4087129680.mount: Deactivated successfully. Sep 10 00:32:48.204104 containerd[1458]: time="2025-09-10T00:32:48.204011966Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:48.204897 containerd[1458]: time="2025-09-10T00:32:48.204805702Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 10 00:32:48.206010 containerd[1458]: time="2025-09-10T00:32:48.205975509Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:48.208353 containerd[1458]: time="2025-09-10T00:32:48.208291316Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:48.209685 containerd[1458]: time="2025-09-10T00:32:48.209630492Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 575.521502ms" Sep 10 00:32:48.209758 containerd[1458]: time="2025-09-10T00:32:48.209688588Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 10 00:32:48.210529 containerd[1458]: time="2025-09-10T00:32:48.210494919Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 10 00:32:48.747620 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1948617477.mount: Deactivated successfully. Sep 10 00:32:52.784435 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 10 00:32:52.799570 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:32:52.995626 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:32:53.002957 (kubelet)[2013]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 10 00:32:53.090463 kubelet[2013]: E0910 00:32:53.089361 2013 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 10 00:32:53.094206 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 10 00:32:53.094503 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 10 00:32:53.382625 containerd[1458]: time="2025-09-10T00:32:53.382434081Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:53.383565 containerd[1458]: time="2025-09-10T00:32:53.383494384Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58377871" Sep 10 00:32:53.384744 containerd[1458]: time="2025-09-10T00:32:53.384703840Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:53.388147 containerd[1458]: time="2025-09-10T00:32:53.388117668Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:32:53.389529 containerd[1458]: time="2025-09-10T00:32:53.389476701Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 5.17893417s" Sep 10 00:32:53.389701 containerd[1458]: time="2025-09-10T00:32:53.389537073Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Sep 10 00:32:56.954560 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:32:56.970768 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:32:56.997626 systemd[1]: Reloading requested from client PID 2051 ('systemctl') (unit session-7.scope)... Sep 10 00:32:56.997641 systemd[1]: Reloading... Sep 10 00:32:57.092433 zram_generator::config[2093]: No configuration found. Sep 10 00:32:57.344854 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:32:57.424312 systemd[1]: Reloading finished in 426 ms. Sep 10 00:32:57.477240 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 10 00:32:57.477342 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 10 00:32:57.477816 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:32:57.481322 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:32:57.686005 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:32:57.695890 (kubelet)[2139]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 10 00:32:57.812403 kubelet[2139]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:32:57.812403 kubelet[2139]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 10 00:32:57.812403 kubelet[2139]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:32:57.812403 kubelet[2139]: I0910 00:32:57.811860 2139 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 10 00:32:58.208321 kubelet[2139]: I0910 00:32:58.208262 2139 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 10 00:32:58.208321 kubelet[2139]: I0910 00:32:58.208305 2139 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 10 00:32:58.208726 kubelet[2139]: I0910 00:32:58.208703 2139 server.go:956] "Client rotation is on, will bootstrap in background" Sep 10 00:32:58.470196 kubelet[2139]: E0910 00:32:58.470017 2139 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.29:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 10 00:32:58.473627 kubelet[2139]: I0910 00:32:58.473583 2139 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 10 00:32:58.487833 kubelet[2139]: E0910 00:32:58.487772 2139 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 10 00:32:58.487833 kubelet[2139]: I0910 00:32:58.487832 2139 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 10 00:32:58.495963 kubelet[2139]: I0910 00:32:58.494902 2139 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 10 00:32:58.495963 kubelet[2139]: I0910 00:32:58.495404 2139 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 10 00:32:58.495963 kubelet[2139]: I0910 00:32:58.495437 2139 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 10 00:32:58.495963 kubelet[2139]: I0910 00:32:58.495722 2139 topology_manager.go:138] "Creating topology manager with none policy" Sep 10 00:32:58.496213 kubelet[2139]: I0910 00:32:58.495735 2139 container_manager_linux.go:303] "Creating device plugin manager" Sep 10 00:32:58.496213 kubelet[2139]: I0910 00:32:58.495966 2139 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:32:58.502499 kubelet[2139]: I0910 00:32:58.502454 2139 kubelet.go:480] "Attempting to sync node with API server" Sep 10 00:32:58.502499 kubelet[2139]: I0910 00:32:58.502493 2139 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 10 00:32:58.502597 kubelet[2139]: I0910 00:32:58.502538 2139 kubelet.go:386] "Adding apiserver pod source" Sep 10 00:32:58.502597 kubelet[2139]: I0910 00:32:58.502571 2139 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 10 00:32:58.511329 kubelet[2139]: E0910 00:32:58.511241 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.29:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 10 00:32:58.511525 kubelet[2139]: I0910 00:32:58.511379 2139 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 10 00:32:58.511919 kubelet[2139]: I0910 00:32:58.511886 2139 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 10 00:32:58.512829 kubelet[2139]: W0910 00:32:58.512802 2139 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 10 00:32:58.514679 kubelet[2139]: E0910 00:32:58.514633 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.29:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 10 00:32:58.517801 kubelet[2139]: I0910 00:32:58.517578 2139 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 10 00:32:58.517801 kubelet[2139]: I0910 00:32:58.517701 2139 server.go:1289] "Started kubelet" Sep 10 00:32:58.518047 kubelet[2139]: I0910 00:32:58.517981 2139 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 10 00:32:58.519845 kubelet[2139]: I0910 00:32:58.519180 2139 server.go:317] "Adding debug handlers to kubelet server" Sep 10 00:32:58.521043 kubelet[2139]: I0910 00:32:58.520966 2139 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 10 00:32:58.521465 kubelet[2139]: I0910 00:32:58.521443 2139 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 10 00:32:58.522535 kubelet[2139]: I0910 00:32:58.522515 2139 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 10 00:32:58.522892 kubelet[2139]: E0910 00:32:58.522777 2139 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:32:58.522892 kubelet[2139]: I0910 00:32:58.522823 2139 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 10 00:32:58.522892 kubelet[2139]: I0910 00:32:58.522871 2139 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 10 00:32:58.523175 kubelet[2139]: I0910 00:32:58.523148 2139 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 10 00:32:58.523458 kubelet[2139]: I0910 00:32:58.523282 2139 reconciler.go:26] "Reconciler: start to sync state" Sep 10 00:32:58.523834 kubelet[2139]: E0910 00:32:58.523800 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.29:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 10 00:32:58.525400 kubelet[2139]: E0910 00:32:58.522347 2139 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.29:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.29:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1863c48a72f21d4e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-10 00:32:58.51762619 +0000 UTC m=+0.810746459,LastTimestamp:2025-09-10 00:32:58.51762619 +0000 UTC m=+0.810746459,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 10 00:32:58.525935 kubelet[2139]: I0910 00:32:58.525905 2139 factory.go:223] Registration of the systemd container factory successfully Sep 10 00:32:58.526002 kubelet[2139]: I0910 00:32:58.525981 2139 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 10 00:32:58.526970 kubelet[2139]: E0910 00:32:58.526930 2139 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 10 00:32:58.527268 kubelet[2139]: E0910 00:32:58.527204 2139 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="200ms" Sep 10 00:32:58.528234 kubelet[2139]: I0910 00:32:58.528208 2139 factory.go:223] Registration of the containerd container factory successfully Sep 10 00:32:58.548132 kubelet[2139]: I0910 00:32:58.548091 2139 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 10 00:32:58.549076 kubelet[2139]: I0910 00:32:58.549040 2139 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 10 00:32:58.549076 kubelet[2139]: I0910 00:32:58.549060 2139 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 10 00:32:58.549076 kubelet[2139]: I0910 00:32:58.549080 2139 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:32:58.549651 kubelet[2139]: I0910 00:32:58.549629 2139 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 10 00:32:58.549687 kubelet[2139]: I0910 00:32:58.549655 2139 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 10 00:32:58.549687 kubelet[2139]: I0910 00:32:58.549678 2139 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 10 00:32:58.549757 kubelet[2139]: I0910 00:32:58.549694 2139 kubelet.go:2436] "Starting kubelet main sync loop" Sep 10 00:32:58.549757 kubelet[2139]: E0910 00:32:58.549736 2139 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 10 00:32:58.551304 kubelet[2139]: E0910 00:32:58.550316 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.29:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 10 00:32:58.623857 kubelet[2139]: E0910 00:32:58.623783 2139 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:32:58.650317 kubelet[2139]: E0910 00:32:58.650207 2139 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 10 00:32:58.724105 kubelet[2139]: E0910 00:32:58.723933 2139 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:32:58.728565 kubelet[2139]: E0910 00:32:58.728517 2139 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="400ms" Sep 10 00:32:58.761682 kubelet[2139]: I0910 00:32:58.761604 2139 policy_none.go:49] "None policy: Start" Sep 10 00:32:58.761682 kubelet[2139]: I0910 00:32:58.761672 2139 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 10 00:32:58.761682 kubelet[2139]: I0910 00:32:58.761699 2139 state_mem.go:35] "Initializing new in-memory state store" Sep 10 00:32:58.775450 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 10 00:32:58.790863 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 10 00:32:58.805871 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 10 00:32:58.807586 kubelet[2139]: E0910 00:32:58.807553 2139 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 10 00:32:58.807915 kubelet[2139]: I0910 00:32:58.807886 2139 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 10 00:32:58.808042 kubelet[2139]: I0910 00:32:58.807910 2139 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 10 00:32:58.808510 kubelet[2139]: I0910 00:32:58.808195 2139 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 10 00:32:58.810313 kubelet[2139]: E0910 00:32:58.810286 2139 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 10 00:32:58.810444 kubelet[2139]: E0910 00:32:58.810405 2139 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 10 00:32:58.866034 systemd[1]: Created slice kubepods-burstable-pod693363976a025a035618fa0cc32f4434.slice - libcontainer container kubepods-burstable-pod693363976a025a035618fa0cc32f4434.slice. Sep 10 00:32:58.883791 kubelet[2139]: E0910 00:32:58.883725 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:32:58.886996 systemd[1]: Created slice kubepods-burstable-pod8de7187202bee21b84740a213836f615.slice - libcontainer container kubepods-burstable-pod8de7187202bee21b84740a213836f615.slice. Sep 10 00:32:58.901055 kubelet[2139]: E0910 00:32:58.901016 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:32:58.903792 systemd[1]: Created slice kubepods-burstable-podd75e6f6978d9f275ea19380916c9cccd.slice - libcontainer container kubepods-burstable-podd75e6f6978d9f275ea19380916c9cccd.slice. Sep 10 00:32:58.905870 kubelet[2139]: E0910 00:32:58.905838 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:32:58.910254 kubelet[2139]: I0910 00:32:58.910224 2139 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:32:58.910834 kubelet[2139]: E0910 00:32:58.910785 2139 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 10 00:32:58.925301 kubelet[2139]: I0910 00:32:58.925223 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/693363976a025a035618fa0cc32f4434-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"693363976a025a035618fa0cc32f4434\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:32:58.925356 kubelet[2139]: I0910 00:32:58.925297 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:32:58.925356 kubelet[2139]: I0910 00:32:58.925330 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:32:58.925440 kubelet[2139]: I0910 00:32:58.925358 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:32:58.925473 kubelet[2139]: I0910 00:32:58.925452 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d75e6f6978d9f275ea19380916c9cccd-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d75e6f6978d9f275ea19380916c9cccd\") " pod="kube-system/kube-scheduler-localhost" Sep 10 00:32:58.925500 kubelet[2139]: I0910 00:32:58.925474 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/693363976a025a035618fa0cc32f4434-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"693363976a025a035618fa0cc32f4434\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:32:58.925525 kubelet[2139]: I0910 00:32:58.925498 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/693363976a025a035618fa0cc32f4434-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"693363976a025a035618fa0cc32f4434\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:32:58.925525 kubelet[2139]: I0910 00:32:58.925520 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:32:58.925599 kubelet[2139]: I0910 00:32:58.925565 2139 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:32:59.113642 kubelet[2139]: I0910 00:32:59.113595 2139 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:32:59.114120 kubelet[2139]: E0910 00:32:59.114088 2139 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 10 00:32:59.129220 kubelet[2139]: E0910 00:32:59.129148 2139 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="800ms" Sep 10 00:32:59.184456 kubelet[2139]: E0910 00:32:59.184368 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:32:59.185286 containerd[1458]: time="2025-09-10T00:32:59.185236161Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:693363976a025a035618fa0cc32f4434,Namespace:kube-system,Attempt:0,}" Sep 10 00:32:59.202823 kubelet[2139]: E0910 00:32:59.202758 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:32:59.204528 containerd[1458]: time="2025-09-10T00:32:59.204474162Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8de7187202bee21b84740a213836f615,Namespace:kube-system,Attempt:0,}" Sep 10 00:32:59.206886 kubelet[2139]: E0910 00:32:59.206842 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:32:59.207502 containerd[1458]: time="2025-09-10T00:32:59.207464062Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d75e6f6978d9f275ea19380916c9cccd,Namespace:kube-system,Attempt:0,}" Sep 10 00:32:59.492471 kubelet[2139]: E0910 00:32:59.492197 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.29:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 10 00:32:59.516188 kubelet[2139]: I0910 00:32:59.516148 2139 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:32:59.516563 kubelet[2139]: E0910 00:32:59.516525 2139 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 10 00:32:59.659351 kubelet[2139]: E0910 00:32:59.659301 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.29:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 10 00:32:59.743325 kubelet[2139]: E0910 00:32:59.743191 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.29:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 10 00:32:59.749892 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1496782699.mount: Deactivated successfully. Sep 10 00:32:59.757959 containerd[1458]: time="2025-09-10T00:32:59.757899604Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:32:59.760247 containerd[1458]: time="2025-09-10T00:32:59.760129298Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 10 00:32:59.761162 containerd[1458]: time="2025-09-10T00:32:59.761113622Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:32:59.762162 containerd[1458]: time="2025-09-10T00:32:59.762098668Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:32:59.762914 containerd[1458]: time="2025-09-10T00:32:59.762863214Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:32:59.763874 containerd[1458]: time="2025-09-10T00:32:59.763775735Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Sep 10 00:32:59.764754 containerd[1458]: time="2025-09-10T00:32:59.764696999Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 10 00:32:59.766332 containerd[1458]: time="2025-09-10T00:32:59.766290303Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 10 00:32:59.769365 containerd[1458]: time="2025-09-10T00:32:59.769338770Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 561.796073ms" Sep 10 00:32:59.771179 containerd[1458]: time="2025-09-10T00:32:59.771149898Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 566.562697ms" Sep 10 00:32:59.774029 containerd[1458]: time="2025-09-10T00:32:59.773974524Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 588.653293ms" Sep 10 00:32:59.875262 kubelet[2139]: E0910 00:32:59.875180 2139 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.29:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.29:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 10 00:32:59.930589 kubelet[2139]: E0910 00:32:59.930540 2139 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.29:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.29:6443: connect: connection refused" interval="1.6s" Sep 10 00:33:00.118211 containerd[1458]: time="2025-09-10T00:33:00.118071794Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:00.118211 containerd[1458]: time="2025-09-10T00:33:00.118142599Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:00.118211 containerd[1458]: time="2025-09-10T00:33:00.118157158Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:00.120994 containerd[1458]: time="2025-09-10T00:33:00.119574279Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:00.183781 systemd[1]: Started cri-containerd-4830362c221620d1bb33b676db54ad75760d1d84675254c40575183cac245043.scope - libcontainer container 4830362c221620d1bb33b676db54ad75760d1d84675254c40575183cac245043. Sep 10 00:33:00.188812 containerd[1458]: time="2025-09-10T00:33:00.188417799Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:00.188812 containerd[1458]: time="2025-09-10T00:33:00.188536401Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:00.188812 containerd[1458]: time="2025-09-10T00:33:00.188548472Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:00.188812 containerd[1458]: time="2025-09-10T00:33:00.188648784Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:00.190882 containerd[1458]: time="2025-09-10T00:33:00.190546746Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:00.190882 containerd[1458]: time="2025-09-10T00:33:00.190636169Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:00.190882 containerd[1458]: time="2025-09-10T00:33:00.190656362Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:00.190882 containerd[1458]: time="2025-09-10T00:33:00.190769659Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:00.216534 systemd[1]: Started cri-containerd-1f382b51de94969cefe89a79ae45b479960364bbb110bc90204b7ef14190711b.scope - libcontainer container 1f382b51de94969cefe89a79ae45b479960364bbb110bc90204b7ef14190711b. Sep 10 00:33:00.220473 systemd[1]: Started cri-containerd-6b8049b83542730a681d6ca820259c3dc0221ac1a261e3074292e8b908dc62db.scope - libcontainer container 6b8049b83542730a681d6ca820259c3dc0221ac1a261e3074292e8b908dc62db. Sep 10 00:33:00.234311 containerd[1458]: time="2025-09-10T00:33:00.234239968Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8de7187202bee21b84740a213836f615,Namespace:kube-system,Attempt:0,} returns sandbox id \"4830362c221620d1bb33b676db54ad75760d1d84675254c40575183cac245043\"" Sep 10 00:33:00.236151 kubelet[2139]: E0910 00:33:00.236110 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:00.242112 containerd[1458]: time="2025-09-10T00:33:00.242074095Z" level=info msg="CreateContainer within sandbox \"4830362c221620d1bb33b676db54ad75760d1d84675254c40575183cac245043\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 10 00:33:00.256364 containerd[1458]: time="2025-09-10T00:33:00.256327224Z" level=info msg="CreateContainer within sandbox \"4830362c221620d1bb33b676db54ad75760d1d84675254c40575183cac245043\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"160e7daa10d079d49200f971bd8d763880f282ad1580bf993ee1bac7a10dc7ed\"" Sep 10 00:33:00.258698 containerd[1458]: time="2025-09-10T00:33:00.258656365Z" level=info msg="StartContainer for \"160e7daa10d079d49200f971bd8d763880f282ad1580bf993ee1bac7a10dc7ed\"" Sep 10 00:33:00.266857 containerd[1458]: time="2025-09-10T00:33:00.266793275Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:693363976a025a035618fa0cc32f4434,Namespace:kube-system,Attempt:0,} returns sandbox id \"1f382b51de94969cefe89a79ae45b479960364bbb110bc90204b7ef14190711b\"" Sep 10 00:33:00.268026 kubelet[2139]: E0910 00:33:00.267993 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:00.273843 containerd[1458]: time="2025-09-10T00:33:00.273806292Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d75e6f6978d9f275ea19380916c9cccd,Namespace:kube-system,Attempt:0,} returns sandbox id \"6b8049b83542730a681d6ca820259c3dc0221ac1a261e3074292e8b908dc62db\"" Sep 10 00:33:00.274546 kubelet[2139]: E0910 00:33:00.274512 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:00.274605 containerd[1458]: time="2025-09-10T00:33:00.274511259Z" level=info msg="CreateContainer within sandbox \"1f382b51de94969cefe89a79ae45b479960364bbb110bc90204b7ef14190711b\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 10 00:33:00.279282 containerd[1458]: time="2025-09-10T00:33:00.279226357Z" level=info msg="CreateContainer within sandbox \"6b8049b83542730a681d6ca820259c3dc0221ac1a261e3074292e8b908dc62db\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 10 00:33:00.288562 systemd[1]: Started cri-containerd-160e7daa10d079d49200f971bd8d763880f282ad1580bf993ee1bac7a10dc7ed.scope - libcontainer container 160e7daa10d079d49200f971bd8d763880f282ad1580bf993ee1bac7a10dc7ed. Sep 10 00:33:00.299867 containerd[1458]: time="2025-09-10T00:33:00.299752876Z" level=info msg="CreateContainer within sandbox \"6b8049b83542730a681d6ca820259c3dc0221ac1a261e3074292e8b908dc62db\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"4b1ed221b639f79ce40dde424ab6becf9831a2017fc2bffd66f5e52fe9e0e669\"" Sep 10 00:33:00.300312 containerd[1458]: time="2025-09-10T00:33:00.300248915Z" level=info msg="StartContainer for \"4b1ed221b639f79ce40dde424ab6becf9831a2017fc2bffd66f5e52fe9e0e669\"" Sep 10 00:33:00.301550 containerd[1458]: time="2025-09-10T00:33:00.301523914Z" level=info msg="CreateContainer within sandbox \"1f382b51de94969cefe89a79ae45b479960364bbb110bc90204b7ef14190711b\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9c8be395af2f1b0faa3dc2c0c1fb1bb78bceb6ef3e283d0d688d516832fdc0f4\"" Sep 10 00:33:00.302436 containerd[1458]: time="2025-09-10T00:33:00.301834739Z" level=info msg="StartContainer for \"9c8be395af2f1b0faa3dc2c0c1fb1bb78bceb6ef3e283d0d688d516832fdc0f4\"" Sep 10 00:33:00.319305 kubelet[2139]: I0910 00:33:00.318927 2139 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:33:00.319305 kubelet[2139]: E0910 00:33:00.319250 2139 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.29:6443/api/v1/nodes\": dial tcp 10.0.0.29:6443: connect: connection refused" node="localhost" Sep 10 00:33:00.332561 systemd[1]: Started cri-containerd-9c8be395af2f1b0faa3dc2c0c1fb1bb78bceb6ef3e283d0d688d516832fdc0f4.scope - libcontainer container 9c8be395af2f1b0faa3dc2c0c1fb1bb78bceb6ef3e283d0d688d516832fdc0f4. Sep 10 00:33:00.336744 systemd[1]: Started cri-containerd-4b1ed221b639f79ce40dde424ab6becf9831a2017fc2bffd66f5e52fe9e0e669.scope - libcontainer container 4b1ed221b639f79ce40dde424ab6becf9831a2017fc2bffd66f5e52fe9e0e669. Sep 10 00:33:00.354618 containerd[1458]: time="2025-09-10T00:33:00.354571135Z" level=info msg="StartContainer for \"160e7daa10d079d49200f971bd8d763880f282ad1580bf993ee1bac7a10dc7ed\" returns successfully" Sep 10 00:33:00.385590 containerd[1458]: time="2025-09-10T00:33:00.384662434Z" level=info msg="StartContainer for \"9c8be395af2f1b0faa3dc2c0c1fb1bb78bceb6ef3e283d0d688d516832fdc0f4\" returns successfully" Sep 10 00:33:00.398226 containerd[1458]: time="2025-09-10T00:33:00.396565328Z" level=info msg="StartContainer for \"4b1ed221b639f79ce40dde424ab6becf9831a2017fc2bffd66f5e52fe9e0e669\" returns successfully" Sep 10 00:33:00.557126 kubelet[2139]: E0910 00:33:00.557080 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:33:00.557274 kubelet[2139]: E0910 00:33:00.557212 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:00.559287 kubelet[2139]: E0910 00:33:00.559244 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:33:00.559474 kubelet[2139]: E0910 00:33:00.559351 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:00.561641 kubelet[2139]: E0910 00:33:00.561616 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:33:00.561741 kubelet[2139]: E0910 00:33:00.561720 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:01.565064 kubelet[2139]: E0910 00:33:01.565018 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:33:01.565529 kubelet[2139]: E0910 00:33:01.565215 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:01.566220 kubelet[2139]: E0910 00:33:01.566185 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:33:01.566361 kubelet[2139]: E0910 00:33:01.566341 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:01.566402 kubelet[2139]: E0910 00:33:01.566371 2139 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 10 00:33:01.566589 kubelet[2139]: E0910 00:33:01.566561 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:02.092649 kubelet[2139]: I0910 00:33:02.088374 2139 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:33:02.276573 kubelet[2139]: E0910 00:33:02.276506 2139 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 10 00:33:02.367261 kubelet[2139]: I0910 00:33:02.365262 2139 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 10 00:33:02.367261 kubelet[2139]: E0910 00:33:02.365311 2139 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 10 00:33:02.375415 kubelet[2139]: E0910 00:33:02.375326 2139 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:33:02.475809 kubelet[2139]: E0910 00:33:02.475721 2139 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:33:02.575984 kubelet[2139]: E0910 00:33:02.575922 2139 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 10 00:33:02.627996 kubelet[2139]: I0910 00:33:02.627887 2139 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:02.634109 kubelet[2139]: E0910 00:33:02.634075 2139 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:02.634109 kubelet[2139]: I0910 00:33:02.634108 2139 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:02.635657 kubelet[2139]: E0910 00:33:02.635621 2139 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:02.635657 kubelet[2139]: I0910 00:33:02.635652 2139 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:02.637148 kubelet[2139]: E0910 00:33:02.637127 2139 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:02.997192 kubelet[2139]: I0910 00:33:02.997035 2139 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:02.999065 kubelet[2139]: E0910 00:33:02.999032 2139 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:02.999304 kubelet[2139]: E0910 00:33:02.999277 2139 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:03.512841 kubelet[2139]: I0910 00:33:03.512776 2139 apiserver.go:52] "Watching apiserver" Sep 10 00:33:03.524208 kubelet[2139]: I0910 00:33:03.524176 2139 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 10 00:33:04.141965 systemd[1]: Reloading requested from client PID 2426 ('systemctl') (unit session-7.scope)... Sep 10 00:33:04.141982 systemd[1]: Reloading... Sep 10 00:33:04.217463 zram_generator::config[2471]: No configuration found. Sep 10 00:33:04.331564 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 10 00:33:04.437434 systemd[1]: Reloading finished in 295 ms. Sep 10 00:33:04.485848 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:33:04.515459 systemd[1]: kubelet.service: Deactivated successfully. Sep 10 00:33:04.515847 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:33:04.515913 systemd[1]: kubelet.service: Consumed 1.119s CPU time, 131.1M memory peak, 0B memory swap peak. Sep 10 00:33:04.526703 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 10 00:33:04.693943 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 10 00:33:04.699871 (kubelet)[2510]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 10 00:33:04.742986 kubelet[2510]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:33:04.742986 kubelet[2510]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 10 00:33:04.742986 kubelet[2510]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 10 00:33:04.743443 kubelet[2510]: I0910 00:33:04.743011 2510 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 10 00:33:04.748976 kubelet[2510]: I0910 00:33:04.748939 2510 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 10 00:33:04.748976 kubelet[2510]: I0910 00:33:04.748967 2510 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 10 00:33:04.749208 kubelet[2510]: I0910 00:33:04.749184 2510 server.go:956] "Client rotation is on, will bootstrap in background" Sep 10 00:33:04.750317 kubelet[2510]: I0910 00:33:04.750291 2510 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 10 00:33:04.754567 kubelet[2510]: I0910 00:33:04.754549 2510 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 10 00:33:04.758693 kubelet[2510]: E0910 00:33:04.758643 2510 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 10 00:33:04.758693 kubelet[2510]: I0910 00:33:04.758684 2510 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 10 00:33:04.764224 kubelet[2510]: I0910 00:33:04.764195 2510 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 10 00:33:04.764484 kubelet[2510]: I0910 00:33:04.764433 2510 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 10 00:33:04.764691 kubelet[2510]: I0910 00:33:04.764472 2510 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 10 00:33:04.764691 kubelet[2510]: I0910 00:33:04.764691 2510 topology_manager.go:138] "Creating topology manager with none policy" Sep 10 00:33:04.764812 kubelet[2510]: I0910 00:33:04.764701 2510 container_manager_linux.go:303] "Creating device plugin manager" Sep 10 00:33:04.764812 kubelet[2510]: I0910 00:33:04.764768 2510 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:33:04.764962 kubelet[2510]: I0910 00:33:04.764946 2510 kubelet.go:480] "Attempting to sync node with API server" Sep 10 00:33:04.764962 kubelet[2510]: I0910 00:33:04.764962 2510 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 10 00:33:04.765024 kubelet[2510]: I0910 00:33:04.764987 2510 kubelet.go:386] "Adding apiserver pod source" Sep 10 00:33:04.768410 kubelet[2510]: I0910 00:33:04.766381 2510 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 10 00:33:04.768410 kubelet[2510]: I0910 00:33:04.767260 2510 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 10 00:33:04.768410 kubelet[2510]: I0910 00:33:04.767697 2510 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 10 00:33:04.774832 kubelet[2510]: I0910 00:33:04.774791 2510 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 10 00:33:04.774832 kubelet[2510]: I0910 00:33:04.774841 2510 server.go:1289] "Started kubelet" Sep 10 00:33:04.775352 kubelet[2510]: I0910 00:33:04.775267 2510 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 10 00:33:04.777120 kubelet[2510]: I0910 00:33:04.775770 2510 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 10 00:33:04.777420 kubelet[2510]: I0910 00:33:04.777345 2510 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 10 00:33:04.777550 kubelet[2510]: I0910 00:33:04.777533 2510 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 10 00:33:04.777740 kubelet[2510]: I0910 00:33:04.777717 2510 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 10 00:33:04.805321 kubelet[2510]: I0910 00:33:04.804524 2510 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 10 00:33:04.808803 kubelet[2510]: E0910 00:33:04.808772 2510 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 10 00:33:04.811027 kubelet[2510]: I0910 00:33:04.809331 2510 factory.go:223] Registration of the systemd container factory successfully Sep 10 00:33:04.811027 kubelet[2510]: I0910 00:33:04.809398 2510 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 10 00:33:04.811027 kubelet[2510]: I0910 00:33:04.809454 2510 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 10 00:33:04.811027 kubelet[2510]: I0910 00:33:04.809574 2510 reconciler.go:26] "Reconciler: start to sync state" Sep 10 00:33:04.811027 kubelet[2510]: I0910 00:33:04.810243 2510 server.go:317] "Adding debug handlers to kubelet server" Sep 10 00:33:04.811027 kubelet[2510]: I0910 00:33:04.810334 2510 factory.go:223] Registration of the containerd container factory successfully Sep 10 00:33:04.823364 kubelet[2510]: I0910 00:33:04.823324 2510 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 10 00:33:04.825363 kubelet[2510]: I0910 00:33:04.825338 2510 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 10 00:33:04.825363 kubelet[2510]: I0910 00:33:04.825360 2510 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 10 00:33:04.825363 kubelet[2510]: I0910 00:33:04.825397 2510 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 10 00:33:04.825500 kubelet[2510]: I0910 00:33:04.825407 2510 kubelet.go:2436] "Starting kubelet main sync loop" Sep 10 00:33:04.825500 kubelet[2510]: E0910 00:33:04.825475 2510 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 10 00:33:04.847183 kubelet[2510]: I0910 00:33:04.847142 2510 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 10 00:33:04.847183 kubelet[2510]: I0910 00:33:04.847170 2510 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 10 00:33:04.847183 kubelet[2510]: I0910 00:33:04.847191 2510 state_mem.go:36] "Initialized new in-memory state store" Sep 10 00:33:04.847381 kubelet[2510]: I0910 00:33:04.847320 2510 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 10 00:33:04.847381 kubelet[2510]: I0910 00:33:04.847331 2510 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 10 00:33:04.847381 kubelet[2510]: I0910 00:33:04.847348 2510 policy_none.go:49] "None policy: Start" Sep 10 00:33:04.847381 kubelet[2510]: I0910 00:33:04.847359 2510 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 10 00:33:04.847381 kubelet[2510]: I0910 00:33:04.847369 2510 state_mem.go:35] "Initializing new in-memory state store" Sep 10 00:33:04.847505 kubelet[2510]: I0910 00:33:04.847483 2510 state_mem.go:75] "Updated machine memory state" Sep 10 00:33:04.852095 kubelet[2510]: E0910 00:33:04.851959 2510 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 10 00:33:04.852183 kubelet[2510]: I0910 00:33:04.852168 2510 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 10 00:33:04.852217 kubelet[2510]: I0910 00:33:04.852182 2510 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 10 00:33:04.852427 kubelet[2510]: I0910 00:33:04.852413 2510 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 10 00:33:04.853339 kubelet[2510]: E0910 00:33:04.853298 2510 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 10 00:33:04.927348 kubelet[2510]: I0910 00:33:04.927044 2510 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:04.927348 kubelet[2510]: I0910 00:33:04.927149 2510 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:04.927348 kubelet[2510]: I0910 00:33:04.927211 2510 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:04.961359 kubelet[2510]: I0910 00:33:04.961278 2510 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 10 00:33:04.967366 kubelet[2510]: I0910 00:33:04.967320 2510 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 10 00:33:04.967514 kubelet[2510]: I0910 00:33:04.967412 2510 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 10 00:33:05.011198 kubelet[2510]: I0910 00:33:05.011136 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:05.011198 kubelet[2510]: I0910 00:33:05.011192 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d75e6f6978d9f275ea19380916c9cccd-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d75e6f6978d9f275ea19380916c9cccd\") " pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:05.011405 kubelet[2510]: I0910 00:33:05.011221 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/693363976a025a035618fa0cc32f4434-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"693363976a025a035618fa0cc32f4434\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:05.011405 kubelet[2510]: I0910 00:33:05.011240 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:05.011405 kubelet[2510]: I0910 00:33:05.011270 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:05.011405 kubelet[2510]: I0910 00:33:05.011353 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:05.011513 kubelet[2510]: I0910 00:33:05.011438 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/693363976a025a035618fa0cc32f4434-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"693363976a025a035618fa0cc32f4434\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:05.011513 kubelet[2510]: I0910 00:33:05.011480 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/693363976a025a035618fa0cc32f4434-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"693363976a025a035618fa0cc32f4434\") " pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:05.011559 kubelet[2510]: I0910 00:33:05.011533 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8de7187202bee21b84740a213836f615-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8de7187202bee21b84740a213836f615\") " pod="kube-system/kube-controller-manager-localhost" Sep 10 00:33:05.148999 sudo[2553]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 10 00:33:05.149431 sudo[2553]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 10 00:33:05.231768 kubelet[2510]: E0910 00:33:05.231640 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:05.234610 kubelet[2510]: E0910 00:33:05.234222 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:05.234610 kubelet[2510]: E0910 00:33:05.234561 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:05.728328 sudo[2553]: pam_unix(sudo:session): session closed for user root Sep 10 00:33:05.767122 kubelet[2510]: I0910 00:33:05.766959 2510 apiserver.go:52] "Watching apiserver" Sep 10 00:33:05.809603 kubelet[2510]: I0910 00:33:05.809562 2510 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 10 00:33:05.836161 kubelet[2510]: I0910 00:33:05.836138 2510 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:05.837122 kubelet[2510]: I0910 00:33:05.836470 2510 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:05.837122 kubelet[2510]: E0910 00:33:05.836639 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:05.842592 kubelet[2510]: E0910 00:33:05.842531 2510 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 10 00:33:05.842786 kubelet[2510]: E0910 00:33:05.842750 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:05.845626 kubelet[2510]: E0910 00:33:05.845583 2510 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 10 00:33:05.845841 kubelet[2510]: E0910 00:33:05.845736 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:05.856073 kubelet[2510]: I0910 00:33:05.855615 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.855578944 podStartE2EDuration="1.855578944s" podCreationTimestamp="2025-09-10 00:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:33:05.855477859 +0000 UTC m=+1.150946429" watchObservedRunningTime="2025-09-10 00:33:05.855578944 +0000 UTC m=+1.151047516" Sep 10 00:33:05.860848 kubelet[2510]: I0910 00:33:05.860714 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.86068584 podStartE2EDuration="1.86068584s" podCreationTimestamp="2025-09-10 00:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:33:05.86062014 +0000 UTC m=+1.156088710" watchObservedRunningTime="2025-09-10 00:33:05.86068584 +0000 UTC m=+1.156154410" Sep 10 00:33:05.866476 kubelet[2510]: I0910 00:33:05.866130 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.8661197619999998 podStartE2EDuration="1.866119762s" podCreationTimestamp="2025-09-10 00:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:33:05.865977195 +0000 UTC m=+1.161445765" watchObservedRunningTime="2025-09-10 00:33:05.866119762 +0000 UTC m=+1.161588322" Sep 10 00:33:06.837315 kubelet[2510]: E0910 00:33:06.837159 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:06.837825 kubelet[2510]: E0910 00:33:06.837340 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:06.894973 sudo[1637]: pam_unix(sudo:session): session closed for user root Sep 10 00:33:06.897534 sshd[1634]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:06.901828 systemd[1]: sshd@6-10.0.0.29:22-10.0.0.1:43326.service: Deactivated successfully. Sep 10 00:33:06.904138 systemd[1]: session-7.scope: Deactivated successfully. Sep 10 00:33:06.904354 systemd[1]: session-7.scope: Consumed 5.828s CPU time, 156.4M memory peak, 0B memory swap peak. Sep 10 00:33:06.905144 systemd-logind[1442]: Session 7 logged out. Waiting for processes to exit. Sep 10 00:33:06.906214 systemd-logind[1442]: Removed session 7. Sep 10 00:33:07.839414 kubelet[2510]: E0910 00:33:07.839358 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:11.021339 kubelet[2510]: I0910 00:33:11.021250 2510 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 10 00:33:11.022070 kubelet[2510]: I0910 00:33:11.021932 2510 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 10 00:33:11.022136 containerd[1458]: time="2025-09-10T00:33:11.021715874Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 10 00:33:12.125872 systemd[1]: Created slice kubepods-besteffort-pod6e2d7f75_3dad_41e7_9e66_83baceba9bb4.slice - libcontainer container kubepods-besteffort-pod6e2d7f75_3dad_41e7_9e66_83baceba9bb4.slice. Sep 10 00:33:12.138579 systemd[1]: Created slice kubepods-burstable-podadce3b8f_7730_4b8e_a3fc_beee0d18d7bd.slice - libcontainer container kubepods-burstable-podadce3b8f_7730_4b8e_a3fc_beee0d18d7bd.slice. Sep 10 00:33:12.154555 kubelet[2510]: I0910 00:33:12.154207 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6e2d7f75-3dad-41e7-9e66-83baceba9bb4-lib-modules\") pod \"kube-proxy-x9f9q\" (UID: \"6e2d7f75-3dad-41e7-9e66-83baceba9bb4\") " pod="kube-system/kube-proxy-x9f9q" Sep 10 00:33:12.154555 kubelet[2510]: I0910 00:33:12.154282 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944md\" (UniqueName: \"kubernetes.io/projected/6e2d7f75-3dad-41e7-9e66-83baceba9bb4-kube-api-access-944md\") pod \"kube-proxy-x9f9q\" (UID: \"6e2d7f75-3dad-41e7-9e66-83baceba9bb4\") " pod="kube-system/kube-proxy-x9f9q" Sep 10 00:33:12.154555 kubelet[2510]: I0910 00:33:12.154338 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-xtables-lock\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.154555 kubelet[2510]: I0910 00:33:12.154423 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6e2d7f75-3dad-41e7-9e66-83baceba9bb4-xtables-lock\") pod \"kube-proxy-x9f9q\" (UID: \"6e2d7f75-3dad-41e7-9e66-83baceba9bb4\") " pod="kube-system/kube-proxy-x9f9q" Sep 10 00:33:12.154555 kubelet[2510]: I0910 00:33:12.154450 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cni-path\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155071 kubelet[2510]: I0910 00:33:12.154482 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-config-path\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155071 kubelet[2510]: I0910 00:33:12.154507 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-net\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155071 kubelet[2510]: I0910 00:33:12.154529 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbf2m\" (UniqueName: \"kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-kube-api-access-kbf2m\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155071 kubelet[2510]: I0910 00:33:12.154560 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/6e2d7f75-3dad-41e7-9e66-83baceba9bb4-kube-proxy\") pod \"kube-proxy-x9f9q\" (UID: \"6e2d7f75-3dad-41e7-9e66-83baceba9bb4\") " pod="kube-system/kube-proxy-x9f9q" Sep 10 00:33:12.155071 kubelet[2510]: I0910 00:33:12.154615 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-run\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155071 kubelet[2510]: I0910 00:33:12.154669 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-bpf-maps\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155238 kubelet[2510]: I0910 00:33:12.154698 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-etc-cni-netd\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155238 kubelet[2510]: I0910 00:33:12.154722 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-lib-modules\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155238 kubelet[2510]: I0910 00:33:12.154745 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hostproc\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155238 kubelet[2510]: I0910 00:33:12.154763 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-cgroup\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155238 kubelet[2510]: I0910 00:33:12.154783 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-clustermesh-secrets\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155238 kubelet[2510]: I0910 00:33:12.154819 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-kernel\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.155377 kubelet[2510]: I0910 00:33:12.154866 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hubble-tls\") pod \"cilium-kmtv9\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " pod="kube-system/cilium-kmtv9" Sep 10 00:33:12.292721 systemd[1]: Created slice kubepods-besteffort-pod98239aeb_952f_4233_b412_f2be18ac8dd6.slice - libcontainer container kubepods-besteffort-pod98239aeb_952f_4233_b412_f2be18ac8dd6.slice. Sep 10 00:33:12.356717 kubelet[2510]: I0910 00:33:12.356681 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znglq\" (UniqueName: \"kubernetes.io/projected/98239aeb-952f-4233-b412-f2be18ac8dd6-kube-api-access-znglq\") pod \"cilium-operator-6c4d7847fc-72fcf\" (UID: \"98239aeb-952f-4233-b412-f2be18ac8dd6\") " pod="kube-system/cilium-operator-6c4d7847fc-72fcf" Sep 10 00:33:12.356717 kubelet[2510]: I0910 00:33:12.356714 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/98239aeb-952f-4233-b412-f2be18ac8dd6-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-72fcf\" (UID: \"98239aeb-952f-4233-b412-f2be18ac8dd6\") " pod="kube-system/cilium-operator-6c4d7847fc-72fcf" Sep 10 00:33:12.436004 kubelet[2510]: E0910 00:33:12.435822 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.436440 containerd[1458]: time="2025-09-10T00:33:12.436313661Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-x9f9q,Uid:6e2d7f75-3dad-41e7-9e66-83baceba9bb4,Namespace:kube-system,Attempt:0,}" Sep 10 00:33:12.442085 kubelet[2510]: E0910 00:33:12.441677 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.442433 containerd[1458]: time="2025-09-10T00:33:12.442394521Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kmtv9,Uid:adce3b8f-7730-4b8e-a3fc-beee0d18d7bd,Namespace:kube-system,Attempt:0,}" Sep 10 00:33:12.478912 containerd[1458]: time="2025-09-10T00:33:12.478677057Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:12.478912 containerd[1458]: time="2025-09-10T00:33:12.478721157Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:12.478912 containerd[1458]: time="2025-09-10T00:33:12.478768795Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:12.479197 containerd[1458]: time="2025-09-10T00:33:12.478882743Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:12.479197 containerd[1458]: time="2025-09-10T00:33:12.478632597Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:12.479197 containerd[1458]: time="2025-09-10T00:33:12.478699048Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:12.479197 containerd[1458]: time="2025-09-10T00:33:12.478714272Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:12.479197 containerd[1458]: time="2025-09-10T00:33:12.478816423Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:12.504553 systemd[1]: Started cri-containerd-d50245eb7708116931d748ba8d6ccfeec7d2e47a199d4681c94fa926b95064a1.scope - libcontainer container d50245eb7708116931d748ba8d6ccfeec7d2e47a199d4681c94fa926b95064a1. Sep 10 00:33:12.508169 systemd[1]: Started cri-containerd-147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff.scope - libcontainer container 147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff. Sep 10 00:33:12.533924 containerd[1458]: time="2025-09-10T00:33:12.533883830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-x9f9q,Uid:6e2d7f75-3dad-41e7-9e66-83baceba9bb4,Namespace:kube-system,Attempt:0,} returns sandbox id \"d50245eb7708116931d748ba8d6ccfeec7d2e47a199d4681c94fa926b95064a1\"" Sep 10 00:33:12.534920 kubelet[2510]: E0910 00:33:12.534895 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.540871 containerd[1458]: time="2025-09-10T00:33:12.540666769Z" level=info msg="CreateContainer within sandbox \"d50245eb7708116931d748ba8d6ccfeec7d2e47a199d4681c94fa926b95064a1\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 10 00:33:12.541880 containerd[1458]: time="2025-09-10T00:33:12.541841578Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kmtv9,Uid:adce3b8f-7730-4b8e-a3fc-beee0d18d7bd,Namespace:kube-system,Attempt:0,} returns sandbox id \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\"" Sep 10 00:33:12.543630 kubelet[2510]: E0910 00:33:12.543589 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.544940 containerd[1458]: time="2025-09-10T00:33:12.544907747Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 10 00:33:12.560515 containerd[1458]: time="2025-09-10T00:33:12.560452574Z" level=info msg="CreateContainer within sandbox \"d50245eb7708116931d748ba8d6ccfeec7d2e47a199d4681c94fa926b95064a1\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"81b82ebcb3f7a71e4ad129e0e80267c373237ff37217fc7cd2e9f0a16cae3f3f\"" Sep 10 00:33:12.561134 containerd[1458]: time="2025-09-10T00:33:12.561097645Z" level=info msg="StartContainer for \"81b82ebcb3f7a71e4ad129e0e80267c373237ff37217fc7cd2e9f0a16cae3f3f\"" Sep 10 00:33:12.595349 kubelet[2510]: E0910 00:33:12.595304 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.595989 containerd[1458]: time="2025-09-10T00:33:12.595796297Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-72fcf,Uid:98239aeb-952f-4233-b412-f2be18ac8dd6,Namespace:kube-system,Attempt:0,}" Sep 10 00:33:12.597954 systemd[1]: Started cri-containerd-81b82ebcb3f7a71e4ad129e0e80267c373237ff37217fc7cd2e9f0a16cae3f3f.scope - libcontainer container 81b82ebcb3f7a71e4ad129e0e80267c373237ff37217fc7cd2e9f0a16cae3f3f. Sep 10 00:33:12.671244 containerd[1458]: time="2025-09-10T00:33:12.671157609Z" level=info msg="StartContainer for \"81b82ebcb3f7a71e4ad129e0e80267c373237ff37217fc7cd2e9f0a16cae3f3f\" returns successfully" Sep 10 00:33:12.690169 containerd[1458]: time="2025-09-10T00:33:12.689765839Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:12.690169 containerd[1458]: time="2025-09-10T00:33:12.689879576Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:12.690169 containerd[1458]: time="2025-09-10T00:33:12.689903782Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:12.690169 containerd[1458]: time="2025-09-10T00:33:12.690069196Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:12.716533 systemd[1]: Started cri-containerd-c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa.scope - libcontainer container c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa. Sep 10 00:33:12.755966 update_engine[1445]: I20250910 00:33:12.755169 1445 update_attempter.cc:509] Updating boot flags... Sep 10 00:33:12.756503 containerd[1458]: time="2025-09-10T00:33:12.755534844Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-72fcf,Uid:98239aeb-952f-4233-b412-f2be18ac8dd6,Namespace:kube-system,Attempt:0,} returns sandbox id \"c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa\"" Sep 10 00:33:12.756583 kubelet[2510]: E0910 00:33:12.756317 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.801193 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (2800) Sep 10 00:33:12.848422 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 34 scanned by (udev-worker) (2806) Sep 10 00:33:12.851832 kubelet[2510]: E0910 00:33:12.851790 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:12.866832 kubelet[2510]: I0910 00:33:12.866762 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-x9f9q" podStartSLOduration=0.866741483 podStartE2EDuration="866.741483ms" podCreationTimestamp="2025-09-10 00:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:33:12.866555993 +0000 UTC m=+8.162024563" watchObservedRunningTime="2025-09-10 00:33:12.866741483 +0000 UTC m=+8.162210053" Sep 10 00:33:13.898447 kubelet[2510]: E0910 00:33:13.898370 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:14.455307 kubelet[2510]: E0910 00:33:14.452929 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:14.782043 kubelet[2510]: E0910 00:33:14.781898 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:14.864666 kubelet[2510]: E0910 00:33:14.864497 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:14.864666 kubelet[2510]: E0910 00:33:14.864572 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:14.864666 kubelet[2510]: E0910 00:33:14.864575 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:15.866027 kubelet[2510]: E0910 00:33:15.865995 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:15.866526 kubelet[2510]: E0910 00:33:15.866233 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:18.545932 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount849553843.mount: Deactivated successfully. Sep 10 00:33:22.276769 containerd[1458]: time="2025-09-10T00:33:22.276684541Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:33:22.285682 containerd[1458]: time="2025-09-10T00:33:22.285563858Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Sep 10 00:33:22.326234 containerd[1458]: time="2025-09-10T00:33:22.326166049Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:33:22.328381 containerd[1458]: time="2025-09-10T00:33:22.328320754Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 9.783372506s" Sep 10 00:33:22.328381 containerd[1458]: time="2025-09-10T00:33:22.328357652Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 10 00:33:22.329987 containerd[1458]: time="2025-09-10T00:33:22.329948702Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 10 00:33:22.431272 containerd[1458]: time="2025-09-10T00:33:22.431208936Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 10 00:33:22.779945 containerd[1458]: time="2025-09-10T00:33:22.779889513Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\"" Sep 10 00:33:22.780679 containerd[1458]: time="2025-09-10T00:33:22.780627827Z" level=info msg="StartContainer for \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\"" Sep 10 00:33:22.824723 systemd[1]: Started cri-containerd-c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819.scope - libcontainer container c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819. Sep 10 00:33:22.856996 containerd[1458]: time="2025-09-10T00:33:22.856948225Z" level=info msg="StartContainer for \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\" returns successfully" Sep 10 00:33:22.869341 systemd[1]: cri-containerd-c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819.scope: Deactivated successfully. Sep 10 00:33:23.078658 containerd[1458]: time="2025-09-10T00:33:23.075173144Z" level=info msg="shim disconnected" id=c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819 namespace=k8s.io Sep 10 00:33:23.078658 containerd[1458]: time="2025-09-10T00:33:23.078631585Z" level=warning msg="cleaning up after shim disconnected" id=c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819 namespace=k8s.io Sep 10 00:33:23.078658 containerd[1458]: time="2025-09-10T00:33:23.078642808Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:33:23.267526 kubelet[2510]: E0910 00:33:23.267486 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:23.273463 containerd[1458]: time="2025-09-10T00:33:23.273378544Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 10 00:33:23.287855 containerd[1458]: time="2025-09-10T00:33:23.287795366Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\"" Sep 10 00:33:23.292005 containerd[1458]: time="2025-09-10T00:33:23.291981472Z" level=info msg="StartContainer for \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\"" Sep 10 00:33:23.327565 systemd[1]: Started cri-containerd-e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4.scope - libcontainer container e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4. Sep 10 00:33:23.358145 containerd[1458]: time="2025-09-10T00:33:23.357986952Z" level=info msg="StartContainer for \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\" returns successfully" Sep 10 00:33:23.368806 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 10 00:33:23.369057 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:33:23.369143 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:33:23.374794 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 10 00:33:23.375013 systemd[1]: cri-containerd-e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4.scope: Deactivated successfully. Sep 10 00:33:23.399054 containerd[1458]: time="2025-09-10T00:33:23.398961453Z" level=info msg="shim disconnected" id=e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4 namespace=k8s.io Sep 10 00:33:23.399054 containerd[1458]: time="2025-09-10T00:33:23.399047193Z" level=warning msg="cleaning up after shim disconnected" id=e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4 namespace=k8s.io Sep 10 00:33:23.399054 containerd[1458]: time="2025-09-10T00:33:23.399057485Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:33:23.405494 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 10 00:33:23.776589 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819-rootfs.mount: Deactivated successfully. Sep 10 00:33:24.132760 containerd[1458]: time="2025-09-10T00:33:24.132686684Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:33:24.133648 containerd[1458]: time="2025-09-10T00:33:24.133497616Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Sep 10 00:33:24.134701 containerd[1458]: time="2025-09-10T00:33:24.134642860Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 10 00:33:24.136414 containerd[1458]: time="2025-09-10T00:33:24.136368241Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 1.806378312s" Sep 10 00:33:24.136461 containerd[1458]: time="2025-09-10T00:33:24.136418557Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 10 00:33:24.141420 containerd[1458]: time="2025-09-10T00:33:24.141358314Z" level=info msg="CreateContainer within sandbox \"c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 10 00:33:24.156675 containerd[1458]: time="2025-09-10T00:33:24.156612449Z" level=info msg="CreateContainer within sandbox \"c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\"" Sep 10 00:33:24.157605 containerd[1458]: time="2025-09-10T00:33:24.157577465Z" level=info msg="StartContainer for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\"" Sep 10 00:33:24.190618 systemd[1]: Started cri-containerd-2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716.scope - libcontainer container 2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716. Sep 10 00:33:24.222901 containerd[1458]: time="2025-09-10T00:33:24.222851352Z" level=info msg="StartContainer for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" returns successfully" Sep 10 00:33:24.271613 kubelet[2510]: E0910 00:33:24.271572 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:24.276372 kubelet[2510]: E0910 00:33:24.276321 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:24.284483 containerd[1458]: time="2025-09-10T00:33:24.284410272Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 10 00:33:24.358002 kubelet[2510]: I0910 00:33:24.357919 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-72fcf" podStartSLOduration=0.977559911 podStartE2EDuration="12.357900803s" podCreationTimestamp="2025-09-10 00:33:12 +0000 UTC" firstStartedPulling="2025-09-10 00:33:12.756823128 +0000 UTC m=+8.052291699" lastFinishedPulling="2025-09-10 00:33:24.137164021 +0000 UTC m=+19.432632591" observedRunningTime="2025-09-10 00:33:24.288912441 +0000 UTC m=+19.584381001" watchObservedRunningTime="2025-09-10 00:33:24.357900803 +0000 UTC m=+19.653369373" Sep 10 00:33:24.434482 containerd[1458]: time="2025-09-10T00:33:24.433643161Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\"" Sep 10 00:33:24.435761 containerd[1458]: time="2025-09-10T00:33:24.435032878Z" level=info msg="StartContainer for \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\"" Sep 10 00:33:24.495540 systemd[1]: Started cri-containerd-79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d.scope - libcontainer container 79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d. Sep 10 00:33:24.534928 systemd[1]: cri-containerd-79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d.scope: Deactivated successfully. Sep 10 00:33:24.539578 containerd[1458]: time="2025-09-10T00:33:24.539540598Z" level=info msg="StartContainer for \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\" returns successfully" Sep 10 00:33:24.567594 containerd[1458]: time="2025-09-10T00:33:24.567513619Z" level=info msg="shim disconnected" id=79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d namespace=k8s.io Sep 10 00:33:24.567594 containerd[1458]: time="2025-09-10T00:33:24.567579677Z" level=warning msg="cleaning up after shim disconnected" id=79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d namespace=k8s.io Sep 10 00:33:24.567594 containerd[1458]: time="2025-09-10T00:33:24.567591412Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:33:25.281138 kubelet[2510]: E0910 00:33:25.280218 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:25.281138 kubelet[2510]: E0910 00:33:25.280340 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:25.285873 containerd[1458]: time="2025-09-10T00:33:25.285829019Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 10 00:33:25.305556 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3469014545.mount: Deactivated successfully. Sep 10 00:33:25.307712 containerd[1458]: time="2025-09-10T00:33:25.307648041Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\"" Sep 10 00:33:25.308279 containerd[1458]: time="2025-09-10T00:33:25.308238816Z" level=info msg="StartContainer for \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\"" Sep 10 00:33:25.360717 systemd[1]: Started cri-containerd-2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266.scope - libcontainer container 2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266. Sep 10 00:33:25.389416 systemd[1]: cri-containerd-2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266.scope: Deactivated successfully. Sep 10 00:33:25.392731 containerd[1458]: time="2025-09-10T00:33:25.392513192Z" level=info msg="StartContainer for \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\" returns successfully" Sep 10 00:33:25.439090 containerd[1458]: time="2025-09-10T00:33:25.439016144Z" level=info msg="shim disconnected" id=2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266 namespace=k8s.io Sep 10 00:33:25.439090 containerd[1458]: time="2025-09-10T00:33:25.439081541Z" level=warning msg="cleaning up after shim disconnected" id=2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266 namespace=k8s.io Sep 10 00:33:25.439090 containerd[1458]: time="2025-09-10T00:33:25.439090931Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:33:25.776168 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266-rootfs.mount: Deactivated successfully. Sep 10 00:33:26.285730 kubelet[2510]: E0910 00:33:26.285692 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:26.290686 containerd[1458]: time="2025-09-10T00:33:26.290636603Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 10 00:33:26.307275 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount154145418.mount: Deactivated successfully. Sep 10 00:33:26.307668 containerd[1458]: time="2025-09-10T00:33:26.307342370Z" level=info msg="CreateContainer within sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\"" Sep 10 00:33:26.308014 containerd[1458]: time="2025-09-10T00:33:26.307969787Z" level=info msg="StartContainer for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\"" Sep 10 00:33:26.340539 systemd[1]: Started cri-containerd-574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157.scope - libcontainer container 574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157. Sep 10 00:33:26.376779 containerd[1458]: time="2025-09-10T00:33:26.376724950Z" level=info msg="StartContainer for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" returns successfully" Sep 10 00:33:26.556752 kubelet[2510]: I0910 00:33:26.556596 2510 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 10 00:33:26.628829 systemd[1]: Created slice kubepods-burstable-podb71e0fdc_4124_4b47_a639_abab5cb152df.slice - libcontainer container kubepods-burstable-podb71e0fdc_4124_4b47_a639_abab5cb152df.slice. Sep 10 00:33:26.636772 systemd[1]: Created slice kubepods-burstable-podc42dd0f0_4947_46ba_aa9b_ca67de218319.slice - libcontainer container kubepods-burstable-podc42dd0f0_4947_46ba_aa9b_ca67de218319.slice. Sep 10 00:33:26.683521 kubelet[2510]: I0910 00:33:26.683421 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm6dr\" (UniqueName: \"kubernetes.io/projected/b71e0fdc-4124-4b47-a639-abab5cb152df-kube-api-access-qm6dr\") pod \"coredns-674b8bbfcf-gwzcd\" (UID: \"b71e0fdc-4124-4b47-a639-abab5cb152df\") " pod="kube-system/coredns-674b8bbfcf-gwzcd" Sep 10 00:33:26.683521 kubelet[2510]: I0910 00:33:26.683456 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c42dd0f0-4947-46ba-aa9b-ca67de218319-config-volume\") pod \"coredns-674b8bbfcf-4q8px\" (UID: \"c42dd0f0-4947-46ba-aa9b-ca67de218319\") " pod="kube-system/coredns-674b8bbfcf-4q8px" Sep 10 00:33:26.683521 kubelet[2510]: I0910 00:33:26.683479 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lhkm\" (UniqueName: \"kubernetes.io/projected/c42dd0f0-4947-46ba-aa9b-ca67de218319-kube-api-access-4lhkm\") pod \"coredns-674b8bbfcf-4q8px\" (UID: \"c42dd0f0-4947-46ba-aa9b-ca67de218319\") " pod="kube-system/coredns-674b8bbfcf-4q8px" Sep 10 00:33:26.683521 kubelet[2510]: I0910 00:33:26.683499 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b71e0fdc-4124-4b47-a639-abab5cb152df-config-volume\") pod \"coredns-674b8bbfcf-gwzcd\" (UID: \"b71e0fdc-4124-4b47-a639-abab5cb152df\") " pod="kube-system/coredns-674b8bbfcf-gwzcd" Sep 10 00:33:26.933512 kubelet[2510]: E0910 00:33:26.932983 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:26.933800 containerd[1458]: time="2025-09-10T00:33:26.933767409Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gwzcd,Uid:b71e0fdc-4124-4b47-a639-abab5cb152df,Namespace:kube-system,Attempt:0,}" Sep 10 00:33:26.941041 kubelet[2510]: E0910 00:33:26.941007 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:26.941676 containerd[1458]: time="2025-09-10T00:33:26.941641822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-4q8px,Uid:c42dd0f0-4947-46ba-aa9b-ca67de218319,Namespace:kube-system,Attempt:0,}" Sep 10 00:33:27.291600 kubelet[2510]: E0910 00:33:27.291555 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:27.313021 kubelet[2510]: I0910 00:33:27.312931 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-kmtv9" podStartSLOduration=5.527613071 podStartE2EDuration="15.312907457s" podCreationTimestamp="2025-09-10 00:33:12 +0000 UTC" firstStartedPulling="2025-09-10 00:33:12.544396089 +0000 UTC m=+7.839864659" lastFinishedPulling="2025-09-10 00:33:22.329690475 +0000 UTC m=+17.625159045" observedRunningTime="2025-09-10 00:33:27.309785933 +0000 UTC m=+22.605254503" watchObservedRunningTime="2025-09-10 00:33:27.312907457 +0000 UTC m=+22.608376027" Sep 10 00:33:28.293371 kubelet[2510]: E0910 00:33:28.293332 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:28.751416 systemd-networkd[1408]: cilium_host: Link UP Sep 10 00:33:28.752733 systemd-networkd[1408]: cilium_net: Link UP Sep 10 00:33:28.752938 systemd-networkd[1408]: cilium_net: Gained carrier Sep 10 00:33:28.753118 systemd-networkd[1408]: cilium_host: Gained carrier Sep 10 00:33:28.863136 systemd-networkd[1408]: cilium_vxlan: Link UP Sep 10 00:33:28.863146 systemd-networkd[1408]: cilium_vxlan: Gained carrier Sep 10 00:33:28.891641 systemd-networkd[1408]: cilium_net: Gained IPv6LL Sep 10 00:33:28.963571 systemd-networkd[1408]: cilium_host: Gained IPv6LL Sep 10 00:33:29.088462 kernel: NET: Registered PF_ALG protocol family Sep 10 00:33:29.295862 kubelet[2510]: E0910 00:33:29.295820 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:29.523628 systemd[1]: Started sshd@7-10.0.0.29:22-10.0.0.1:42554.service - OpenSSH per-connection server daemon (10.0.0.1:42554). Sep 10 00:33:29.565720 sshd[3587]: Accepted publickey for core from 10.0.0.1 port 42554 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:29.566450 sshd[3587]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:29.570864 systemd-logind[1442]: New session 8 of user core. Sep 10 00:33:29.574666 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 10 00:33:29.731262 sshd[3587]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:29.735901 systemd[1]: sshd@7-10.0.0.29:22-10.0.0.1:42554.service: Deactivated successfully. Sep 10 00:33:29.738451 systemd[1]: session-8.scope: Deactivated successfully. Sep 10 00:33:29.739165 systemd-logind[1442]: Session 8 logged out. Waiting for processes to exit. Sep 10 00:33:29.740162 systemd-logind[1442]: Removed session 8. Sep 10 00:33:29.814323 systemd-networkd[1408]: lxc_health: Link UP Sep 10 00:33:29.821847 systemd-networkd[1408]: lxc_health: Gained carrier Sep 10 00:33:29.931569 systemd-networkd[1408]: cilium_vxlan: Gained IPv6LL Sep 10 00:33:30.020604 systemd-networkd[1408]: lxc8fbb4ce8ea6f: Link UP Sep 10 00:33:30.027414 kernel: eth0: renamed from tmp491dc Sep 10 00:33:30.036516 systemd-networkd[1408]: lxc8fbb4ce8ea6f: Gained carrier Sep 10 00:33:30.076941 systemd-networkd[1408]: lxc4f0c35c65b03: Link UP Sep 10 00:33:30.085409 kernel: eth0: renamed from tmp08883 Sep 10 00:33:30.091675 systemd-networkd[1408]: lxc4f0c35c65b03: Gained carrier Sep 10 00:33:30.445351 kubelet[2510]: E0910 00:33:30.445214 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:31.083663 systemd-networkd[1408]: lxc8fbb4ce8ea6f: Gained IPv6LL Sep 10 00:33:31.851671 systemd-networkd[1408]: lxc_health: Gained IPv6LL Sep 10 00:33:31.856556 kubelet[2510]: I0910 00:33:31.856455 2510 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 10 00:33:31.857122 kubelet[2510]: E0910 00:33:31.857080 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:31.979709 systemd-networkd[1408]: lxc4f0c35c65b03: Gained IPv6LL Sep 10 00:33:32.301928 kubelet[2510]: E0910 00:33:32.301880 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:33.562750 containerd[1458]: time="2025-09-10T00:33:33.562501656Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:33.562750 containerd[1458]: time="2025-09-10T00:33:33.562563241Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:33.562750 containerd[1458]: time="2025-09-10T00:33:33.562573953Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:33.562750 containerd[1458]: time="2025-09-10T00:33:33.562665699Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:33.571916 containerd[1458]: time="2025-09-10T00:33:33.571793312Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:33:33.572152 containerd[1458]: time="2025-09-10T00:33:33.571967517Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:33:33.572772 containerd[1458]: time="2025-09-10T00:33:33.572725370Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:33.572858 containerd[1458]: time="2025-09-10T00:33:33.572824993Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:33:33.586528 systemd[1]: Started cri-containerd-491dc5c996706ac08a12ea72b8b4c03e70e603addd24d4c0fe4ecc8267cae27b.scope - libcontainer container 491dc5c996706ac08a12ea72b8b4c03e70e603addd24d4c0fe4ecc8267cae27b. Sep 10 00:33:33.591959 systemd[1]: Started cri-containerd-08883b345bb748932e9f00ee73f7e6b57873c3dfc1a7bee556631efb981f5f93.scope - libcontainer container 08883b345bb748932e9f00ee73f7e6b57873c3dfc1a7bee556631efb981f5f93. Sep 10 00:33:33.598843 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 10 00:33:33.604758 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 10 00:33:33.626735 containerd[1458]: time="2025-09-10T00:33:33.626680714Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gwzcd,Uid:b71e0fdc-4124-4b47-a639-abab5cb152df,Namespace:kube-system,Attempt:0,} returns sandbox id \"491dc5c996706ac08a12ea72b8b4c03e70e603addd24d4c0fe4ecc8267cae27b\"" Sep 10 00:33:33.627347 kubelet[2510]: E0910 00:33:33.627317 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:33.632765 containerd[1458]: time="2025-09-10T00:33:33.632703973Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-4q8px,Uid:c42dd0f0-4947-46ba-aa9b-ca67de218319,Namespace:kube-system,Attempt:0,} returns sandbox id \"08883b345bb748932e9f00ee73f7e6b57873c3dfc1a7bee556631efb981f5f93\"" Sep 10 00:33:33.633370 kubelet[2510]: E0910 00:33:33.633325 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:33.634460 containerd[1458]: time="2025-09-10T00:33:33.634432443Z" level=info msg="CreateContainer within sandbox \"491dc5c996706ac08a12ea72b8b4c03e70e603addd24d4c0fe4ecc8267cae27b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 10 00:33:33.637959 containerd[1458]: time="2025-09-10T00:33:33.637915819Z" level=info msg="CreateContainer within sandbox \"08883b345bb748932e9f00ee73f7e6b57873c3dfc1a7bee556631efb981f5f93\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 10 00:33:33.653151 containerd[1458]: time="2025-09-10T00:33:33.653103295Z" level=info msg="CreateContainer within sandbox \"491dc5c996706ac08a12ea72b8b4c03e70e603addd24d4c0fe4ecc8267cae27b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"feb205795a37c55cad9170bad5cf75c0c2fca65c09893e24d4bf7674a7d766f7\"" Sep 10 00:33:33.653662 containerd[1458]: time="2025-09-10T00:33:33.653631641Z" level=info msg="StartContainer for \"feb205795a37c55cad9170bad5cf75c0c2fca65c09893e24d4bf7674a7d766f7\"" Sep 10 00:33:33.660406 containerd[1458]: time="2025-09-10T00:33:33.660224871Z" level=info msg="CreateContainer within sandbox \"08883b345bb748932e9f00ee73f7e6b57873c3dfc1a7bee556631efb981f5f93\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4679dc4b9e545684939c0177f92ccc26e2d41e0ceb2d414bee998d145f1cb4ed\"" Sep 10 00:33:33.661009 containerd[1458]: time="2025-09-10T00:33:33.660852299Z" level=info msg="StartContainer for \"4679dc4b9e545684939c0177f92ccc26e2d41e0ceb2d414bee998d145f1cb4ed\"" Sep 10 00:33:33.693679 systemd[1]: Started cri-containerd-feb205795a37c55cad9170bad5cf75c0c2fca65c09893e24d4bf7674a7d766f7.scope - libcontainer container feb205795a37c55cad9170bad5cf75c0c2fca65c09893e24d4bf7674a7d766f7. Sep 10 00:33:33.696602 systemd[1]: Started cri-containerd-4679dc4b9e545684939c0177f92ccc26e2d41e0ceb2d414bee998d145f1cb4ed.scope - libcontainer container 4679dc4b9e545684939c0177f92ccc26e2d41e0ceb2d414bee998d145f1cb4ed. Sep 10 00:33:33.731027 containerd[1458]: time="2025-09-10T00:33:33.730977679Z" level=info msg="StartContainer for \"feb205795a37c55cad9170bad5cf75c0c2fca65c09893e24d4bf7674a7d766f7\" returns successfully" Sep 10 00:33:33.739107 containerd[1458]: time="2025-09-10T00:33:33.739052507Z" level=info msg="StartContainer for \"4679dc4b9e545684939c0177f92ccc26e2d41e0ceb2d414bee998d145f1cb4ed\" returns successfully" Sep 10 00:33:34.306747 kubelet[2510]: E0910 00:33:34.306622 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:34.308704 kubelet[2510]: E0910 00:33:34.308663 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:34.327015 kubelet[2510]: I0910 00:33:34.326952 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-gwzcd" podStartSLOduration=22.326905019 podStartE2EDuration="22.326905019s" podCreationTimestamp="2025-09-10 00:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:33:34.326326915 +0000 UTC m=+29.621795485" watchObservedRunningTime="2025-09-10 00:33:34.326905019 +0000 UTC m=+29.622373589" Sep 10 00:33:34.327329 kubelet[2510]: I0910 00:33:34.327215 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-4q8px" podStartSLOduration=22.327209669 podStartE2EDuration="22.327209669s" podCreationTimestamp="2025-09-10 00:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:33:34.318000255 +0000 UTC m=+29.613468825" watchObservedRunningTime="2025-09-10 00:33:34.327209669 +0000 UTC m=+29.622678239" Sep 10 00:33:34.568234 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1956645381.mount: Deactivated successfully. Sep 10 00:33:34.743892 systemd[1]: Started sshd@8-10.0.0.29:22-10.0.0.1:35250.service - OpenSSH per-connection server daemon (10.0.0.1:35250). Sep 10 00:33:34.788602 sshd[3937]: Accepted publickey for core from 10.0.0.1 port 35250 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:34.790688 sshd[3937]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:34.795877 systemd-logind[1442]: New session 9 of user core. Sep 10 00:33:34.804545 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 10 00:33:35.081960 sshd[3937]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:35.085877 systemd[1]: sshd@8-10.0.0.29:22-10.0.0.1:35250.service: Deactivated successfully. Sep 10 00:33:35.088404 systemd[1]: session-9.scope: Deactivated successfully. Sep 10 00:33:35.089123 systemd-logind[1442]: Session 9 logged out. Waiting for processes to exit. Sep 10 00:33:35.090017 systemd-logind[1442]: Removed session 9. Sep 10 00:33:35.310354 kubelet[2510]: E0910 00:33:35.310313 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:35.310932 kubelet[2510]: E0910 00:33:35.310701 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:36.311884 kubelet[2510]: E0910 00:33:36.311849 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:33:40.094466 systemd[1]: Started sshd@9-10.0.0.29:22-10.0.0.1:54498.service - OpenSSH per-connection server daemon (10.0.0.1:54498). Sep 10 00:33:40.133968 sshd[3955]: Accepted publickey for core from 10.0.0.1 port 54498 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:40.135894 sshd[3955]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:40.140356 systemd-logind[1442]: New session 10 of user core. Sep 10 00:33:40.150501 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 10 00:33:40.264555 sshd[3955]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:40.268700 systemd[1]: sshd@9-10.0.0.29:22-10.0.0.1:54498.service: Deactivated successfully. Sep 10 00:33:40.271021 systemd[1]: session-10.scope: Deactivated successfully. Sep 10 00:33:40.271699 systemd-logind[1442]: Session 10 logged out. Waiting for processes to exit. Sep 10 00:33:40.272636 systemd-logind[1442]: Removed session 10. Sep 10 00:33:45.277546 systemd[1]: Started sshd@10-10.0.0.29:22-10.0.0.1:54508.service - OpenSSH per-connection server daemon (10.0.0.1:54508). Sep 10 00:33:45.317225 sshd[3973]: Accepted publickey for core from 10.0.0.1 port 54508 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:45.319105 sshd[3973]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:45.322974 systemd-logind[1442]: New session 11 of user core. Sep 10 00:33:45.329528 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 10 00:33:45.444249 sshd[3973]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:45.456369 systemd[1]: sshd@10-10.0.0.29:22-10.0.0.1:54508.service: Deactivated successfully. Sep 10 00:33:45.458423 systemd[1]: session-11.scope: Deactivated successfully. Sep 10 00:33:45.459921 systemd-logind[1442]: Session 11 logged out. Waiting for processes to exit. Sep 10 00:33:45.466675 systemd[1]: Started sshd@11-10.0.0.29:22-10.0.0.1:54514.service - OpenSSH per-connection server daemon (10.0.0.1:54514). Sep 10 00:33:45.467637 systemd-logind[1442]: Removed session 11. Sep 10 00:33:45.501567 sshd[3989]: Accepted publickey for core from 10.0.0.1 port 54514 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:45.503095 sshd[3989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:45.507061 systemd-logind[1442]: New session 12 of user core. Sep 10 00:33:45.515543 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 10 00:33:45.667859 sshd[3989]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:45.675910 systemd[1]: sshd@11-10.0.0.29:22-10.0.0.1:54514.service: Deactivated successfully. Sep 10 00:33:45.679064 systemd[1]: session-12.scope: Deactivated successfully. Sep 10 00:33:45.682697 systemd-logind[1442]: Session 12 logged out. Waiting for processes to exit. Sep 10 00:33:45.688835 systemd[1]: Started sshd@12-10.0.0.29:22-10.0.0.1:54520.service - OpenSSH per-connection server daemon (10.0.0.1:54520). Sep 10 00:33:45.690120 systemd-logind[1442]: Removed session 12. Sep 10 00:33:45.726093 sshd[4002]: Accepted publickey for core from 10.0.0.1 port 54520 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:45.727696 sshd[4002]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:45.731780 systemd-logind[1442]: New session 13 of user core. Sep 10 00:33:45.742531 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 10 00:33:45.853563 sshd[4002]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:45.857916 systemd[1]: sshd@12-10.0.0.29:22-10.0.0.1:54520.service: Deactivated successfully. Sep 10 00:33:45.859880 systemd[1]: session-13.scope: Deactivated successfully. Sep 10 00:33:45.860647 systemd-logind[1442]: Session 13 logged out. Waiting for processes to exit. Sep 10 00:33:45.861443 systemd-logind[1442]: Removed session 13. Sep 10 00:33:50.865648 systemd[1]: Started sshd@13-10.0.0.29:22-10.0.0.1:40170.service - OpenSSH per-connection server daemon (10.0.0.1:40170). Sep 10 00:33:50.905532 sshd[4016]: Accepted publickey for core from 10.0.0.1 port 40170 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:50.907288 sshd[4016]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:50.911826 systemd-logind[1442]: New session 14 of user core. Sep 10 00:33:50.923539 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 10 00:33:51.041036 sshd[4016]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:51.046865 systemd[1]: sshd@13-10.0.0.29:22-10.0.0.1:40170.service: Deactivated successfully. Sep 10 00:33:51.049357 systemd[1]: session-14.scope: Deactivated successfully. Sep 10 00:33:51.049990 systemd-logind[1442]: Session 14 logged out. Waiting for processes to exit. Sep 10 00:33:51.050919 systemd-logind[1442]: Removed session 14. Sep 10 00:33:56.052350 systemd[1]: Started sshd@14-10.0.0.29:22-10.0.0.1:40178.service - OpenSSH per-connection server daemon (10.0.0.1:40178). Sep 10 00:33:56.092307 sshd[4030]: Accepted publickey for core from 10.0.0.1 port 40178 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:33:56.094136 sshd[4030]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:33:56.098096 systemd-logind[1442]: New session 15 of user core. Sep 10 00:33:56.107539 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 10 00:33:56.239879 sshd[4030]: pam_unix(sshd:session): session closed for user core Sep 10 00:33:56.244336 systemd[1]: sshd@14-10.0.0.29:22-10.0.0.1:40178.service: Deactivated successfully. Sep 10 00:33:56.246665 systemd[1]: session-15.scope: Deactivated successfully. Sep 10 00:33:56.247314 systemd-logind[1442]: Session 15 logged out. Waiting for processes to exit. Sep 10 00:33:56.248244 systemd-logind[1442]: Removed session 15. Sep 10 00:34:01.253047 systemd[1]: Started sshd@15-10.0.0.29:22-10.0.0.1:57066.service - OpenSSH per-connection server daemon (10.0.0.1:57066). Sep 10 00:34:01.291973 sshd[4044]: Accepted publickey for core from 10.0.0.1 port 57066 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:01.293413 sshd[4044]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:01.297272 systemd-logind[1442]: New session 16 of user core. Sep 10 00:34:01.309523 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 10 00:34:01.423971 sshd[4044]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:01.436348 systemd[1]: sshd@15-10.0.0.29:22-10.0.0.1:57066.service: Deactivated successfully. Sep 10 00:34:01.438235 systemd[1]: session-16.scope: Deactivated successfully. Sep 10 00:34:01.439718 systemd-logind[1442]: Session 16 logged out. Waiting for processes to exit. Sep 10 00:34:01.444694 systemd[1]: Started sshd@16-10.0.0.29:22-10.0.0.1:57082.service - OpenSSH per-connection server daemon (10.0.0.1:57082). Sep 10 00:34:01.445628 systemd-logind[1442]: Removed session 16. Sep 10 00:34:01.484035 sshd[4058]: Accepted publickey for core from 10.0.0.1 port 57082 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:01.485612 sshd[4058]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:01.489511 systemd-logind[1442]: New session 17 of user core. Sep 10 00:34:01.496533 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 10 00:34:02.288184 sshd[4058]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:02.300592 systemd[1]: sshd@16-10.0.0.29:22-10.0.0.1:57082.service: Deactivated successfully. Sep 10 00:34:02.302460 systemd[1]: session-17.scope: Deactivated successfully. Sep 10 00:34:02.304364 systemd-logind[1442]: Session 17 logged out. Waiting for processes to exit. Sep 10 00:34:02.309945 systemd[1]: Started sshd@17-10.0.0.29:22-10.0.0.1:57096.service - OpenSSH per-connection server daemon (10.0.0.1:57096). Sep 10 00:34:02.310881 systemd-logind[1442]: Removed session 17. Sep 10 00:34:02.350837 sshd[4071]: Accepted publickey for core from 10.0.0.1 port 57096 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:02.352788 sshd[4071]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:02.357168 systemd-logind[1442]: New session 18 of user core. Sep 10 00:34:02.373707 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 10 00:34:02.869887 sshd[4071]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:02.882985 systemd[1]: sshd@17-10.0.0.29:22-10.0.0.1:57096.service: Deactivated successfully. Sep 10 00:34:02.884915 systemd[1]: session-18.scope: Deactivated successfully. Sep 10 00:34:02.889027 systemd-logind[1442]: Session 18 logged out. Waiting for processes to exit. Sep 10 00:34:02.895921 systemd[1]: Started sshd@18-10.0.0.29:22-10.0.0.1:57112.service - OpenSSH per-connection server daemon (10.0.0.1:57112). Sep 10 00:34:02.896966 systemd-logind[1442]: Removed session 18. Sep 10 00:34:02.934223 sshd[4091]: Accepted publickey for core from 10.0.0.1 port 57112 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:02.936054 sshd[4091]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:02.940113 systemd-logind[1442]: New session 19 of user core. Sep 10 00:34:02.947527 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 10 00:34:03.216413 sshd[4091]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:03.227642 systemd[1]: sshd@18-10.0.0.29:22-10.0.0.1:57112.service: Deactivated successfully. Sep 10 00:34:03.229753 systemd[1]: session-19.scope: Deactivated successfully. Sep 10 00:34:03.231849 systemd-logind[1442]: Session 19 logged out. Waiting for processes to exit. Sep 10 00:34:03.233760 systemd[1]: Started sshd@19-10.0.0.29:22-10.0.0.1:57116.service - OpenSSH per-connection server daemon (10.0.0.1:57116). Sep 10 00:34:03.235030 systemd-logind[1442]: Removed session 19. Sep 10 00:34:03.275144 sshd[4104]: Accepted publickey for core from 10.0.0.1 port 57116 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:03.277427 sshd[4104]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:03.282708 systemd-logind[1442]: New session 20 of user core. Sep 10 00:34:03.294596 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 10 00:34:03.414535 sshd[4104]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:03.418955 systemd[1]: sshd@19-10.0.0.29:22-10.0.0.1:57116.service: Deactivated successfully. Sep 10 00:34:03.421507 systemd[1]: session-20.scope: Deactivated successfully. Sep 10 00:34:03.422228 systemd-logind[1442]: Session 20 logged out. Waiting for processes to exit. Sep 10 00:34:03.423217 systemd-logind[1442]: Removed session 20. Sep 10 00:34:08.430629 systemd[1]: Started sshd@20-10.0.0.29:22-10.0.0.1:57122.service - OpenSSH per-connection server daemon (10.0.0.1:57122). Sep 10 00:34:08.469607 sshd[4121]: Accepted publickey for core from 10.0.0.1 port 57122 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:08.471112 sshd[4121]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:08.475105 systemd-logind[1442]: New session 21 of user core. Sep 10 00:34:08.482541 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 10 00:34:08.585351 sshd[4121]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:08.589485 systemd[1]: sshd@20-10.0.0.29:22-10.0.0.1:57122.service: Deactivated successfully. Sep 10 00:34:08.592026 systemd[1]: session-21.scope: Deactivated successfully. Sep 10 00:34:08.592787 systemd-logind[1442]: Session 21 logged out. Waiting for processes to exit. Sep 10 00:34:08.593745 systemd-logind[1442]: Removed session 21. Sep 10 00:34:13.604790 systemd[1]: Started sshd@21-10.0.0.29:22-10.0.0.1:40730.service - OpenSSH per-connection server daemon (10.0.0.1:40730). Sep 10 00:34:13.644865 sshd[4139]: Accepted publickey for core from 10.0.0.1 port 40730 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:13.646612 sshd[4139]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:13.650806 systemd-logind[1442]: New session 22 of user core. Sep 10 00:34:13.661532 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 10 00:34:14.018302 sshd[4139]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:14.022938 systemd[1]: sshd@21-10.0.0.29:22-10.0.0.1:40730.service: Deactivated successfully. Sep 10 00:34:14.025615 systemd[1]: session-22.scope: Deactivated successfully. Sep 10 00:34:14.026380 systemd-logind[1442]: Session 22 logged out. Waiting for processes to exit. Sep 10 00:34:14.027328 systemd-logind[1442]: Removed session 22. Sep 10 00:34:19.030906 systemd[1]: Started sshd@22-10.0.0.29:22-10.0.0.1:40742.service - OpenSSH per-connection server daemon (10.0.0.1:40742). Sep 10 00:34:19.074312 sshd[4153]: Accepted publickey for core from 10.0.0.1 port 40742 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:19.076287 sshd[4153]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:19.081004 systemd-logind[1442]: New session 23 of user core. Sep 10 00:34:19.090558 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 10 00:34:19.199275 sshd[4153]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:19.213562 systemd[1]: sshd@22-10.0.0.29:22-10.0.0.1:40742.service: Deactivated successfully. Sep 10 00:34:19.215655 systemd[1]: session-23.scope: Deactivated successfully. Sep 10 00:34:19.217250 systemd-logind[1442]: Session 23 logged out. Waiting for processes to exit. Sep 10 00:34:19.226710 systemd[1]: Started sshd@23-10.0.0.29:22-10.0.0.1:40744.service - OpenSSH per-connection server daemon (10.0.0.1:40744). Sep 10 00:34:19.227635 systemd-logind[1442]: Removed session 23. Sep 10 00:34:19.262829 sshd[4167]: Accepted publickey for core from 10.0.0.1 port 40744 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:19.264523 sshd[4167]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:19.268774 systemd-logind[1442]: New session 24 of user core. Sep 10 00:34:19.282571 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 10 00:34:20.714005 containerd[1458]: time="2025-09-10T00:34:20.713951522Z" level=info msg="StopContainer for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" with timeout 30 (s)" Sep 10 00:34:20.715784 containerd[1458]: time="2025-09-10T00:34:20.715752521Z" level=info msg="Stop container \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" with signal terminated" Sep 10 00:34:20.731527 systemd[1]: cri-containerd-2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716.scope: Deactivated successfully. Sep 10 00:34:20.751648 containerd[1458]: time="2025-09-10T00:34:20.751575751Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 10 00:34:20.754739 containerd[1458]: time="2025-09-10T00:34:20.754665960Z" level=info msg="StopContainer for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" with timeout 2 (s)" Sep 10 00:34:20.754975 containerd[1458]: time="2025-09-10T00:34:20.754956415Z" level=info msg="Stop container \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" with signal terminated" Sep 10 00:34:20.757094 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716-rootfs.mount: Deactivated successfully. Sep 10 00:34:20.762044 systemd-networkd[1408]: lxc_health: Link DOWN Sep 10 00:34:20.762113 systemd-networkd[1408]: lxc_health: Lost carrier Sep 10 00:34:20.767605 containerd[1458]: time="2025-09-10T00:34:20.767542779Z" level=info msg="shim disconnected" id=2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716 namespace=k8s.io Sep 10 00:34:20.767725 containerd[1458]: time="2025-09-10T00:34:20.767610035Z" level=warning msg="cleaning up after shim disconnected" id=2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716 namespace=k8s.io Sep 10 00:34:20.767725 containerd[1458]: time="2025-09-10T00:34:20.767623230Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:20.785124 systemd[1]: cri-containerd-574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157.scope: Deactivated successfully. Sep 10 00:34:20.785521 systemd[1]: cri-containerd-574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157.scope: Consumed 7.094s CPU time. Sep 10 00:34:20.813033 containerd[1458]: time="2025-09-10T00:34:20.812637504Z" level=info msg="StopContainer for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" returns successfully" Sep 10 00:34:20.814290 containerd[1458]: time="2025-09-10T00:34:20.814251472Z" level=info msg="StopPodSandbox for \"c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa\"" Sep 10 00:34:20.814375 containerd[1458]: time="2025-09-10T00:34:20.814318568Z" level=info msg="Container to stop \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:34:20.817202 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa-shm.mount: Deactivated successfully. Sep 10 00:34:20.832635 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157-rootfs.mount: Deactivated successfully. Sep 10 00:34:20.834007 systemd[1]: cri-containerd-c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa.scope: Deactivated successfully. Sep 10 00:34:20.844222 containerd[1458]: time="2025-09-10T00:34:20.844140750Z" level=info msg="shim disconnected" id=574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157 namespace=k8s.io Sep 10 00:34:20.844468 containerd[1458]: time="2025-09-10T00:34:20.844446293Z" level=warning msg="cleaning up after shim disconnected" id=574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157 namespace=k8s.io Sep 10 00:34:20.844540 containerd[1458]: time="2025-09-10T00:34:20.844523568Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:20.863882 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa-rootfs.mount: Deactivated successfully. Sep 10 00:34:20.865680 containerd[1458]: time="2025-09-10T00:34:20.865637150Z" level=info msg="StopContainer for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" returns successfully" Sep 10 00:34:20.866092 containerd[1458]: time="2025-09-10T00:34:20.866043513Z" level=info msg="shim disconnected" id=c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa namespace=k8s.io Sep 10 00:34:20.866143 containerd[1458]: time="2025-09-10T00:34:20.866093457Z" level=warning msg="cleaning up after shim disconnected" id=c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa namespace=k8s.io Sep 10 00:34:20.866143 containerd[1458]: time="2025-09-10T00:34:20.866106391Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:20.866523 containerd[1458]: time="2025-09-10T00:34:20.866500380Z" level=info msg="StopPodSandbox for \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\"" Sep 10 00:34:20.866574 containerd[1458]: time="2025-09-10T00:34:20.866534554Z" level=info msg="Container to stop \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:34:20.866574 containerd[1458]: time="2025-09-10T00:34:20.866551476Z" level=info msg="Container to stop \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:34:20.866574 containerd[1458]: time="2025-09-10T00:34:20.866563458Z" level=info msg="Container to stop \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:34:20.866704 containerd[1458]: time="2025-09-10T00:34:20.866575591Z" level=info msg="Container to stop \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:34:20.866704 containerd[1458]: time="2025-09-10T00:34:20.866587564Z" level=info msg="Container to stop \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 10 00:34:20.874428 systemd[1]: cri-containerd-147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff.scope: Deactivated successfully. Sep 10 00:34:20.902470 containerd[1458]: time="2025-09-10T00:34:20.902243921Z" level=info msg="TearDown network for sandbox \"c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa\" successfully" Sep 10 00:34:20.902470 containerd[1458]: time="2025-09-10T00:34:20.902298122Z" level=info msg="StopPodSandbox for \"c4ad74879f02310abb2475d115fc4a952771a8e2d83c77d858a1c9ccc99e65fa\" returns successfully" Sep 10 00:34:20.921806 containerd[1458]: time="2025-09-10T00:34:20.921736402Z" level=info msg="shim disconnected" id=147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff namespace=k8s.io Sep 10 00:34:20.921806 containerd[1458]: time="2025-09-10T00:34:20.921795112Z" level=warning msg="cleaning up after shim disconnected" id=147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff namespace=k8s.io Sep 10 00:34:20.921806 containerd[1458]: time="2025-09-10T00:34:20.921803668Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:20.948328 containerd[1458]: time="2025-09-10T00:34:20.948271737Z" level=info msg="TearDown network for sandbox \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" successfully" Sep 10 00:34:20.948328 containerd[1458]: time="2025-09-10T00:34:20.948309578Z" level=info msg="StopPodSandbox for \"147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff\" returns successfully" Sep 10 00:34:20.999907 kubelet[2510]: I0910 00:34:20.999748 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cni-path\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:20.999907 kubelet[2510]: I0910 00:34:20.999806 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-cgroup\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:20.999907 kubelet[2510]: I0910 00:34:20.999837 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbf2m\" (UniqueName: \"kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-kube-api-access-kbf2m\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:20.999907 kubelet[2510]: I0910 00:34:20.999867 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-config-path\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:20.999907 kubelet[2510]: I0910 00:34:20.999886 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-bpf-maps\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000513 kubelet[2510]: I0910 00:34:20.999893 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.000513 kubelet[2510]: I0910 00:34:20.999928 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.000513 kubelet[2510]: I0910 00:34:20.999905 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-etc-cni-netd\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000513 kubelet[2510]: I0910 00:34:20.999966 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-net\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000513 kubelet[2510]: I0910 00:34:20.999989 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-kernel\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000672 kubelet[2510]: I0910 00:34:21.000005 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/98239aeb-952f-4233-b412-f2be18ac8dd6-cilium-config-path\") pod \"98239aeb-952f-4233-b412-f2be18ac8dd6\" (UID: \"98239aeb-952f-4233-b412-f2be18ac8dd6\") " Sep 10 00:34:21.000672 kubelet[2510]: I0910 00:34:21.000019 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znglq\" (UniqueName: \"kubernetes.io/projected/98239aeb-952f-4233-b412-f2be18ac8dd6-kube-api-access-znglq\") pod \"98239aeb-952f-4233-b412-f2be18ac8dd6\" (UID: \"98239aeb-952f-4233-b412-f2be18ac8dd6\") " Sep 10 00:34:21.000672 kubelet[2510]: I0910 00:34:21.000033 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-run\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000672 kubelet[2510]: I0910 00:34:21.000045 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-lib-modules\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000672 kubelet[2510]: I0910 00:34:21.000057 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hostproc\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000672 kubelet[2510]: I0910 00:34:21.000072 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-clustermesh-secrets\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000851 kubelet[2510]: I0910 00:34:21.000085 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hubble-tls\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000851 kubelet[2510]: I0910 00:34:21.000098 2510 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-xtables-lock\") pod \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\" (UID: \"adce3b8f-7730-4b8e-a3fc-beee0d18d7bd\") " Sep 10 00:34:21.000851 kubelet[2510]: I0910 00:34:21.000129 2510 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.000851 kubelet[2510]: I0910 00:34:21.000137 2510 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.000851 kubelet[2510]: I0910 00:34:21.000154 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.000851 kubelet[2510]: I0910 00:34:21.000171 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.000992 kubelet[2510]: I0910 00:34:21.000192 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.002822 kubelet[2510]: I0910 00:34:21.002462 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cni-path" (OuterVolumeSpecName: "cni-path") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.004715 kubelet[2510]: I0910 00:34:21.004681 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-kube-api-access-kbf2m" (OuterVolumeSpecName: "kube-api-access-kbf2m") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "kube-api-access-kbf2m". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 10 00:34:21.005522 kubelet[2510]: I0910 00:34:21.004866 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.005522 kubelet[2510]: I0910 00:34:21.004909 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hostproc" (OuterVolumeSpecName: "hostproc") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.005522 kubelet[2510]: I0910 00:34:21.004933 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.005522 kubelet[2510]: I0910 00:34:21.004955 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 10 00:34:21.005804 kubelet[2510]: I0910 00:34:21.005780 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98239aeb-952f-4233-b412-f2be18ac8dd6-kube-api-access-znglq" (OuterVolumeSpecName: "kube-api-access-znglq") pod "98239aeb-952f-4233-b412-f2be18ac8dd6" (UID: "98239aeb-952f-4233-b412-f2be18ac8dd6"). InnerVolumeSpecName "kube-api-access-znglq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 10 00:34:21.009118 kubelet[2510]: I0910 00:34:21.009066 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98239aeb-952f-4233-b412-f2be18ac8dd6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "98239aeb-952f-4233-b412-f2be18ac8dd6" (UID: "98239aeb-952f-4233-b412-f2be18ac8dd6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 10 00:34:21.009988 kubelet[2510]: I0910 00:34:21.009965 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 10 00:34:21.010069 kubelet[2510]: I0910 00:34:21.010007 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 10 00:34:21.010917 kubelet[2510]: I0910 00:34:21.010890 2510 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" (UID: "adce3b8f-7730-4b8e-a3fc-beee0d18d7bd"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 10 00:34:21.100460 kubelet[2510]: I0910 00:34:21.100415 2510 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kbf2m\" (UniqueName: \"kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-kube-api-access-kbf2m\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100460 kubelet[2510]: I0910 00:34:21.100453 2510 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100505 2510 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100520 2510 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100531 2510 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100540 2510 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/98239aeb-952f-4233-b412-f2be18ac8dd6-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100549 2510 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-znglq\" (UniqueName: \"kubernetes.io/projected/98239aeb-952f-4233-b412-f2be18ac8dd6-kube-api-access-znglq\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100560 2510 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100570 2510 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100653 kubelet[2510]: I0910 00:34:21.100580 2510 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100840 kubelet[2510]: I0910 00:34:21.100590 2510 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100840 kubelet[2510]: I0910 00:34:21.100598 2510 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100840 kubelet[2510]: I0910 00:34:21.100605 2510 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.100840 kubelet[2510]: I0910 00:34:21.100612 2510 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 10 00:34:21.416639 kubelet[2510]: I0910 00:34:21.416572 2510 scope.go:117] "RemoveContainer" containerID="2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716" Sep 10 00:34:21.419733 containerd[1458]: time="2025-09-10T00:34:21.419686179Z" level=info msg="RemoveContainer for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\"" Sep 10 00:34:21.423014 systemd[1]: Removed slice kubepods-besteffort-pod98239aeb_952f_4233_b412_f2be18ac8dd6.slice - libcontainer container kubepods-besteffort-pod98239aeb_952f_4233_b412_f2be18ac8dd6.slice. Sep 10 00:34:21.425892 systemd[1]: Removed slice kubepods-burstable-podadce3b8f_7730_4b8e_a3fc_beee0d18d7bd.slice - libcontainer container kubepods-burstable-podadce3b8f_7730_4b8e_a3fc_beee0d18d7bd.slice. Sep 10 00:34:21.425976 systemd[1]: kubepods-burstable-podadce3b8f_7730_4b8e_a3fc_beee0d18d7bd.slice: Consumed 7.205s CPU time. Sep 10 00:34:21.429717 containerd[1458]: time="2025-09-10T00:34:21.429684830Z" level=info msg="RemoveContainer for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" returns successfully" Sep 10 00:34:21.429959 kubelet[2510]: I0910 00:34:21.429938 2510 scope.go:117] "RemoveContainer" containerID="2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716" Sep 10 00:34:21.433231 containerd[1458]: time="2025-09-10T00:34:21.433148676Z" level=error msg="ContainerStatus for \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\": not found" Sep 10 00:34:21.433443 kubelet[2510]: E0910 00:34:21.433347 2510 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\": not found" containerID="2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716" Sep 10 00:34:21.433501 kubelet[2510]: I0910 00:34:21.433424 2510 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716"} err="failed to get container status \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\": rpc error: code = NotFound desc = an error occurred when try to find container \"2a83783038a39a641fcb8c74d133f736d7d0714baab2e2c6ffc7a5eed0f1a716\": not found" Sep 10 00:34:21.433501 kubelet[2510]: I0910 00:34:21.433478 2510 scope.go:117] "RemoveContainer" containerID="574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157" Sep 10 00:34:21.434920 containerd[1458]: time="2025-09-10T00:34:21.434873977Z" level=info msg="RemoveContainer for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\"" Sep 10 00:34:21.440077 containerd[1458]: time="2025-09-10T00:34:21.440011837Z" level=info msg="RemoveContainer for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" returns successfully" Sep 10 00:34:21.440315 kubelet[2510]: I0910 00:34:21.440270 2510 scope.go:117] "RemoveContainer" containerID="2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266" Sep 10 00:34:21.442358 containerd[1458]: time="2025-09-10T00:34:21.442328228Z" level=info msg="RemoveContainer for \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\"" Sep 10 00:34:21.446326 containerd[1458]: time="2025-09-10T00:34:21.446284800Z" level=info msg="RemoveContainer for \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\" returns successfully" Sep 10 00:34:21.446537 kubelet[2510]: I0910 00:34:21.446507 2510 scope.go:117] "RemoveContainer" containerID="79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d" Sep 10 00:34:21.447609 containerd[1458]: time="2025-09-10T00:34:21.447562630Z" level=info msg="RemoveContainer for \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\"" Sep 10 00:34:21.451056 containerd[1458]: time="2025-09-10T00:34:21.451010225Z" level=info msg="RemoveContainer for \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\" returns successfully" Sep 10 00:34:21.451229 kubelet[2510]: I0910 00:34:21.451197 2510 scope.go:117] "RemoveContainer" containerID="e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4" Sep 10 00:34:21.452319 containerd[1458]: time="2025-09-10T00:34:21.452288437Z" level=info msg="RemoveContainer for \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\"" Sep 10 00:34:21.455543 containerd[1458]: time="2025-09-10T00:34:21.455500389Z" level=info msg="RemoveContainer for \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\" returns successfully" Sep 10 00:34:21.455696 kubelet[2510]: I0910 00:34:21.455638 2510 scope.go:117] "RemoveContainer" containerID="c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819" Sep 10 00:34:21.456536 containerd[1458]: time="2025-09-10T00:34:21.456509335Z" level=info msg="RemoveContainer for \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\"" Sep 10 00:34:21.459666 containerd[1458]: time="2025-09-10T00:34:21.459639945Z" level=info msg="RemoveContainer for \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\" returns successfully" Sep 10 00:34:21.459852 kubelet[2510]: I0910 00:34:21.459819 2510 scope.go:117] "RemoveContainer" containerID="574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157" Sep 10 00:34:21.460076 containerd[1458]: time="2025-09-10T00:34:21.460013917Z" level=error msg="ContainerStatus for \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\": not found" Sep 10 00:34:21.460183 kubelet[2510]: E0910 00:34:21.460146 2510 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\": not found" containerID="574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157" Sep 10 00:34:21.460226 kubelet[2510]: I0910 00:34:21.460194 2510 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157"} err="failed to get container status \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\": rpc error: code = NotFound desc = an error occurred when try to find container \"574b7591de4c25a8054843a10ea80bae207e8b6f0b801ec719eb772bde21a157\": not found" Sep 10 00:34:21.460226 kubelet[2510]: I0910 00:34:21.460222 2510 scope.go:117] "RemoveContainer" containerID="2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266" Sep 10 00:34:21.460513 containerd[1458]: time="2025-09-10T00:34:21.460457420Z" level=error msg="ContainerStatus for \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\": not found" Sep 10 00:34:21.460655 kubelet[2510]: E0910 00:34:21.460625 2510 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\": not found" containerID="2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266" Sep 10 00:34:21.460716 kubelet[2510]: I0910 00:34:21.460658 2510 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266"} err="failed to get container status \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\": rpc error: code = NotFound desc = an error occurred when try to find container \"2605804ac7cf5da3729114a8527aecf74951d71576a37314ede4d80877ccc266\": not found" Sep 10 00:34:21.460716 kubelet[2510]: I0910 00:34:21.460679 2510 scope.go:117] "RemoveContainer" containerID="79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d" Sep 10 00:34:21.460865 containerd[1458]: time="2025-09-10T00:34:21.460832475Z" level=error msg="ContainerStatus for \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\": not found" Sep 10 00:34:21.460991 kubelet[2510]: E0910 00:34:21.460964 2510 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\": not found" containerID="79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d" Sep 10 00:34:21.461026 kubelet[2510]: I0910 00:34:21.460996 2510 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d"} err="failed to get container status \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\": rpc error: code = NotFound desc = an error occurred when try to find container \"79a38f6c265c9d26f823e6f3903c69fbde6830b9a9f52202abecf70c5512789d\": not found" Sep 10 00:34:21.461026 kubelet[2510]: I0910 00:34:21.461018 2510 scope.go:117] "RemoveContainer" containerID="e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4" Sep 10 00:34:21.461242 containerd[1458]: time="2025-09-10T00:34:21.461205135Z" level=error msg="ContainerStatus for \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\": not found" Sep 10 00:34:21.461353 kubelet[2510]: E0910 00:34:21.461331 2510 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\": not found" containerID="e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4" Sep 10 00:34:21.461412 kubelet[2510]: I0910 00:34:21.461359 2510 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4"} err="failed to get container status \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\": rpc error: code = NotFound desc = an error occurred when try to find container \"e27af97106f61983efd3faa3448971170d8f90903f6eb3795395af48339c3ce4\": not found" Sep 10 00:34:21.461412 kubelet[2510]: I0910 00:34:21.461374 2510 scope.go:117] "RemoveContainer" containerID="c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819" Sep 10 00:34:21.461611 containerd[1458]: time="2025-09-10T00:34:21.461577063Z" level=error msg="ContainerStatus for \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\": not found" Sep 10 00:34:21.461760 kubelet[2510]: E0910 00:34:21.461731 2510 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\": not found" containerID="c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819" Sep 10 00:34:21.461760 kubelet[2510]: I0910 00:34:21.461752 2510 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819"} err="failed to get container status \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\": rpc error: code = NotFound desc = an error occurred when try to find container \"c2e49bb5c019013cd8ac696c35304e625bdfc5e621ff60ee3c72fa397508f819\": not found" Sep 10 00:34:21.723479 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff-rootfs.mount: Deactivated successfully. Sep 10 00:34:21.723593 systemd[1]: var-lib-kubelet-pods-98239aeb\x2d952f\x2d4233\x2db412\x2df2be18ac8dd6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dznglq.mount: Deactivated successfully. Sep 10 00:34:21.723684 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-147f9826ec98460903afad116139aef9a1f88258c2c689fa7d643aab0eadf2ff-shm.mount: Deactivated successfully. Sep 10 00:34:21.723764 systemd[1]: var-lib-kubelet-pods-adce3b8f\x2d7730\x2d4b8e\x2da3fc\x2dbeee0d18d7bd-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkbf2m.mount: Deactivated successfully. Sep 10 00:34:21.723847 systemd[1]: var-lib-kubelet-pods-adce3b8f\x2d7730\x2d4b8e\x2da3fc\x2dbeee0d18d7bd-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 10 00:34:21.723921 systemd[1]: var-lib-kubelet-pods-adce3b8f\x2d7730\x2d4b8e\x2da3fc\x2dbeee0d18d7bd-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 10 00:34:22.726543 sshd[4167]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:22.736178 systemd[1]: sshd@23-10.0.0.29:22-10.0.0.1:40744.service: Deactivated successfully. Sep 10 00:34:22.738001 systemd[1]: session-24.scope: Deactivated successfully. Sep 10 00:34:22.739565 systemd-logind[1442]: Session 24 logged out. Waiting for processes to exit. Sep 10 00:34:22.740828 systemd[1]: Started sshd@24-10.0.0.29:22-10.0.0.1:41530.service - OpenSSH per-connection server daemon (10.0.0.1:41530). Sep 10 00:34:22.741653 systemd-logind[1442]: Removed session 24. Sep 10 00:34:22.782725 sshd[4330]: Accepted publickey for core from 10.0.0.1 port 41530 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:22.784253 sshd[4330]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:22.787948 systemd-logind[1442]: New session 25 of user core. Sep 10 00:34:22.798514 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 10 00:34:22.828682 kubelet[2510]: I0910 00:34:22.828628 2510 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98239aeb-952f-4233-b412-f2be18ac8dd6" path="/var/lib/kubelet/pods/98239aeb-952f-4233-b412-f2be18ac8dd6/volumes" Sep 10 00:34:22.829411 kubelet[2510]: I0910 00:34:22.829354 2510 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adce3b8f-7730-4b8e-a3fc-beee0d18d7bd" path="/var/lib/kubelet/pods/adce3b8f-7730-4b8e-a3fc-beee0d18d7bd/volumes" Sep 10 00:34:23.448799 sshd[4330]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:23.461768 systemd[1]: sshd@24-10.0.0.29:22-10.0.0.1:41530.service: Deactivated successfully. Sep 10 00:34:23.463923 systemd[1]: session-25.scope: Deactivated successfully. Sep 10 00:34:23.467710 systemd-logind[1442]: Session 25 logged out. Waiting for processes to exit. Sep 10 00:34:23.479036 systemd[1]: Started sshd@25-10.0.0.29:22-10.0.0.1:41538.service - OpenSSH per-connection server daemon (10.0.0.1:41538). Sep 10 00:34:23.481695 systemd-logind[1442]: Removed session 25. Sep 10 00:34:23.490478 systemd[1]: Created slice kubepods-burstable-podf27b1a9d_874a_4f49_b390_17ba39afe082.slice - libcontainer container kubepods-burstable-podf27b1a9d_874a_4f49_b390_17ba39afe082.slice. Sep 10 00:34:23.518075 sshd[4343]: Accepted publickey for core from 10.0.0.1 port 41538 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:23.519757 sshd[4343]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:23.523579 systemd-logind[1442]: New session 26 of user core. Sep 10 00:34:23.533518 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 10 00:34:23.583839 sshd[4343]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:23.591186 systemd[1]: sshd@25-10.0.0.29:22-10.0.0.1:41538.service: Deactivated successfully. Sep 10 00:34:23.593010 systemd[1]: session-26.scope: Deactivated successfully. Sep 10 00:34:23.594509 systemd-logind[1442]: Session 26 logged out. Waiting for processes to exit. Sep 10 00:34:23.604666 systemd[1]: Started sshd@26-10.0.0.29:22-10.0.0.1:41552.service - OpenSSH per-connection server daemon (10.0.0.1:41552). Sep 10 00:34:23.605545 systemd-logind[1442]: Removed session 26. Sep 10 00:34:23.615376 kubelet[2510]: I0910 00:34:23.615336 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-host-proc-sys-net\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615376 kubelet[2510]: I0910 00:34:23.615373 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-etc-cni-netd\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615543 kubelet[2510]: I0910 00:34:23.615412 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-host-proc-sys-kernel\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615543 kubelet[2510]: I0910 00:34:23.615438 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f27b1a9d-874a-4f49-b390-17ba39afe082-hubble-tls\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615543 kubelet[2510]: I0910 00:34:23.615457 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-hostproc\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615543 kubelet[2510]: I0910 00:34:23.615472 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-cilium-cgroup\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615543 kubelet[2510]: I0910 00:34:23.615502 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-xtables-lock\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615543 kubelet[2510]: I0910 00:34:23.615518 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sgst\" (UniqueName: \"kubernetes.io/projected/f27b1a9d-874a-4f49-b390-17ba39afe082-kube-api-access-4sgst\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615681 kubelet[2510]: I0910 00:34:23.615535 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f27b1a9d-874a-4f49-b390-17ba39afe082-clustermesh-secrets\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615681 kubelet[2510]: I0910 00:34:23.615550 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-bpf-maps\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615681 kubelet[2510]: I0910 00:34:23.615618 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-cni-path\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615681 kubelet[2510]: I0910 00:34:23.615659 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-lib-modules\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615773 kubelet[2510]: I0910 00:34:23.615682 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f27b1a9d-874a-4f49-b390-17ba39afe082-cilium-config-path\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615773 kubelet[2510]: I0910 00:34:23.615698 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f27b1a9d-874a-4f49-b390-17ba39afe082-cilium-run\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.615773 kubelet[2510]: I0910 00:34:23.615728 2510 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/f27b1a9d-874a-4f49-b390-17ba39afe082-cilium-ipsec-secrets\") pod \"cilium-vtdrm\" (UID: \"f27b1a9d-874a-4f49-b390-17ba39afe082\") " pod="kube-system/cilium-vtdrm" Sep 10 00:34:23.642099 sshd[4351]: Accepted publickey for core from 10.0.0.1 port 41552 ssh2: RSA SHA256:yotFPVH/8pVol0IcCMTpL4axYdSEk1J0cKg1+3rpd1s Sep 10 00:34:23.643798 sshd[4351]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 10 00:34:23.647777 systemd-logind[1442]: New session 27 of user core. Sep 10 00:34:23.658538 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 10 00:34:23.793741 kubelet[2510]: E0910 00:34:23.793700 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:23.794642 containerd[1458]: time="2025-09-10T00:34:23.794259659Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vtdrm,Uid:f27b1a9d-874a-4f49-b390-17ba39afe082,Namespace:kube-system,Attempt:0,}" Sep 10 00:34:23.816496 containerd[1458]: time="2025-09-10T00:34:23.816407940Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 10 00:34:23.816496 containerd[1458]: time="2025-09-10T00:34:23.816465097Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 10 00:34:23.816496 containerd[1458]: time="2025-09-10T00:34:23.816478592Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:34:23.816668 containerd[1458]: time="2025-09-10T00:34:23.816557681Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 10 00:34:23.838523 systemd[1]: Started cri-containerd-d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87.scope - libcontainer container d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87. Sep 10 00:34:23.860582 containerd[1458]: time="2025-09-10T00:34:23.860489396Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vtdrm,Uid:f27b1a9d-874a-4f49-b390-17ba39afe082,Namespace:kube-system,Attempt:0,} returns sandbox id \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\"" Sep 10 00:34:23.861418 kubelet[2510]: E0910 00:34:23.861394 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:23.867068 containerd[1458]: time="2025-09-10T00:34:23.867026001Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 10 00:34:23.878608 containerd[1458]: time="2025-09-10T00:34:23.878554731Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22\"" Sep 10 00:34:23.879199 containerd[1458]: time="2025-09-10T00:34:23.879115517Z" level=info msg="StartContainer for \"477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22\"" Sep 10 00:34:23.916649 systemd[1]: Started cri-containerd-477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22.scope - libcontainer container 477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22. Sep 10 00:34:23.943671 containerd[1458]: time="2025-09-10T00:34:23.943614527Z" level=info msg="StartContainer for \"477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22\" returns successfully" Sep 10 00:34:23.954638 systemd[1]: cri-containerd-477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22.scope: Deactivated successfully. Sep 10 00:34:23.985587 containerd[1458]: time="2025-09-10T00:34:23.985511852Z" level=info msg="shim disconnected" id=477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22 namespace=k8s.io Sep 10 00:34:23.985587 containerd[1458]: time="2025-09-10T00:34:23.985580111Z" level=warning msg="cleaning up after shim disconnected" id=477ecacdc72db7fe768f0502bbc505f7cc3fedd1072b93bcb447278d007bce22 namespace=k8s.io Sep 10 00:34:23.985587 containerd[1458]: time="2025-09-10T00:34:23.985592805Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:24.439547 kubelet[2510]: E0910 00:34:24.439512 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:24.445956 containerd[1458]: time="2025-09-10T00:34:24.445905566Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 10 00:34:24.460167 containerd[1458]: time="2025-09-10T00:34:24.460088680Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f\"" Sep 10 00:34:24.460681 containerd[1458]: time="2025-09-10T00:34:24.460644608Z" level=info msg="StartContainer for \"8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f\"" Sep 10 00:34:24.492539 systemd[1]: Started cri-containerd-8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f.scope - libcontainer container 8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f. Sep 10 00:34:24.522085 containerd[1458]: time="2025-09-10T00:34:24.522031899Z" level=info msg="StartContainer for \"8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f\" returns successfully" Sep 10 00:34:24.528951 systemd[1]: cri-containerd-8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f.scope: Deactivated successfully. Sep 10 00:34:24.551010 containerd[1458]: time="2025-09-10T00:34:24.550921655Z" level=info msg="shim disconnected" id=8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f namespace=k8s.io Sep 10 00:34:24.551010 containerd[1458]: time="2025-09-10T00:34:24.550986879Z" level=warning msg="cleaning up after shim disconnected" id=8396cec9aa93d1ab29d2d6dcd0a89e5e5d5c495275b18967f1a1a6043771aa0f namespace=k8s.io Sep 10 00:34:24.551010 containerd[1458]: time="2025-09-10T00:34:24.550998170Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:24.881166 kubelet[2510]: E0910 00:34:24.881090 2510 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 10 00:34:25.442918 kubelet[2510]: E0910 00:34:25.442886 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:25.447195 containerd[1458]: time="2025-09-10T00:34:25.447150281Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 10 00:34:25.462236 containerd[1458]: time="2025-09-10T00:34:25.462192263Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1\"" Sep 10 00:34:25.462738 containerd[1458]: time="2025-09-10T00:34:25.462708486Z" level=info msg="StartContainer for \"5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1\"" Sep 10 00:34:25.492518 systemd[1]: Started cri-containerd-5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1.scope - libcontainer container 5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1. Sep 10 00:34:25.523266 containerd[1458]: time="2025-09-10T00:34:25.523223749Z" level=info msg="StartContainer for \"5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1\" returns successfully" Sep 10 00:34:25.525528 systemd[1]: cri-containerd-5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1.scope: Deactivated successfully. Sep 10 00:34:25.550537 containerd[1458]: time="2025-09-10T00:34:25.550455492Z" level=info msg="shim disconnected" id=5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1 namespace=k8s.io Sep 10 00:34:25.550537 containerd[1458]: time="2025-09-10T00:34:25.550510896Z" level=warning msg="cleaning up after shim disconnected" id=5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1 namespace=k8s.io Sep 10 00:34:25.550537 containerd[1458]: time="2025-09-10T00:34:25.550519722Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:25.721023 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5b41f79dd4993e4cb0c2a960a1906611f8d4cc781c69e0e24e08bf143b175cf1-rootfs.mount: Deactivated successfully. Sep 10 00:34:26.446717 kubelet[2510]: E0910 00:34:26.446681 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:26.585872 containerd[1458]: time="2025-09-10T00:34:26.585819874Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 10 00:34:26.631069 kubelet[2510]: I0910 00:34:26.629575 2510 setters.go:618] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-10T00:34:26Z","lastTransitionTime":"2025-09-10T00:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 10 00:34:26.631250 containerd[1458]: time="2025-09-10T00:34:26.630603935Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf\"" Sep 10 00:34:26.632192 containerd[1458]: time="2025-09-10T00:34:26.632145284Z" level=info msg="StartContainer for \"60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf\"" Sep 10 00:34:26.665574 systemd[1]: Started cri-containerd-60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf.scope - libcontainer container 60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf. Sep 10 00:34:26.691161 systemd[1]: cri-containerd-60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf.scope: Deactivated successfully. Sep 10 00:34:26.693164 containerd[1458]: time="2025-09-10T00:34:26.693121175Z" level=info msg="StartContainer for \"60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf\" returns successfully" Sep 10 00:34:26.715960 containerd[1458]: time="2025-09-10T00:34:26.715812687Z" level=info msg="shim disconnected" id=60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf namespace=k8s.io Sep 10 00:34:26.715960 containerd[1458]: time="2025-09-10T00:34:26.715867220Z" level=warning msg="cleaning up after shim disconnected" id=60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf namespace=k8s.io Sep 10 00:34:26.715960 containerd[1458]: time="2025-09-10T00:34:26.715875315Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 10 00:34:26.720816 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-60ce46adbcb4ed388c0478c6fe0e906605ce65022b4cf01121f96fbaf408baaf-rootfs.mount: Deactivated successfully. Sep 10 00:34:27.450406 kubelet[2510]: E0910 00:34:27.450347 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:27.456440 containerd[1458]: time="2025-09-10T00:34:27.456374908Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 10 00:34:27.472220 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2320183576.mount: Deactivated successfully. Sep 10 00:34:27.473275 containerd[1458]: time="2025-09-10T00:34:27.473230361Z" level=info msg="CreateContainer within sandbox \"d63f01e3cf5f71a8af740622e4271b3cf4ce3932d8ba3e6121aae9baf0535e87\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"720c1b4930ae3003bb9095341dd2ee8291c31c4728188c770be3ac7e80e1da42\"" Sep 10 00:34:27.473859 containerd[1458]: time="2025-09-10T00:34:27.473821127Z" level=info msg="StartContainer for \"720c1b4930ae3003bb9095341dd2ee8291c31c4728188c770be3ac7e80e1da42\"" Sep 10 00:34:27.504525 systemd[1]: Started cri-containerd-720c1b4930ae3003bb9095341dd2ee8291c31c4728188c770be3ac7e80e1da42.scope - libcontainer container 720c1b4930ae3003bb9095341dd2ee8291c31c4728188c770be3ac7e80e1da42. Sep 10 00:34:27.537714 containerd[1458]: time="2025-09-10T00:34:27.537657901Z" level=info msg="StartContainer for \"720c1b4930ae3003bb9095341dd2ee8291c31c4728188c770be3ac7e80e1da42\" returns successfully" Sep 10 00:34:27.976418 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 10 00:34:28.455712 kubelet[2510]: E0910 00:34:28.455676 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:29.795255 kubelet[2510]: E0910 00:34:29.795212 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:31.146524 systemd-networkd[1408]: lxc_health: Link UP Sep 10 00:34:31.157837 systemd-networkd[1408]: lxc_health: Gained carrier Sep 10 00:34:31.798440 kubelet[2510]: E0910 00:34:31.796459 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:31.823886 kubelet[2510]: I0910 00:34:31.823787 2510 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-vtdrm" podStartSLOduration=8.823770332 podStartE2EDuration="8.823770332s" podCreationTimestamp="2025-09-10 00:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-10 00:34:28.469123988 +0000 UTC m=+83.764592578" watchObservedRunningTime="2025-09-10 00:34:31.823770332 +0000 UTC m=+87.119238902" Sep 10 00:34:32.331691 systemd-networkd[1408]: lxc_health: Gained IPv6LL Sep 10 00:34:32.464070 kubelet[2510]: E0910 00:34:32.464021 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:33.465685 kubelet[2510]: E0910 00:34:33.465646 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:34.826676 kubelet[2510]: E0910 00:34:34.826635 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:36.372455 sshd[4351]: pam_unix(sshd:session): session closed for user core Sep 10 00:34:36.376710 systemd[1]: sshd@26-10.0.0.29:22-10.0.0.1:41552.service: Deactivated successfully. Sep 10 00:34:36.379555 systemd[1]: session-27.scope: Deactivated successfully. Sep 10 00:34:36.380241 systemd-logind[1442]: Session 27 logged out. Waiting for processes to exit. Sep 10 00:34:36.381374 systemd-logind[1442]: Removed session 27. Sep 10 00:34:36.826126 kubelet[2510]: E0910 00:34:36.826078 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 10 00:34:37.826461 kubelet[2510]: E0910 00:34:37.826373 2510 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"