Sep 9 00:28:12.999445 kernel: Linux version 6.6.104-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Sep 8 22:41:17 -00 2025 Sep 9 00:28:12.999475 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=99a67175ee6aabbc03a22dabcade16d60ad192b31c4118a259bf1f24bbfa2d29 Sep 9 00:28:12.999491 kernel: BIOS-provided physical RAM map: Sep 9 00:28:12.999500 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 9 00:28:12.999509 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 9 00:28:12.999518 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 9 00:28:12.999528 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Sep 9 00:28:12.999538 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Sep 9 00:28:12.999547 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Sep 9 00:28:12.999559 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Sep 9 00:28:12.999569 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 9 00:28:12.999577 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 9 00:28:12.999590 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Sep 9 00:28:12.999600 kernel: NX (Execute Disable) protection: active Sep 9 00:28:12.999611 kernel: APIC: Static calls initialized Sep 9 00:28:12.999628 kernel: SMBIOS 2.8 present. Sep 9 00:28:12.999638 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Sep 9 00:28:12.999648 kernel: Hypervisor detected: KVM Sep 9 00:28:12.999658 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 9 00:28:12.999667 kernel: kvm-clock: using sched offset of 3035526191 cycles Sep 9 00:28:12.999678 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 9 00:28:12.999688 kernel: tsc: Detected 2794.750 MHz processor Sep 9 00:28:12.999698 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 9 00:28:12.999709 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 9 00:28:12.999719 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Sep 9 00:28:12.999733 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Sep 9 00:28:12.999743 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 9 00:28:12.999753 kernel: Using GB pages for direct mapping Sep 9 00:28:12.999763 kernel: ACPI: Early table checksum verification disabled Sep 9 00:28:12.999773 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Sep 9 00:28:12.999784 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999794 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999804 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999817 kernel: ACPI: FACS 0x000000009CFE0000 000040 Sep 9 00:28:12.999827 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999837 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999848 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999858 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 9 00:28:12.999868 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Sep 9 00:28:12.999878 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Sep 9 00:28:12.999894 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Sep 9 00:28:12.999907 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Sep 9 00:28:12.999918 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Sep 9 00:28:12.999928 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Sep 9 00:28:12.999939 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Sep 9 00:28:12.999953 kernel: No NUMA configuration found Sep 9 00:28:12.999964 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Sep 9 00:28:12.999974 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Sep 9 00:28:12.999988 kernel: Zone ranges: Sep 9 00:28:12.999999 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 9 00:28:13.000009 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Sep 9 00:28:13.000020 kernel: Normal empty Sep 9 00:28:13.000031 kernel: Movable zone start for each node Sep 9 00:28:13.000041 kernel: Early memory node ranges Sep 9 00:28:13.000051 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 9 00:28:13.000062 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Sep 9 00:28:13.000091 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Sep 9 00:28:13.000135 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 9 00:28:13.000160 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 9 00:28:13.000171 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Sep 9 00:28:13.000182 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 9 00:28:13.000192 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 9 00:28:13.000203 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 9 00:28:13.000213 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 9 00:28:13.000224 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 9 00:28:13.000235 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 9 00:28:13.000249 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 9 00:28:13.000260 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 9 00:28:13.000270 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 9 00:28:13.000281 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 9 00:28:13.000291 kernel: TSC deadline timer available Sep 9 00:28:13.000302 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Sep 9 00:28:13.000312 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 9 00:28:13.000323 kernel: kvm-guest: KVM setup pv remote TLB flush Sep 9 00:28:13.000337 kernel: kvm-guest: setup PV sched yield Sep 9 00:28:13.000351 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Sep 9 00:28:13.000362 kernel: Booting paravirtualized kernel on KVM Sep 9 00:28:13.000373 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 9 00:28:13.000383 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Sep 9 00:28:13.000394 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u524288 Sep 9 00:28:13.000405 kernel: pcpu-alloc: s197160 r8192 d32216 u524288 alloc=1*2097152 Sep 9 00:28:13.000415 kernel: pcpu-alloc: [0] 0 1 2 3 Sep 9 00:28:13.000425 kernel: kvm-guest: PV spinlocks enabled Sep 9 00:28:13.000436 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 9 00:28:13.000451 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=99a67175ee6aabbc03a22dabcade16d60ad192b31c4118a259bf1f24bbfa2d29 Sep 9 00:28:13.000463 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 9 00:28:13.000473 kernel: random: crng init done Sep 9 00:28:13.000484 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 9 00:28:13.000494 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 9 00:28:13.000505 kernel: Fallback order for Node 0: 0 Sep 9 00:28:13.000516 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Sep 9 00:28:13.000526 kernel: Policy zone: DMA32 Sep 9 00:28:13.000537 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 9 00:28:13.000552 kernel: Memory: 2434592K/2571752K available (12288K kernel code, 2293K rwdata, 22744K rodata, 42880K init, 2316K bss, 136900K reserved, 0K cma-reserved) Sep 9 00:28:13.000562 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 9 00:28:13.000573 kernel: ftrace: allocating 37969 entries in 149 pages Sep 9 00:28:13.000584 kernel: ftrace: allocated 149 pages with 4 groups Sep 9 00:28:13.000594 kernel: Dynamic Preempt: voluntary Sep 9 00:28:13.000605 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 9 00:28:13.000616 kernel: rcu: RCU event tracing is enabled. Sep 9 00:28:13.000627 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 9 00:28:13.000637 kernel: Trampoline variant of Tasks RCU enabled. Sep 9 00:28:13.000651 kernel: Rude variant of Tasks RCU enabled. Sep 9 00:28:13.000662 kernel: Tracing variant of Tasks RCU enabled. Sep 9 00:28:13.000672 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 9 00:28:13.000686 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 9 00:28:13.000697 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Sep 9 00:28:13.000708 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 9 00:28:13.000718 kernel: Console: colour VGA+ 80x25 Sep 9 00:28:13.000728 kernel: printk: console [ttyS0] enabled Sep 9 00:28:13.000738 kernel: ACPI: Core revision 20230628 Sep 9 00:28:13.000753 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 9 00:28:13.000763 kernel: APIC: Switch to symmetric I/O mode setup Sep 9 00:28:13.000774 kernel: x2apic enabled Sep 9 00:28:13.000784 kernel: APIC: Switched APIC routing to: physical x2apic Sep 9 00:28:13.000795 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Sep 9 00:28:13.000806 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Sep 9 00:28:13.000816 kernel: kvm-guest: setup PV IPIs Sep 9 00:28:13.000840 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 9 00:28:13.000852 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Sep 9 00:28:13.000863 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Sep 9 00:28:13.000874 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 9 00:28:13.000889 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Sep 9 00:28:13.000899 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Sep 9 00:28:13.000909 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 9 00:28:13.000920 kernel: Spectre V2 : Mitigation: Retpolines Sep 9 00:28:13.000931 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 9 00:28:13.000946 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Sep 9 00:28:13.000957 kernel: active return thunk: retbleed_return_thunk Sep 9 00:28:13.000972 kernel: RETBleed: Mitigation: untrained return thunk Sep 9 00:28:13.000983 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 9 00:28:13.000994 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 9 00:28:13.001005 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Sep 9 00:28:13.001016 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Sep 9 00:28:13.001027 kernel: active return thunk: srso_return_thunk Sep 9 00:28:13.001042 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Sep 9 00:28:13.001053 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 9 00:28:13.001064 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 9 00:28:13.001111 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 9 00:28:13.001123 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 9 00:28:13.001133 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Sep 9 00:28:13.001152 kernel: Freeing SMP alternatives memory: 32K Sep 9 00:28:13.001162 kernel: pid_max: default: 32768 minimum: 301 Sep 9 00:28:13.001181 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 9 00:28:13.001199 kernel: landlock: Up and running. Sep 9 00:28:13.001210 kernel: SELinux: Initializing. Sep 9 00:28:13.001221 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 9 00:28:13.001232 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 9 00:28:13.001243 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Sep 9 00:28:13.001255 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 9 00:28:13.001266 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 9 00:28:13.001277 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 9 00:28:13.001292 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Sep 9 00:28:13.001307 kernel: ... version: 0 Sep 9 00:28:13.001318 kernel: ... bit width: 48 Sep 9 00:28:13.001329 kernel: ... generic registers: 6 Sep 9 00:28:13.001340 kernel: ... value mask: 0000ffffffffffff Sep 9 00:28:13.001350 kernel: ... max period: 00007fffffffffff Sep 9 00:28:13.001360 kernel: ... fixed-purpose events: 0 Sep 9 00:28:13.001370 kernel: ... event mask: 000000000000003f Sep 9 00:28:13.001381 kernel: signal: max sigframe size: 1776 Sep 9 00:28:13.001391 kernel: rcu: Hierarchical SRCU implementation. Sep 9 00:28:13.001406 kernel: rcu: Max phase no-delay instances is 400. Sep 9 00:28:13.001418 kernel: smp: Bringing up secondary CPUs ... Sep 9 00:28:13.001429 kernel: smpboot: x86: Booting SMP configuration: Sep 9 00:28:13.001440 kernel: .... node #0, CPUs: #1 #2 #3 Sep 9 00:28:13.001451 kernel: smp: Brought up 1 node, 4 CPUs Sep 9 00:28:13.001462 kernel: smpboot: Max logical packages: 1 Sep 9 00:28:13.001473 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Sep 9 00:28:13.001484 kernel: devtmpfs: initialized Sep 9 00:28:13.001495 kernel: x86/mm: Memory block size: 128MB Sep 9 00:28:13.001507 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 9 00:28:13.001522 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 9 00:28:13.001533 kernel: pinctrl core: initialized pinctrl subsystem Sep 9 00:28:13.001544 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 9 00:28:13.001555 kernel: audit: initializing netlink subsys (disabled) Sep 9 00:28:13.001567 kernel: audit: type=2000 audit(1757377691.270:1): state=initialized audit_enabled=0 res=1 Sep 9 00:28:13.001578 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 9 00:28:13.001589 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 9 00:28:13.001600 kernel: cpuidle: using governor menu Sep 9 00:28:13.001611 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 9 00:28:13.001625 kernel: dca service started, version 1.12.1 Sep 9 00:28:13.001637 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Sep 9 00:28:13.001648 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Sep 9 00:28:13.001659 kernel: PCI: Using configuration type 1 for base access Sep 9 00:28:13.001671 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 9 00:28:13.001682 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 9 00:28:13.001693 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 9 00:28:13.001705 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 9 00:28:13.001720 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 9 00:28:13.001731 kernel: ACPI: Added _OSI(Module Device) Sep 9 00:28:13.001742 kernel: ACPI: Added _OSI(Processor Device) Sep 9 00:28:13.001753 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 9 00:28:13.001764 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 9 00:28:13.001775 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 9 00:28:13.001786 kernel: ACPI: Interpreter enabled Sep 9 00:28:13.001797 kernel: ACPI: PM: (supports S0 S3 S5) Sep 9 00:28:13.001808 kernel: ACPI: Using IOAPIC for interrupt routing Sep 9 00:28:13.001819 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 9 00:28:13.001834 kernel: PCI: Using E820 reservations for host bridge windows Sep 9 00:28:13.001845 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 9 00:28:13.001856 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 9 00:28:13.002192 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 9 00:28:13.002382 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Sep 9 00:28:13.002552 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Sep 9 00:28:13.002568 kernel: PCI host bridge to bus 0000:00 Sep 9 00:28:13.002759 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 9 00:28:13.002919 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 9 00:28:13.003093 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 9 00:28:13.003266 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Sep 9 00:28:13.003420 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Sep 9 00:28:13.003579 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Sep 9 00:28:13.003731 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 9 00:28:13.004182 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Sep 9 00:28:13.004383 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Sep 9 00:28:13.004555 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Sep 9 00:28:13.004724 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Sep 9 00:28:13.004889 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Sep 9 00:28:13.005056 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 9 00:28:13.005322 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Sep 9 00:28:13.005493 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Sep 9 00:28:13.005659 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Sep 9 00:28:13.005823 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Sep 9 00:28:13.006006 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Sep 9 00:28:13.006208 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Sep 9 00:28:13.006376 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Sep 9 00:28:13.006571 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Sep 9 00:28:13.006778 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Sep 9 00:28:13.006975 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Sep 9 00:28:13.007167 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Sep 9 00:28:13.007332 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Sep 9 00:28:13.007489 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Sep 9 00:28:13.007663 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Sep 9 00:28:13.007829 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 9 00:28:13.008006 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Sep 9 00:28:13.008243 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Sep 9 00:28:13.008401 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Sep 9 00:28:13.008570 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Sep 9 00:28:13.008725 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Sep 9 00:28:13.008745 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 9 00:28:13.008757 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 9 00:28:13.008769 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 9 00:28:13.008780 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 9 00:28:13.008791 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 9 00:28:13.008802 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 9 00:28:13.008813 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 9 00:28:13.008824 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 9 00:28:13.008835 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 9 00:28:13.008849 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 9 00:28:13.008860 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 9 00:28:13.008872 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 9 00:28:13.008883 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 9 00:28:13.008894 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 9 00:28:13.008905 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 9 00:28:13.008916 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 9 00:28:13.008927 kernel: iommu: Default domain type: Translated Sep 9 00:28:13.008938 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 9 00:28:13.008952 kernel: PCI: Using ACPI for IRQ routing Sep 9 00:28:13.008964 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 9 00:28:13.008975 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 9 00:28:13.008986 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Sep 9 00:28:13.009183 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 9 00:28:13.009340 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 9 00:28:13.009494 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 9 00:28:13.009508 kernel: vgaarb: loaded Sep 9 00:28:13.009520 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 9 00:28:13.009536 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 9 00:28:13.009547 kernel: clocksource: Switched to clocksource kvm-clock Sep 9 00:28:13.009558 kernel: VFS: Disk quotas dquot_6.6.0 Sep 9 00:28:13.009569 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 9 00:28:13.009580 kernel: pnp: PnP ACPI init Sep 9 00:28:13.009965 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Sep 9 00:28:13.009982 kernel: pnp: PnP ACPI: found 6 devices Sep 9 00:28:13.009994 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 9 00:28:13.010010 kernel: NET: Registered PF_INET protocol family Sep 9 00:28:13.010021 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 9 00:28:13.010032 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 9 00:28:13.010043 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 9 00:28:13.010055 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 9 00:28:13.010066 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 9 00:28:13.010095 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 9 00:28:13.010107 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 9 00:28:13.010118 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 9 00:28:13.010133 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 9 00:28:13.010152 kernel: NET: Registered PF_XDP protocol family Sep 9 00:28:13.010302 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 9 00:28:13.010442 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 9 00:28:13.010584 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 9 00:28:13.010727 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Sep 9 00:28:13.010870 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Sep 9 00:28:13.011010 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Sep 9 00:28:13.011028 kernel: PCI: CLS 0 bytes, default 64 Sep 9 00:28:13.011039 kernel: Initialise system trusted keyrings Sep 9 00:28:13.011051 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 9 00:28:13.011062 kernel: Key type asymmetric registered Sep 9 00:28:13.011151 kernel: Asymmetric key parser 'x509' registered Sep 9 00:28:13.011164 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 9 00:28:13.011175 kernel: io scheduler mq-deadline registered Sep 9 00:28:13.011187 kernel: io scheduler kyber registered Sep 9 00:28:13.011198 kernel: io scheduler bfq registered Sep 9 00:28:13.011213 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 9 00:28:13.011225 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 9 00:28:13.011236 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 9 00:28:13.011247 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 9 00:28:13.011259 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 9 00:28:13.011270 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 9 00:28:13.011282 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 9 00:28:13.011293 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 9 00:28:13.011304 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 9 00:28:13.011485 kernel: rtc_cmos 00:04: RTC can wake from S4 Sep 9 00:28:13.011502 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 9 00:28:13.011645 kernel: rtc_cmos 00:04: registered as rtc0 Sep 9 00:28:13.011790 kernel: rtc_cmos 00:04: setting system clock to 2025-09-09T00:28:12 UTC (1757377692) Sep 9 00:28:13.011938 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Sep 9 00:28:13.011955 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Sep 9 00:28:13.011966 kernel: NET: Registered PF_INET6 protocol family Sep 9 00:28:13.011977 kernel: Segment Routing with IPv6 Sep 9 00:28:13.011993 kernel: In-situ OAM (IOAM) with IPv6 Sep 9 00:28:13.012004 kernel: NET: Registered PF_PACKET protocol family Sep 9 00:28:13.012016 kernel: Key type dns_resolver registered Sep 9 00:28:13.012027 kernel: IPI shorthand broadcast: enabled Sep 9 00:28:13.012038 kernel: sched_clock: Marking stable (940004399, 133911709)->(1150671048, -76754940) Sep 9 00:28:13.012049 kernel: registered taskstats version 1 Sep 9 00:28:13.012060 kernel: Loading compiled-in X.509 certificates Sep 9 00:28:13.012084 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.104-flatcar: cc5240ef94b546331b2896cdc739274c03278c51' Sep 9 00:28:13.012096 kernel: Key type .fscrypt registered Sep 9 00:28:13.012111 kernel: Key type fscrypt-provisioning registered Sep 9 00:28:13.012122 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 9 00:28:13.012133 kernel: ima: Allocated hash algorithm: sha1 Sep 9 00:28:13.012153 kernel: ima: No architecture policies found Sep 9 00:28:13.012164 kernel: clk: Disabling unused clocks Sep 9 00:28:13.012175 kernel: Freeing unused kernel image (initmem) memory: 42880K Sep 9 00:28:13.012186 kernel: Write protecting the kernel read-only data: 36864k Sep 9 00:28:13.012197 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Sep 9 00:28:13.012209 kernel: Run /init as init process Sep 9 00:28:13.012223 kernel: with arguments: Sep 9 00:28:13.012233 kernel: /init Sep 9 00:28:13.012244 kernel: with environment: Sep 9 00:28:13.012255 kernel: HOME=/ Sep 9 00:28:13.012265 kernel: TERM=linux Sep 9 00:28:13.012275 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 9 00:28:13.012292 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 9 00:28:13.012305 systemd[1]: Detected virtualization kvm. Sep 9 00:28:13.012321 systemd[1]: Detected architecture x86-64. Sep 9 00:28:13.012333 systemd[1]: Running in initrd. Sep 9 00:28:13.012344 systemd[1]: No hostname configured, using default hostname. Sep 9 00:28:13.012356 systemd[1]: Hostname set to . Sep 9 00:28:13.012369 systemd[1]: Initializing machine ID from VM UUID. Sep 9 00:28:13.012381 systemd[1]: Queued start job for default target initrd.target. Sep 9 00:28:13.012393 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 9 00:28:13.012405 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 9 00:28:13.012421 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 9 00:28:13.012449 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 9 00:28:13.012465 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 9 00:28:13.012478 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 9 00:28:13.012493 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 9 00:28:13.012509 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 9 00:28:13.012522 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 9 00:28:13.012534 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 9 00:28:13.012546 systemd[1]: Reached target paths.target - Path Units. Sep 9 00:28:13.012562 systemd[1]: Reached target slices.target - Slice Units. Sep 9 00:28:13.012574 systemd[1]: Reached target swap.target - Swaps. Sep 9 00:28:13.012586 systemd[1]: Reached target timers.target - Timer Units. Sep 9 00:28:13.012598 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 9 00:28:13.012613 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 9 00:28:13.012626 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 9 00:28:13.012638 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 9 00:28:13.012651 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 9 00:28:13.012663 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 9 00:28:13.012676 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 9 00:28:13.012688 systemd[1]: Reached target sockets.target - Socket Units. Sep 9 00:28:13.012700 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 9 00:28:13.012716 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 9 00:28:13.012729 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 9 00:28:13.012741 systemd[1]: Starting systemd-fsck-usr.service... Sep 9 00:28:13.012753 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 9 00:28:13.012766 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 9 00:28:13.012778 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 00:28:13.012791 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 9 00:28:13.012804 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 9 00:28:13.012815 systemd[1]: Finished systemd-fsck-usr.service. Sep 9 00:28:13.012862 systemd-journald[193]: Collecting audit messages is disabled. Sep 9 00:28:13.012894 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 9 00:28:13.012910 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 9 00:28:13.012922 systemd-journald[193]: Journal started Sep 9 00:28:13.012952 systemd-journald[193]: Runtime Journal (/run/log/journal/2aa2edc7bd0a4e2ca1b119a71e145e98) is 6.0M, max 48.4M, 42.3M free. Sep 9 00:28:12.993961 systemd-modules-load[194]: Inserted module 'overlay' Sep 9 00:28:13.038348 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 9 00:28:13.038380 kernel: Bridge firewalling registered Sep 9 00:28:13.028269 systemd-modules-load[194]: Inserted module 'br_netfilter' Sep 9 00:28:13.040098 systemd[1]: Started systemd-journald.service - Journal Service. Sep 9 00:28:13.041162 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 9 00:28:13.060401 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 9 00:28:13.061443 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 9 00:28:13.067236 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 9 00:28:13.072984 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 00:28:13.078349 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 9 00:28:13.088667 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 9 00:28:13.089104 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 9 00:28:13.096357 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 9 00:28:13.105434 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 9 00:28:13.105929 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 9 00:28:13.109893 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 9 00:28:13.127891 dracut-cmdline[232]: dracut-dracut-053 Sep 9 00:28:13.131607 dracut-cmdline[232]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=99a67175ee6aabbc03a22dabcade16d60ad192b31c4118a259bf1f24bbfa2d29 Sep 9 00:28:13.145303 systemd-resolved[227]: Positive Trust Anchors: Sep 9 00:28:13.145324 systemd-resolved[227]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 9 00:28:13.145368 systemd-resolved[227]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 9 00:28:13.148390 systemd-resolved[227]: Defaulting to hostname 'linux'. Sep 9 00:28:13.149611 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 9 00:28:13.155680 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 9 00:28:13.256129 kernel: SCSI subsystem initialized Sep 9 00:28:13.267113 kernel: Loading iSCSI transport class v2.0-870. Sep 9 00:28:13.280160 kernel: iscsi: registered transport (tcp) Sep 9 00:28:13.303152 kernel: iscsi: registered transport (qla4xxx) Sep 9 00:28:13.303252 kernel: QLogic iSCSI HBA Driver Sep 9 00:28:13.356970 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 9 00:28:13.363324 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 9 00:28:13.425143 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 9 00:28:13.425262 kernel: device-mapper: uevent: version 1.0.3 Sep 9 00:28:13.425282 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 9 00:28:13.482666 kernel: raid6: avx2x4 gen() 20450 MB/s Sep 9 00:28:13.499156 kernel: raid6: avx2x2 gen() 23998 MB/s Sep 9 00:28:13.516486 kernel: raid6: avx2x1 gen() 16372 MB/s Sep 9 00:28:13.516672 kernel: raid6: using algorithm avx2x2 gen() 23998 MB/s Sep 9 00:28:13.534500 kernel: raid6: .... xor() 13069 MB/s, rmw enabled Sep 9 00:28:13.534619 kernel: raid6: using avx2x2 recovery algorithm Sep 9 00:28:13.563153 kernel: xor: automatically using best checksumming function avx Sep 9 00:28:13.732122 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 9 00:28:13.749329 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 9 00:28:13.761387 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 9 00:28:13.777416 systemd-udevd[415]: Using default interface naming scheme 'v255'. Sep 9 00:28:13.783668 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 9 00:28:13.795306 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 9 00:28:13.817816 dracut-pre-trigger[423]: rd.md=0: removing MD RAID activation Sep 9 00:28:13.855637 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 9 00:28:13.875297 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 9 00:28:13.955943 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 9 00:28:13.968434 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 9 00:28:13.982376 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 9 00:28:13.988663 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 9 00:28:13.990274 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 9 00:28:13.992665 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 9 00:28:14.005290 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 9 00:28:14.013883 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Sep 9 00:28:14.019495 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 9 00:28:14.019707 kernel: cryptd: max_cpu_qlen set to 1000 Sep 9 00:28:14.025241 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 9 00:28:14.025278 kernel: GPT:9289727 != 19775487 Sep 9 00:28:14.025290 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 9 00:28:14.025300 kernel: GPT:9289727 != 19775487 Sep 9 00:28:14.025311 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 9 00:28:14.021767 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 9 00:28:14.028792 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 00:28:14.054141 kernel: AVX2 version of gcm_enc/dec engaged. Sep 9 00:28:14.058108 kernel: AES CTR mode by8 optimization enabled Sep 9 00:28:14.058160 kernel: libata version 3.00 loaded. Sep 9 00:28:14.065293 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 9 00:28:14.069126 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (473) Sep 9 00:28:14.075253 kernel: BTRFS: device fsid 7cd16ef1-c91b-4e35-a9b3-a431b3c1949a devid 1 transid 36 /dev/vda3 scanned by (udev-worker) (460) Sep 9 00:28:14.079347 kernel: ahci 0000:00:1f.2: version 3.0 Sep 9 00:28:14.079611 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 9 00:28:14.081104 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Sep 9 00:28:14.081349 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 9 00:28:14.085490 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 9 00:28:14.089789 kernel: scsi host0: ahci Sep 9 00:28:14.090001 kernel: scsi host1: ahci Sep 9 00:28:14.091359 kernel: scsi host2: ahci Sep 9 00:28:14.091562 kernel: scsi host3: ahci Sep 9 00:28:14.092304 kernel: scsi host4: ahci Sep 9 00:28:14.093984 kernel: scsi host5: ahci Sep 9 00:28:14.094363 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Sep 9 00:28:14.094378 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Sep 9 00:28:14.095945 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Sep 9 00:28:14.095979 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Sep 9 00:28:14.097933 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Sep 9 00:28:14.097963 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Sep 9 00:28:14.099585 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 9 00:28:14.103203 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 9 00:28:14.124585 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 9 00:28:14.149337 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 9 00:28:14.226559 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 9 00:28:14.226677 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 9 00:28:14.230395 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 9 00:28:14.232777 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 9 00:28:14.233803 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 00:28:14.236022 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 00:28:14.239177 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 00:28:14.294591 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 00:28:14.397275 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 9 00:28:14.417394 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 9 00:28:14.486764 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 9 00:28:14.486826 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 9 00:28:14.487090 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 9 00:28:14.488145 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 9 00:28:14.489119 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Sep 9 00:28:14.490102 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 9 00:28:14.491167 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Sep 9 00:28:14.491253 kernel: ata3.00: applying bridge limits Sep 9 00:28:14.492287 kernel: ata3.00: configured for UDMA/100 Sep 9 00:28:14.493122 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Sep 9 00:28:14.557136 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Sep 9 00:28:14.557580 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 9 00:28:14.579189 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Sep 9 00:28:15.140058 disk-uuid[561]: Primary Header is updated. Sep 9 00:28:15.140058 disk-uuid[561]: Secondary Entries is updated. Sep 9 00:28:15.140058 disk-uuid[561]: Secondary Header is updated. Sep 9 00:28:15.169861 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 00:28:15.175109 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 00:28:16.284106 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 9 00:28:16.284197 disk-uuid[580]: The operation has completed successfully. Sep 9 00:28:16.310039 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 9 00:28:16.310245 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 9 00:28:16.381448 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 9 00:28:16.385275 sh[591]: Success Sep 9 00:28:16.452110 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Sep 9 00:28:16.495102 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 9 00:28:16.537378 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 9 00:28:16.540127 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 9 00:28:16.632344 kernel: BTRFS info (device dm-0): first mount of filesystem 7cd16ef1-c91b-4e35-a9b3-a431b3c1949a Sep 9 00:28:16.632385 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 9 00:28:16.632396 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 9 00:28:16.633391 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 9 00:28:16.634248 kernel: BTRFS info (device dm-0): using free space tree Sep 9 00:28:16.639195 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 9 00:28:16.640841 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 9 00:28:16.649273 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 9 00:28:16.651360 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 9 00:28:16.736838 kernel: BTRFS info (device vda6): first mount of filesystem a5263def-4663-4ce6-b873-45a7d7f1ec33 Sep 9 00:28:16.736912 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 9 00:28:16.736923 kernel: BTRFS info (device vda6): using free space tree Sep 9 00:28:16.740132 kernel: BTRFS info (device vda6): auto enabling async discard Sep 9 00:28:16.752960 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 9 00:28:16.755034 kernel: BTRFS info (device vda6): last unmount of filesystem a5263def-4663-4ce6-b873-45a7d7f1ec33 Sep 9 00:28:16.774519 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 9 00:28:16.900288 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 9 00:28:16.928331 systemd-networkd[769]: lo: Link UP Sep 9 00:28:16.928345 systemd-networkd[769]: lo: Gained carrier Sep 9 00:28:16.931460 systemd-networkd[769]: Enumeration completed Sep 9 00:28:16.931570 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 9 00:28:16.932389 systemd[1]: Reached target network.target - Network. Sep 9 00:28:17.066752 systemd-networkd[769]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 00:28:17.066764 systemd-networkd[769]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 9 00:28:17.070941 systemd-networkd[769]: eth0: Link UP Sep 9 00:28:17.070952 systemd-networkd[769]: eth0: Gained carrier Sep 9 00:28:17.070960 systemd-networkd[769]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 00:28:17.102140 systemd-networkd[769]: eth0: DHCPv4 address 10.0.0.113/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 9 00:28:17.545468 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 9 00:28:17.575478 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 9 00:28:17.764565 ignition[775]: Ignition 2.19.0 Sep 9 00:28:17.764581 ignition[775]: Stage: fetch-offline Sep 9 00:28:17.764643 ignition[775]: no configs at "/usr/lib/ignition/base.d" Sep 9 00:28:17.764661 ignition[775]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 00:28:17.764800 ignition[775]: parsed url from cmdline: "" Sep 9 00:28:17.764806 ignition[775]: no config URL provided Sep 9 00:28:17.764813 ignition[775]: reading system config file "/usr/lib/ignition/user.ign" Sep 9 00:28:17.764826 ignition[775]: no config at "/usr/lib/ignition/user.ign" Sep 9 00:28:17.764873 ignition[775]: op(1): [started] loading QEMU firmware config module Sep 9 00:28:17.764881 ignition[775]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 9 00:28:17.783257 ignition[775]: op(1): [finished] loading QEMU firmware config module Sep 9 00:28:17.783290 ignition[775]: QEMU firmware config was not found. Ignoring... Sep 9 00:28:17.825358 ignition[775]: parsing config with SHA512: 9542c8bcf6b99eba95dee8fdefdb8f9638d93c5d518634bf1159e53f09d7acccfdcdeecb86fc129814ca96619c138751cabe2a076cdd3483f6451fd461fd8304 Sep 9 00:28:17.829807 unknown[775]: fetched base config from "system" Sep 9 00:28:17.829824 unknown[775]: fetched user config from "qemu" Sep 9 00:28:17.830275 ignition[775]: fetch-offline: fetch-offline passed Sep 9 00:28:17.830348 ignition[775]: Ignition finished successfully Sep 9 00:28:17.853842 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 9 00:28:17.854194 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 9 00:28:17.870405 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 9 00:28:17.886629 ignition[785]: Ignition 2.19.0 Sep 9 00:28:17.886640 ignition[785]: Stage: kargs Sep 9 00:28:17.886827 ignition[785]: no configs at "/usr/lib/ignition/base.d" Sep 9 00:28:17.886838 ignition[785]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 00:28:17.887735 ignition[785]: kargs: kargs passed Sep 9 00:28:17.887803 ignition[785]: Ignition finished successfully Sep 9 00:28:17.895379 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 9 00:28:17.901426 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 9 00:28:17.921496 ignition[793]: Ignition 2.19.0 Sep 9 00:28:17.921508 ignition[793]: Stage: disks Sep 9 00:28:17.921729 ignition[793]: no configs at "/usr/lib/ignition/base.d" Sep 9 00:28:17.921742 ignition[793]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 00:28:17.922880 ignition[793]: disks: disks passed Sep 9 00:28:17.922930 ignition[793]: Ignition finished successfully Sep 9 00:28:17.928465 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 9 00:28:17.928834 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 9 00:28:17.931631 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 9 00:28:17.933194 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 9 00:28:17.935556 systemd[1]: Reached target sysinit.target - System Initialization. Sep 9 00:28:17.937464 systemd[1]: Reached target basic.target - Basic System. Sep 9 00:28:17.948350 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 9 00:28:18.005305 systemd-fsck[803]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 9 00:28:18.030385 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 9 00:28:18.042438 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 9 00:28:18.162128 kernel: EXT4-fs (vda9): mounted filesystem ee55a213-d578-493d-a79b-e10c399cd35c r/w with ordered data mode. Quota mode: none. Sep 9 00:28:18.164140 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 9 00:28:18.166416 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 9 00:28:18.185331 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 9 00:28:18.189171 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 9 00:28:18.189696 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 9 00:28:18.189774 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 9 00:28:18.189820 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 9 00:28:18.204461 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 9 00:28:18.207401 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (812) Sep 9 00:28:18.207499 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 9 00:28:18.213415 kernel: BTRFS info (device vda6): first mount of filesystem a5263def-4663-4ce6-b873-45a7d7f1ec33 Sep 9 00:28:18.213456 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 9 00:28:18.213472 kernel: BTRFS info (device vda6): using free space tree Sep 9 00:28:18.218308 kernel: BTRFS info (device vda6): auto enabling async discard Sep 9 00:28:18.219924 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 9 00:28:18.267193 initrd-setup-root[836]: cut: /sysroot/etc/passwd: No such file or directory Sep 9 00:28:18.273627 initrd-setup-root[843]: cut: /sysroot/etc/group: No such file or directory Sep 9 00:28:18.278982 initrd-setup-root[850]: cut: /sysroot/etc/shadow: No such file or directory Sep 9 00:28:18.283347 initrd-setup-root[857]: cut: /sysroot/etc/gshadow: No such file or directory Sep 9 00:28:18.408523 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 9 00:28:18.451451 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 9 00:28:18.455826 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 9 00:28:18.462285 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 9 00:28:18.493902 kernel: BTRFS info (device vda6): last unmount of filesystem a5263def-4663-4ce6-b873-45a7d7f1ec33 Sep 9 00:28:18.512152 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 9 00:28:18.531093 systemd-networkd[769]: eth0: Gained IPv6LL Sep 9 00:28:18.578099 ignition[928]: INFO : Ignition 2.19.0 Sep 9 00:28:18.578099 ignition[928]: INFO : Stage: mount Sep 9 00:28:18.614105 ignition[928]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 9 00:28:18.614105 ignition[928]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 00:28:18.617980 ignition[928]: INFO : mount: mount passed Sep 9 00:28:18.618953 ignition[928]: INFO : Ignition finished successfully Sep 9 00:28:18.622742 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 9 00:28:18.632193 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 9 00:28:19.177228 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 9 00:28:19.198094 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (938) Sep 9 00:28:19.198124 kernel: BTRFS info (device vda6): first mount of filesystem a5263def-4663-4ce6-b873-45a7d7f1ec33 Sep 9 00:28:19.200191 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 9 00:28:19.200205 kernel: BTRFS info (device vda6): using free space tree Sep 9 00:28:19.203097 kernel: BTRFS info (device vda6): auto enabling async discard Sep 9 00:28:19.204950 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 9 00:28:19.230886 ignition[955]: INFO : Ignition 2.19.0 Sep 9 00:28:19.230886 ignition[955]: INFO : Stage: files Sep 9 00:28:19.269311 ignition[955]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 9 00:28:19.269311 ignition[955]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 00:28:19.272767 ignition[955]: DEBUG : files: compiled without relabeling support, skipping Sep 9 00:28:19.274798 ignition[955]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 9 00:28:19.274798 ignition[955]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 9 00:28:19.282742 ignition[955]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 9 00:28:19.284563 ignition[955]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 9 00:28:19.284563 ignition[955]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 9 00:28:19.283541 unknown[955]: wrote ssh authorized keys file for user: core Sep 9 00:28:19.289348 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 9 00:28:19.289348 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Sep 9 00:28:19.340812 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 9 00:28:19.693442 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 9 00:28:19.695893 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 9 00:28:19.695893 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 9 00:28:19.950780 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 9 00:28:20.219504 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 9 00:28:20.219504 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 9 00:28:20.247892 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Sep 9 00:28:20.624130 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 9 00:28:21.356982 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 9 00:28:21.356982 ignition[955]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Sep 9 00:28:21.361889 ignition[955]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Sep 9 00:28:21.404688 ignition[955]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 9 00:28:21.412335 ignition[955]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 9 00:28:21.413995 ignition[955]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Sep 9 00:28:21.413995 ignition[955]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Sep 9 00:28:21.413995 ignition[955]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Sep 9 00:28:21.413995 ignition[955]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 9 00:28:21.413995 ignition[955]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 9 00:28:21.413995 ignition[955]: INFO : files: files passed Sep 9 00:28:21.413995 ignition[955]: INFO : Ignition finished successfully Sep 9 00:28:21.416177 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 9 00:28:21.426246 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 9 00:28:21.428463 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 9 00:28:21.432460 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 9 00:28:21.432654 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 9 00:28:21.446591 initrd-setup-root-after-ignition[983]: grep: /sysroot/oem/oem-release: No such file or directory Sep 9 00:28:21.450182 initrd-setup-root-after-ignition[985]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 9 00:28:21.450182 initrd-setup-root-after-ignition[985]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 9 00:28:21.453576 initrd-setup-root-after-ignition[989]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 9 00:28:21.455328 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 9 00:28:21.455625 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 9 00:28:21.462276 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 9 00:28:21.491517 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 9 00:28:21.527247 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 9 00:28:21.530619 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 9 00:28:21.532695 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 9 00:28:21.534988 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 9 00:28:21.549474 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 9 00:28:21.564314 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 9 00:28:21.577306 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 9 00:28:21.588242 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 9 00:28:21.590737 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 9 00:28:21.593201 systemd[1]: Stopped target timers.target - Timer Units. Sep 9 00:28:21.595100 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 9 00:28:21.595317 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 9 00:28:21.598646 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 9 00:28:21.613970 systemd[1]: Stopped target basic.target - Basic System. Sep 9 00:28:21.614628 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 9 00:28:21.617980 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 9 00:28:21.619376 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 9 00:28:21.622938 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 9 00:28:21.624178 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 9 00:28:21.626176 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 9 00:28:21.626650 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 9 00:28:21.626996 systemd[1]: Stopped target swap.target - Swaps. Sep 9 00:28:21.632239 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 9 00:28:21.632436 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 9 00:28:21.634147 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 9 00:28:21.634692 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 9 00:28:21.634969 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 9 00:28:21.635268 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 9 00:28:21.640538 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 9 00:28:21.640691 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 9 00:28:21.645005 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 9 00:28:21.645192 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 9 00:28:21.646466 systemd[1]: Stopped target paths.target - Path Units. Sep 9 00:28:21.648534 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 9 00:28:21.654171 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 9 00:28:21.657023 systemd[1]: Stopped target slices.target - Slice Units. Sep 9 00:28:21.658886 systemd[1]: Stopped target sockets.target - Socket Units. Sep 9 00:28:21.660836 systemd[1]: iscsid.socket: Deactivated successfully. Sep 9 00:28:21.661762 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 9 00:28:21.664899 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 9 00:28:21.665810 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 9 00:28:21.667971 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 9 00:28:21.669180 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 9 00:28:21.671650 systemd[1]: ignition-files.service: Deactivated successfully. Sep 9 00:28:21.672634 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 9 00:28:21.682461 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 9 00:28:21.702048 ignition[1009]: INFO : Ignition 2.19.0 Sep 9 00:28:21.702048 ignition[1009]: INFO : Stage: umount Sep 9 00:28:21.704429 ignition[1009]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 9 00:28:21.704429 ignition[1009]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 9 00:28:21.704429 ignition[1009]: INFO : umount: umount passed Sep 9 00:28:21.704429 ignition[1009]: INFO : Ignition finished successfully Sep 9 00:28:21.702096 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 9 00:28:21.703665 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 9 00:28:21.705434 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 9 00:28:21.710384 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 9 00:28:21.711478 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 9 00:28:21.717474 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 9 00:28:21.738754 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 9 00:28:21.743133 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 9 00:28:21.744345 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 9 00:28:21.747779 systemd[1]: Stopped target network.target - Network. Sep 9 00:28:21.749842 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 9 00:28:21.750952 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 9 00:28:21.753311 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 9 00:28:21.754769 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 9 00:28:21.757402 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 9 00:28:21.758698 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 9 00:28:21.772864 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 9 00:28:21.773913 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 9 00:28:21.776203 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 9 00:28:21.778593 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 9 00:28:21.781916 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 9 00:28:21.784253 systemd-networkd[769]: eth0: DHCPv6 lease lost Sep 9 00:28:21.793169 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 9 00:28:21.793382 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 9 00:28:21.796813 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 9 00:28:21.796978 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 9 00:28:21.802039 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 9 00:28:21.802134 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 9 00:28:21.817250 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 9 00:28:21.818275 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 9 00:28:21.818342 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 9 00:28:21.820652 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 9 00:28:21.820707 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 9 00:28:21.823006 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 9 00:28:21.823063 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 9 00:28:21.825748 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 9 00:28:21.825807 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 9 00:28:21.826937 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 9 00:28:21.839090 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 9 00:28:21.839231 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 9 00:28:21.870111 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 9 00:28:21.870318 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 9 00:28:21.873722 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 9 00:28:21.873795 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 9 00:28:21.875777 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 9 00:28:21.875825 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 9 00:28:21.876836 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 9 00:28:21.876889 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 9 00:28:21.881440 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 9 00:28:21.881499 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 9 00:28:21.882802 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 9 00:28:21.882853 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 9 00:28:21.890440 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 9 00:28:21.892738 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 9 00:28:21.892824 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 9 00:28:21.896396 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 9 00:28:21.897541 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 9 00:28:21.900217 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 9 00:28:21.900277 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 9 00:28:21.903553 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 9 00:28:21.903621 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 00:28:21.911505 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 9 00:28:21.912699 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 9 00:28:22.197889 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 9 00:28:22.199122 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 9 00:28:22.201434 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 9 00:28:22.203679 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 9 00:28:22.204796 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 9 00:28:22.218271 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 9 00:28:22.232148 systemd[1]: Switching root. Sep 9 00:28:22.272564 systemd-journald[193]: Journal stopped Sep 9 00:28:24.319895 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). Sep 9 00:28:24.319976 kernel: SELinux: policy capability network_peer_controls=1 Sep 9 00:28:24.319996 kernel: SELinux: policy capability open_perms=1 Sep 9 00:28:24.320008 kernel: SELinux: policy capability extended_socket_class=1 Sep 9 00:28:24.320019 kernel: SELinux: policy capability always_check_network=0 Sep 9 00:28:24.320031 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 9 00:28:24.320055 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 9 00:28:24.320091 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 9 00:28:24.320103 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 9 00:28:24.320115 kernel: audit: type=1403 audit(1757377703.243:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 9 00:28:24.320132 systemd[1]: Successfully loaded SELinux policy in 41.430ms. Sep 9 00:28:24.320148 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 15.074ms. Sep 9 00:28:24.320162 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 9 00:28:24.320174 systemd[1]: Detected virtualization kvm. Sep 9 00:28:24.320187 systemd[1]: Detected architecture x86-64. Sep 9 00:28:24.320199 systemd[1]: Detected first boot. Sep 9 00:28:24.320218 systemd[1]: Initializing machine ID from VM UUID. Sep 9 00:28:24.320231 zram_generator::config[1053]: No configuration found. Sep 9 00:28:24.320244 systemd[1]: Populated /etc with preset unit settings. Sep 9 00:28:24.320257 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 9 00:28:24.320269 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 9 00:28:24.320281 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 9 00:28:24.320295 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 9 00:28:24.320307 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 9 00:28:24.320326 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 9 00:28:24.320338 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 9 00:28:24.320351 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 9 00:28:24.320366 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 9 00:28:24.320379 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 9 00:28:24.320397 systemd[1]: Created slice user.slice - User and Session Slice. Sep 9 00:28:24.320410 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 9 00:28:24.320422 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 9 00:28:24.320435 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 9 00:28:24.320453 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 9 00:28:24.320466 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 9 00:28:24.320479 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 9 00:28:24.320492 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 9 00:28:24.320504 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 9 00:28:24.320517 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 9 00:28:24.320529 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 9 00:28:24.320542 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 9 00:28:24.320560 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 9 00:28:24.320573 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 9 00:28:24.320586 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 9 00:28:24.320598 systemd[1]: Reached target slices.target - Slice Units. Sep 9 00:28:24.320610 systemd[1]: Reached target swap.target - Swaps. Sep 9 00:28:24.320622 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 9 00:28:24.320635 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 9 00:28:24.320648 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 9 00:28:24.320661 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 9 00:28:24.320679 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 9 00:28:24.320691 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 9 00:28:24.320703 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 9 00:28:24.320720 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 9 00:28:24.320733 systemd[1]: Mounting media.mount - External Media Directory... Sep 9 00:28:24.320745 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 9 00:28:24.320757 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 9 00:28:24.320770 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 9 00:28:24.320787 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 9 00:28:24.320800 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 9 00:28:24.320813 systemd[1]: Reached target machines.target - Containers. Sep 9 00:28:24.320825 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 9 00:28:24.320837 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 9 00:28:24.320850 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 9 00:28:24.320872 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 9 00:28:24.320887 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 9 00:28:24.320903 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 9 00:28:24.320925 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 9 00:28:24.320937 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 9 00:28:24.320952 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 9 00:28:24.320964 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 9 00:28:24.320976 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 9 00:28:24.320989 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 9 00:28:24.321001 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 9 00:28:24.321013 systemd[1]: Stopped systemd-fsck-usr.service. Sep 9 00:28:24.321032 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 9 00:28:24.321044 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 9 00:28:24.321056 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 9 00:28:24.321416 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 9 00:28:24.321755 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 9 00:28:24.321773 systemd[1]: verity-setup.service: Deactivated successfully. Sep 9 00:28:24.321786 systemd[1]: Stopped verity-setup.service. Sep 9 00:28:24.321799 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 9 00:28:24.321812 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 9 00:28:24.321835 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 9 00:28:24.321847 systemd[1]: Mounted media.mount - External Media Directory. Sep 9 00:28:24.321869 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 9 00:28:24.322283 systemd-journald[1116]: Collecting audit messages is disabled. Sep 9 00:28:24.322318 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 9 00:28:24.322331 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 9 00:28:24.322344 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 9 00:28:24.322357 systemd-journald[1116]: Journal started Sep 9 00:28:24.322379 systemd-journald[1116]: Runtime Journal (/run/log/journal/2aa2edc7bd0a4e2ca1b119a71e145e98) is 6.0M, max 48.4M, 42.3M free. Sep 9 00:28:23.893361 systemd[1]: Queued start job for default target multi-user.target. Sep 9 00:28:23.917423 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 9 00:28:23.918013 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 9 00:28:24.339475 systemd[1]: Started systemd-journald.service - Journal Service. Sep 9 00:28:24.339583 kernel: loop: module loaded Sep 9 00:28:24.329472 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 9 00:28:24.329711 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 9 00:28:24.331622 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 9 00:28:24.331813 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 9 00:28:24.333325 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 9 00:28:24.333503 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 9 00:28:24.335113 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 9 00:28:24.335297 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 9 00:28:24.336732 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 9 00:28:24.338253 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 9 00:28:24.340477 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 9 00:28:24.345091 kernel: fuse: init (API version 7.39) Sep 9 00:28:24.346126 kernel: ACPI: bus type drm_connector registered Sep 9 00:28:24.347573 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 9 00:28:24.347814 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 9 00:28:24.350531 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 9 00:28:24.350735 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 9 00:28:24.364496 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 9 00:28:24.401249 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 9 00:28:24.404552 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 9 00:28:24.406072 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 9 00:28:24.406137 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 9 00:28:24.408795 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 9 00:28:24.412158 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 9 00:28:24.428980 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 9 00:28:24.430516 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 9 00:28:24.432832 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 9 00:28:24.435686 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 9 00:28:24.437296 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 9 00:28:24.438745 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 9 00:28:24.440284 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 9 00:28:24.447260 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 9 00:28:24.457527 systemd-journald[1116]: Time spent on flushing to /var/log/journal/2aa2edc7bd0a4e2ca1b119a71e145e98 is 13.601ms for 954 entries. Sep 9 00:28:24.457527 systemd-journald[1116]: System Journal (/var/log/journal/2aa2edc7bd0a4e2ca1b119a71e145e98) is 8.0M, max 195.6M, 187.6M free. Sep 9 00:28:24.952326 systemd-journald[1116]: Received client request to flush runtime journal. Sep 9 00:28:24.952403 kernel: loop0: detected capacity change from 0 to 140768 Sep 9 00:28:24.952442 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 9 00:28:24.952466 kernel: loop1: detected capacity change from 0 to 224512 Sep 9 00:28:24.952495 kernel: loop2: detected capacity change from 0 to 142488 Sep 9 00:28:24.952520 kernel: loop3: detected capacity change from 0 to 140768 Sep 9 00:28:24.450714 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 9 00:28:24.455575 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 9 00:28:24.461297 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 9 00:28:24.474207 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 9 00:28:24.476604 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 9 00:28:24.478632 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 9 00:28:24.497247 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 9 00:28:24.523680 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 9 00:28:24.526659 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 9 00:28:24.528479 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 9 00:28:24.538278 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 9 00:28:24.541661 udevadm[1167]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 9 00:28:24.574818 systemd-tmpfiles[1157]: ACLs are not supported, ignoring. Sep 9 00:28:24.574836 systemd-tmpfiles[1157]: ACLs are not supported, ignoring. Sep 9 00:28:24.598713 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 9 00:28:24.897825 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 9 00:28:24.910339 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 9 00:28:24.948330 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 9 00:28:24.958357 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 9 00:28:24.962810 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 9 00:28:24.990125 kernel: loop4: detected capacity change from 0 to 224512 Sep 9 00:28:24.997468 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Sep 9 00:28:24.997496 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Sep 9 00:28:25.005944 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 9 00:28:25.010482 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 9 00:28:25.011465 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 9 00:28:25.018117 kernel: loop5: detected capacity change from 0 to 142488 Sep 9 00:28:25.033757 (sd-merge)[1187]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 9 00:28:25.034590 (sd-merge)[1187]: Merged extensions into '/usr'. Sep 9 00:28:25.041562 systemd[1]: Reloading requested from client PID 1154 ('systemd-sysext') (unit systemd-sysext.service)... Sep 9 00:28:25.041593 systemd[1]: Reloading... Sep 9 00:28:25.110129 zram_generator::config[1220]: No configuration found. Sep 9 00:28:25.227619 ldconfig[1147]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 9 00:28:25.290852 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 9 00:28:25.349221 systemd[1]: Reloading finished in 307 ms. Sep 9 00:28:25.384908 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 9 00:28:25.386552 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 9 00:28:25.402507 systemd[1]: Starting ensure-sysext.service... Sep 9 00:28:25.405943 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 9 00:28:25.412880 systemd[1]: Reloading requested from client PID 1258 ('systemctl') (unit ensure-sysext.service)... Sep 9 00:28:25.412901 systemd[1]: Reloading... Sep 9 00:28:25.498520 systemd-tmpfiles[1259]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 9 00:28:25.499111 systemd-tmpfiles[1259]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 9 00:28:25.505681 systemd-tmpfiles[1259]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 9 00:28:25.506212 systemd-tmpfiles[1259]: ACLs are not supported, ignoring. Sep 9 00:28:25.506350 systemd-tmpfiles[1259]: ACLs are not supported, ignoring. Sep 9 00:28:25.509163 zram_generator::config[1288]: No configuration found. Sep 9 00:28:25.540298 systemd-tmpfiles[1259]: Detected autofs mount point /boot during canonicalization of boot. Sep 9 00:28:25.540327 systemd-tmpfiles[1259]: Skipping /boot Sep 9 00:28:25.554914 systemd-tmpfiles[1259]: Detected autofs mount point /boot during canonicalization of boot. Sep 9 00:28:25.554935 systemd-tmpfiles[1259]: Skipping /boot Sep 9 00:28:25.646506 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 9 00:28:25.701202 systemd[1]: Reloading finished in 287 ms. Sep 9 00:28:25.721497 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 9 00:28:25.747682 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 9 00:28:25.756311 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 9 00:28:25.759511 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 9 00:28:25.784691 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 9 00:28:25.789586 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 9 00:28:25.802538 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 9 00:28:25.807120 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 9 00:28:25.814255 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 9 00:28:25.814446 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 9 00:28:25.824523 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 9 00:28:25.838387 systemd-udevd[1335]: Using default interface naming scheme 'v255'. Sep 9 00:28:25.841376 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 9 00:28:25.845546 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 9 00:28:25.863600 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 9 00:28:25.864356 augenrules[1348]: No rules Sep 9 00:28:25.866535 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 9 00:28:25.867833 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 9 00:28:25.869439 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 9 00:28:25.871629 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 9 00:28:25.873840 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 9 00:28:25.874106 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 9 00:28:25.879184 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 9 00:28:25.879416 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 9 00:28:25.883240 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 9 00:28:25.883513 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 9 00:28:25.887407 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 9 00:28:25.898357 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 9 00:28:25.909664 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 9 00:28:25.909961 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 9 00:28:25.919903 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 9 00:28:25.924343 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 9 00:28:25.932871 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 9 00:28:25.936337 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 9 00:28:25.938351 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 9 00:28:25.956482 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 9 00:28:25.962502 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 9 00:28:25.963855 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 9 00:28:25.965496 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 9 00:28:25.967508 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 9 00:28:25.969475 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 9 00:28:25.969712 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 9 00:28:25.971728 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 9 00:28:25.971992 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 9 00:28:25.974225 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 9 00:28:25.974480 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 9 00:28:25.976535 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 9 00:28:25.976784 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 9 00:28:25.984186 systemd[1]: Finished ensure-sysext.service. Sep 9 00:28:25.997560 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 9 00:28:25.997672 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 9 00:28:26.008480 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 9 00:28:26.011296 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 9 00:28:26.018495 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 9 00:28:26.042324 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 9 00:28:26.054108 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1358) Sep 9 00:28:26.180265 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Sep 9 00:28:26.186211 kernel: ACPI: button: Power Button [PWRF] Sep 9 00:28:26.195106 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Sep 9 00:28:26.199582 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 9 00:28:26.199907 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Sep 9 00:28:26.200175 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 9 00:28:26.217166 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 9 00:28:26.230304 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 9 00:28:26.239341 systemd-networkd[1384]: lo: Link UP Sep 9 00:28:26.239353 systemd-networkd[1384]: lo: Gained carrier Sep 9 00:28:26.241038 systemd-networkd[1384]: Enumeration completed Sep 9 00:28:26.241163 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 9 00:28:26.246239 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 9 00:28:26.247855 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 9 00:28:26.255563 systemd-networkd[1384]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 00:28:26.255577 systemd-networkd[1384]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 9 00:28:26.256398 systemd-networkd[1384]: eth0: Link UP Sep 9 00:28:26.256403 systemd-networkd[1384]: eth0: Gained carrier Sep 9 00:28:26.256415 systemd-networkd[1384]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 9 00:28:26.298128 systemd-networkd[1384]: eth0: DHCPv4 address 10.0.0.113/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 9 00:28:26.315995 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 9 00:28:26.321453 systemd-resolved[1329]: Positive Trust Anchors: Sep 9 00:28:26.326703 kernel: mousedev: PS/2 mouse device common for all mice Sep 9 00:28:26.321668 systemd-resolved[1329]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 9 00:28:26.321700 systemd-resolved[1329]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 9 00:28:26.326425 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 9 00:28:26.327128 systemd-resolved[1329]: Defaulting to hostname 'linux'. Sep 9 00:28:26.328000 systemd-timesyncd[1401]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 9 00:28:26.328944 systemd[1]: Reached target time-set.target - System Time Set. Sep 9 00:28:26.329418 systemd-timesyncd[1401]: Initial clock synchronization to Tue 2025-09-09 00:28:26.088266 UTC. Sep 9 00:28:26.330136 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 9 00:28:26.332168 systemd[1]: Reached target network.target - Network. Sep 9 00:28:26.334166 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 9 00:28:26.340221 kernel: kvm_amd: TSC scaling supported Sep 9 00:28:26.340257 kernel: kvm_amd: Nested Virtualization enabled Sep 9 00:28:26.340271 kernel: kvm_amd: Nested Paging enabled Sep 9 00:28:26.340284 kernel: kvm_amd: LBR virtualization supported Sep 9 00:28:26.341288 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Sep 9 00:28:26.341323 kernel: kvm_amd: Virtual GIF supported Sep 9 00:28:26.366107 kernel: EDAC MC: Ver: 3.0.0 Sep 9 00:28:26.413906 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 9 00:28:26.439280 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 9 00:28:26.464604 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 9 00:28:26.473708 lvm[1424]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 9 00:28:26.511626 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 9 00:28:26.552921 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 9 00:28:26.554064 systemd[1]: Reached target sysinit.target - System Initialization. Sep 9 00:28:26.555307 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 9 00:28:26.556654 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 9 00:28:26.613761 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 9 00:28:26.615068 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 9 00:28:26.616337 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 9 00:28:26.617584 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 9 00:28:26.617617 systemd[1]: Reached target paths.target - Path Units. Sep 9 00:28:26.618563 systemd[1]: Reached target timers.target - Timer Units. Sep 9 00:28:26.620386 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 9 00:28:26.623263 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 9 00:28:26.633019 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 9 00:28:26.635737 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 9 00:28:26.637454 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 9 00:28:26.650590 systemd[1]: Reached target sockets.target - Socket Units. Sep 9 00:28:26.651604 systemd[1]: Reached target basic.target - Basic System. Sep 9 00:28:26.652617 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 9 00:28:26.652646 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 9 00:28:26.653935 systemd[1]: Starting containerd.service - containerd container runtime... Sep 9 00:28:26.655841 lvm[1429]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 9 00:28:26.658230 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 9 00:28:26.661290 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 9 00:28:26.667015 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 9 00:28:26.669463 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 9 00:28:26.671585 jq[1432]: false Sep 9 00:28:26.672031 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 9 00:28:26.675323 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 9 00:28:26.677523 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 9 00:28:26.682260 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 9 00:28:26.692270 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 9 00:28:26.694234 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 9 00:28:26.694840 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 9 00:28:26.696275 systemd[1]: Starting update-engine.service - Update Engine... Sep 9 00:28:26.699406 extend-filesystems[1433]: Found loop3 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found loop4 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found loop5 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found sr0 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda1 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda2 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda3 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found usr Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda4 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda6 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda7 Sep 9 00:28:26.740013 extend-filesystems[1433]: Found vda9 Sep 9 00:28:26.740013 extend-filesystems[1433]: Checking size of /dev/vda9 Sep 9 00:28:26.790001 dbus-daemon[1431]: [system] SELinux support is enabled Sep 9 00:28:26.740836 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 9 00:28:26.791973 update_engine[1445]: I20250909 00:28:26.762025 1445 main.cc:92] Flatcar Update Engine starting Sep 9 00:28:26.745137 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 9 00:28:26.792349 jq[1447]: true Sep 9 00:28:26.748751 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 9 00:28:26.748983 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 9 00:28:26.749396 systemd[1]: motdgen.service: Deactivated successfully. Sep 9 00:28:26.749601 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 9 00:28:26.752752 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 9 00:28:26.752971 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 9 00:28:26.791756 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 9 00:28:26.796122 update_engine[1445]: I20250909 00:28:26.795624 1445 update_check_scheduler.cc:74] Next update check in 9m14s Sep 9 00:28:26.802106 jq[1453]: true Sep 9 00:28:26.806008 (ntainerd)[1457]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 9 00:28:26.811458 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 9 00:28:26.811539 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 9 00:28:26.813016 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 9 00:28:26.813041 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 9 00:28:26.813759 systemd-logind[1441]: Watching system buttons on /dev/input/event1 (Power Button) Sep 9 00:28:26.813785 systemd-logind[1441]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 9 00:28:26.815299 systemd-logind[1441]: New seat seat0. Sep 9 00:28:26.815655 systemd[1]: Started update-engine.service - Update Engine. Sep 9 00:28:26.821269 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 9 00:28:26.822449 systemd[1]: Started systemd-logind.service - User Login Management. Sep 9 00:28:26.865306 extend-filesystems[1433]: Resized partition /dev/vda9 Sep 9 00:28:26.888136 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1379) Sep 9 00:28:27.012687 sshd_keygen[1455]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 9 00:28:27.043808 extend-filesystems[1491]: resize2fs 1.47.1 (20-May-2024) Sep 9 00:28:27.046484 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 9 00:28:27.086118 locksmithd[1467]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 9 00:28:27.091637 tar[1450]: linux-amd64/LICENSE Sep 9 00:28:27.091931 tar[1450]: linux-amd64/helm Sep 9 00:28:27.132474 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 9 00:28:27.174573 systemd[1]: issuegen.service: Deactivated successfully. Sep 9 00:28:27.174916 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 9 00:28:27.177272 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 9 00:28:27.186391 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 9 00:28:27.343336 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 9 00:28:27.378532 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 9 00:28:27.381325 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 9 00:28:27.425209 systemd[1]: Reached target getty.target - Login Prompts. Sep 9 00:28:27.493709 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 9 00:28:27.529346 systemd[1]: Started sshd@0-10.0.0.113:22-10.0.0.1:59436.service - OpenSSH per-connection server daemon (10.0.0.1:59436). Sep 9 00:28:27.736645 systemd-networkd[1384]: eth0: Gained IPv6LL Sep 9 00:28:27.741585 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 9 00:28:27.743570 systemd[1]: Reached target network-online.target - Network is Online. Sep 9 00:28:27.768289 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 9 00:28:27.769463 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 9 00:28:27.895331 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:28:28.171144 sshd[1507]: Connection closed by authenticating user core 10.0.0.1 port 59436 [preauth] Sep 9 00:28:27.899623 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 9 00:28:27.924256 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 9 00:28:27.924594 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 9 00:28:27.926630 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 9 00:28:28.172748 containerd[1457]: time="2025-09-09T00:28:28.172438659Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Sep 9 00:28:28.115377 systemd[1]: sshd@0-10.0.0.113:22-10.0.0.1:59436.service: Deactivated successfully. Sep 9 00:28:28.175121 extend-filesystems[1491]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 9 00:28:28.175121 extend-filesystems[1491]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 9 00:28:28.175121 extend-filesystems[1491]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 9 00:28:28.183210 extend-filesystems[1433]: Resized filesystem in /dev/vda9 Sep 9 00:28:28.179944 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 9 00:28:28.186794 bash[1480]: Updated "/home/core/.ssh/authorized_keys" Sep 9 00:28:28.181336 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 9 00:28:28.187375 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 9 00:28:28.200209 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 9 00:28:28.206258 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 9 00:28:28.219174 containerd[1457]: time="2025-09-09T00:28:28.219041122Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.222769 containerd[1457]: time="2025-09-09T00:28:28.222680785Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.104-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 9 00:28:28.222769 containerd[1457]: time="2025-09-09T00:28:28.222736026Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 9 00:28:28.222769 containerd[1457]: time="2025-09-09T00:28:28.222761412Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 9 00:28:28.223064 containerd[1457]: time="2025-09-09T00:28:28.222995412Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 9 00:28:28.223064 containerd[1457]: time="2025-09-09T00:28:28.223020681Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223155 containerd[1457]: time="2025-09-09T00:28:28.223126852Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223155 containerd[1457]: time="2025-09-09T00:28:28.223141838Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223378 containerd[1457]: time="2025-09-09T00:28:28.223352988Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223378 containerd[1457]: time="2025-09-09T00:28:28.223372120Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223436 containerd[1457]: time="2025-09-09T00:28:28.223386521Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223436 containerd[1457]: time="2025-09-09T00:28:28.223398151Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223570 containerd[1457]: time="2025-09-09T00:28:28.223512838Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.223855 containerd[1457]: time="2025-09-09T00:28:28.223818003Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 9 00:28:28.224039 containerd[1457]: time="2025-09-09T00:28:28.223990780Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 9 00:28:28.224039 containerd[1457]: time="2025-09-09T00:28:28.224017572Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 9 00:28:28.224238 containerd[1457]: time="2025-09-09T00:28:28.224212760Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 9 00:28:28.224320 containerd[1457]: time="2025-09-09T00:28:28.224298910Z" level=info msg="metadata content store policy set" policy=shared Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.263782142Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.263884575Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.263903249Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.263919610Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.263935718Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264184695Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264509888Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264668315Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264685964Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264700111Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264714141Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264732785Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264746473Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.264855 containerd[1457]: time="2025-09-09T00:28:28.264759860Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264786485Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264799325Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264811111Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264826477Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264848400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264861103Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264874147Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264941028Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264964638Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264983254Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.264999830Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.265015656Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.265030281Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265299 containerd[1457]: time="2025-09-09T00:28:28.265056057Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265099005Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265112557Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265128822Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265145095Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265166823Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265181497Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265192589Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265260329Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265281111Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265300809Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265321463Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265334917Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265396872Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 9 00:28:28.265558 containerd[1457]: time="2025-09-09T00:28:28.265410892Z" level=info msg="NRI interface is disabled by configuration." Sep 9 00:28:28.265913 containerd[1457]: time="2025-09-09T00:28:28.265421155Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 9 00:28:28.265939 containerd[1457]: time="2025-09-09T00:28:28.265794166Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 9 00:28:28.265939 containerd[1457]: time="2025-09-09T00:28:28.265870472Z" level=info msg="Connect containerd service" Sep 9 00:28:28.265939 containerd[1457]: time="2025-09-09T00:28:28.265929898Z" level=info msg="using legacy CRI server" Sep 9 00:28:28.265939 containerd[1457]: time="2025-09-09T00:28:28.265941607Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 9 00:28:28.266516 containerd[1457]: time="2025-09-09T00:28:28.266301309Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 9 00:28:28.267249 containerd[1457]: time="2025-09-09T00:28:28.267221357Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 9 00:28:28.268829 containerd[1457]: time="2025-09-09T00:28:28.268185621Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 9 00:28:28.269194 containerd[1457]: time="2025-09-09T00:28:28.268231555Z" level=info msg="Start subscribing containerd event" Sep 9 00:28:28.269280 containerd[1457]: time="2025-09-09T00:28:28.269257890Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 9 00:28:28.269932 containerd[1457]: time="2025-09-09T00:28:28.269276710Z" level=info msg="Start recovering state" Sep 9 00:28:28.269932 containerd[1457]: time="2025-09-09T00:28:28.269568390Z" level=info msg="Start event monitor" Sep 9 00:28:28.269932 containerd[1457]: time="2025-09-09T00:28:28.269598859Z" level=info msg="Start snapshots syncer" Sep 9 00:28:28.269932 containerd[1457]: time="2025-09-09T00:28:28.269609075Z" level=info msg="Start cni network conf syncer for default" Sep 9 00:28:28.269932 containerd[1457]: time="2025-09-09T00:28:28.269618002Z" level=info msg="Start streaming server" Sep 9 00:28:28.269825 systemd[1]: Started containerd.service - containerd container runtime. Sep 9 00:28:28.270443 containerd[1457]: time="2025-09-09T00:28:28.270419947Z" level=info msg="containerd successfully booted in 0.359644s" Sep 9 00:28:28.365947 tar[1450]: linux-amd64/README.md Sep 9 00:28:28.384528 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 9 00:28:29.592134 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:28:29.595040 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 9 00:28:29.596771 systemd[1]: Startup finished in 1.118s (kernel) + 10.490s (initrd) + 6.394s (userspace) = 18.003s. Sep 9 00:28:29.617844 (kubelet)[1550]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 9 00:28:30.454968 kubelet[1550]: E0909 00:28:30.454870 1550 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 9 00:28:30.459367 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 9 00:28:30.459578 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 9 00:28:30.459920 systemd[1]: kubelet.service: Consumed 1.885s CPU time. Sep 9 00:28:37.977979 systemd[1]: Started sshd@1-10.0.0.113:22-10.0.0.1:51826.service - OpenSSH per-connection server daemon (10.0.0.1:51826). Sep 9 00:28:38.017341 sshd[1563]: Accepted publickey for core from 10.0.0.1 port 51826 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:38.019802 sshd[1563]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:38.028703 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 9 00:28:38.041582 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 9 00:28:38.045014 systemd-logind[1441]: New session 1 of user core. Sep 9 00:28:38.056372 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 9 00:28:38.070594 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 9 00:28:38.075116 (systemd)[1567]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 9 00:28:38.209872 systemd[1567]: Queued start job for default target default.target. Sep 9 00:28:38.221580 systemd[1567]: Created slice app.slice - User Application Slice. Sep 9 00:28:38.221612 systemd[1567]: Reached target paths.target - Paths. Sep 9 00:28:38.221627 systemd[1567]: Reached target timers.target - Timers. Sep 9 00:28:38.223559 systemd[1567]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 9 00:28:38.236680 systemd[1567]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 9 00:28:38.236862 systemd[1567]: Reached target sockets.target - Sockets. Sep 9 00:28:38.236887 systemd[1567]: Reached target basic.target - Basic System. Sep 9 00:28:38.236940 systemd[1567]: Reached target default.target - Main User Target. Sep 9 00:28:38.236986 systemd[1567]: Startup finished in 152ms. Sep 9 00:28:38.237489 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 9 00:28:38.250256 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 9 00:28:38.316215 systemd[1]: Started sshd@2-10.0.0.113:22-10.0.0.1:51832.service - OpenSSH per-connection server daemon (10.0.0.1:51832). Sep 9 00:28:38.364683 sshd[1578]: Accepted publickey for core from 10.0.0.1 port 51832 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:38.366636 sshd[1578]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:38.371141 systemd-logind[1441]: New session 2 of user core. Sep 9 00:28:38.380542 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 9 00:28:38.438540 sshd[1578]: pam_unix(sshd:session): session closed for user core Sep 9 00:28:38.451628 systemd[1]: sshd@2-10.0.0.113:22-10.0.0.1:51832.service: Deactivated successfully. Sep 9 00:28:38.454008 systemd[1]: session-2.scope: Deactivated successfully. Sep 9 00:28:38.455856 systemd-logind[1441]: Session 2 logged out. Waiting for processes to exit. Sep 9 00:28:38.463495 systemd[1]: Started sshd@3-10.0.0.113:22-10.0.0.1:51836.service - OpenSSH per-connection server daemon (10.0.0.1:51836). Sep 9 00:28:38.464799 systemd-logind[1441]: Removed session 2. Sep 9 00:28:38.499541 sshd[1585]: Accepted publickey for core from 10.0.0.1 port 51836 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:38.501369 sshd[1585]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:38.505900 systemd-logind[1441]: New session 3 of user core. Sep 9 00:28:38.515256 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 9 00:28:38.566835 sshd[1585]: pam_unix(sshd:session): session closed for user core Sep 9 00:28:38.575141 systemd[1]: sshd@3-10.0.0.113:22-10.0.0.1:51836.service: Deactivated successfully. Sep 9 00:28:38.577161 systemd[1]: session-3.scope: Deactivated successfully. Sep 9 00:28:38.578986 systemd-logind[1441]: Session 3 logged out. Waiting for processes to exit. Sep 9 00:28:38.586474 systemd[1]: Started sshd@4-10.0.0.113:22-10.0.0.1:51846.service - OpenSSH per-connection server daemon (10.0.0.1:51846). Sep 9 00:28:38.587542 systemd-logind[1441]: Removed session 3. Sep 9 00:28:38.617161 sshd[1592]: Accepted publickey for core from 10.0.0.1 port 51846 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:38.619316 sshd[1592]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:38.623788 systemd-logind[1441]: New session 4 of user core. Sep 9 00:28:38.633223 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 9 00:28:38.690692 sshd[1592]: pam_unix(sshd:session): session closed for user core Sep 9 00:28:38.703273 systemd[1]: sshd@4-10.0.0.113:22-10.0.0.1:51846.service: Deactivated successfully. Sep 9 00:28:38.705683 systemd[1]: session-4.scope: Deactivated successfully. Sep 9 00:28:38.707790 systemd-logind[1441]: Session 4 logged out. Waiting for processes to exit. Sep 9 00:28:38.716577 systemd[1]: Started sshd@5-10.0.0.113:22-10.0.0.1:51848.service - OpenSSH per-connection server daemon (10.0.0.1:51848). Sep 9 00:28:38.717878 systemd-logind[1441]: Removed session 4. Sep 9 00:28:38.749818 sshd[1599]: Accepted publickey for core from 10.0.0.1 port 51848 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:38.751903 sshd[1599]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:38.758583 systemd-logind[1441]: New session 5 of user core. Sep 9 00:28:38.768439 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 9 00:28:38.829547 sudo[1602]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 9 00:28:38.829935 sudo[1602]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 00:28:38.851333 sudo[1602]: pam_unix(sudo:session): session closed for user root Sep 9 00:28:38.853281 sshd[1599]: pam_unix(sshd:session): session closed for user core Sep 9 00:28:38.867677 systemd[1]: sshd@5-10.0.0.113:22-10.0.0.1:51848.service: Deactivated successfully. Sep 9 00:28:38.869714 systemd[1]: session-5.scope: Deactivated successfully. Sep 9 00:28:38.871597 systemd-logind[1441]: Session 5 logged out. Waiting for processes to exit. Sep 9 00:28:38.885427 systemd[1]: Started sshd@6-10.0.0.113:22-10.0.0.1:51852.service - OpenSSH per-connection server daemon (10.0.0.1:51852). Sep 9 00:28:38.886461 systemd-logind[1441]: Removed session 5. Sep 9 00:28:38.920692 sshd[1607]: Accepted publickey for core from 10.0.0.1 port 51852 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:38.923117 sshd[1607]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:38.928940 systemd-logind[1441]: New session 6 of user core. Sep 9 00:28:38.944480 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 9 00:28:39.002577 sudo[1611]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 9 00:28:39.002999 sudo[1611]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 00:28:39.007137 sudo[1611]: pam_unix(sudo:session): session closed for user root Sep 9 00:28:39.014558 sudo[1610]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 9 00:28:39.014934 sudo[1610]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 00:28:39.037452 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 9 00:28:39.039710 auditctl[1614]: No rules Sep 9 00:28:39.040250 systemd[1]: audit-rules.service: Deactivated successfully. Sep 9 00:28:39.040567 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 9 00:28:39.043551 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 9 00:28:39.080524 augenrules[1632]: No rules Sep 9 00:28:39.082824 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 9 00:28:39.084394 sudo[1610]: pam_unix(sudo:session): session closed for user root Sep 9 00:28:39.086608 sshd[1607]: pam_unix(sshd:session): session closed for user core Sep 9 00:28:39.096490 systemd[1]: sshd@6-10.0.0.113:22-10.0.0.1:51852.service: Deactivated successfully. Sep 9 00:28:39.098355 systemd[1]: session-6.scope: Deactivated successfully. Sep 9 00:28:39.100022 systemd-logind[1441]: Session 6 logged out. Waiting for processes to exit. Sep 9 00:28:39.110429 systemd[1]: Started sshd@7-10.0.0.113:22-10.0.0.1:51856.service - OpenSSH per-connection server daemon (10.0.0.1:51856). Sep 9 00:28:39.111644 systemd-logind[1441]: Removed session 6. Sep 9 00:28:39.141753 sshd[1640]: Accepted publickey for core from 10.0.0.1 port 51856 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:28:39.143796 sshd[1640]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:28:39.148226 systemd-logind[1441]: New session 7 of user core. Sep 9 00:28:39.158231 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 9 00:28:39.216058 sudo[1643]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 9 00:28:39.216552 sudo[1643]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 9 00:28:39.743392 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 9 00:28:39.743499 (dockerd)[1661]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 9 00:28:40.255932 dockerd[1661]: time="2025-09-09T00:28:40.255833509Z" level=info msg="Starting up" Sep 9 00:28:40.472992 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 9 00:28:40.489473 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:28:40.550353 dockerd[1661]: time="2025-09-09T00:28:40.550152634Z" level=info msg="Loading containers: start." Sep 9 00:28:40.716745 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:28:40.722144 (kubelet)[1709]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 9 00:28:41.402938 kubelet[1709]: E0909 00:28:41.402859 1709 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 9 00:28:41.410115 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 9 00:28:41.410359 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 9 00:28:41.755109 kernel: Initializing XFRM netlink socket Sep 9 00:28:41.858570 systemd-networkd[1384]: docker0: Link UP Sep 9 00:28:41.886171 dockerd[1661]: time="2025-09-09T00:28:41.886095419Z" level=info msg="Loading containers: done." Sep 9 00:28:41.904463 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2829554850-merged.mount: Deactivated successfully. Sep 9 00:28:41.909338 dockerd[1661]: time="2025-09-09T00:28:41.909284782Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 9 00:28:41.909421 dockerd[1661]: time="2025-09-09T00:28:41.909404771Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 9 00:28:41.909559 dockerd[1661]: time="2025-09-09T00:28:41.909537506Z" level=info msg="Daemon has completed initialization" Sep 9 00:28:41.960491 dockerd[1661]: time="2025-09-09T00:28:41.960369072Z" level=info msg="API listen on /run/docker.sock" Sep 9 00:28:41.960718 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 9 00:28:44.463919 containerd[1457]: time="2025-09-09T00:28:44.463818733Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.8\"" Sep 9 00:28:46.408158 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2589741469.mount: Deactivated successfully. Sep 9 00:28:50.417612 containerd[1457]: time="2025-09-09T00:28:50.417466198Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:50.419379 containerd[1457]: time="2025-09-09T00:28:50.419302445Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.8: active requests=0, bytes read=28800687" Sep 9 00:28:50.421261 containerd[1457]: time="2025-09-09T00:28:50.421120663Z" level=info msg="ImageCreate event name:\"sha256:0d4edaa48e2f940c934e0f7cfd5209fc85e65ab5e842b980f41263d1764661f1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:50.425752 containerd[1457]: time="2025-09-09T00:28:50.425668155Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6e1a2f9b24f69ee77d0c0edaf32b31fdbb5e1a613f4476272197e6e1e239050b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:50.426669 containerd[1457]: time="2025-09-09T00:28:50.426606472Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.8\" with image id \"sha256:0d4edaa48e2f940c934e0f7cfd5209fc85e65ab5e842b980f41263d1764661f1\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6e1a2f9b24f69ee77d0c0edaf32b31fdbb5e1a613f4476272197e6e1e239050b\", size \"28797487\" in 5.962677117s" Sep 9 00:28:50.426669 containerd[1457]: time="2025-09-09T00:28:50.426670505Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.8\" returns image reference \"sha256:0d4edaa48e2f940c934e0f7cfd5209fc85e65ab5e842b980f41263d1764661f1\"" Sep 9 00:28:50.427529 containerd[1457]: time="2025-09-09T00:28:50.427481190Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.8\"" Sep 9 00:28:51.698580 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 9 00:28:51.709287 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:28:51.980685 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:28:51.989601 (kubelet)[1891]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 9 00:28:52.341238 containerd[1457]: time="2025-09-09T00:28:52.341041258Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:52.342511 containerd[1457]: time="2025-09-09T00:28:52.342456560Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.8: active requests=0, bytes read=24784128" Sep 9 00:28:52.344589 containerd[1457]: time="2025-09-09T00:28:52.344538844Z" level=info msg="ImageCreate event name:\"sha256:b248d0b0c74ad8230e0bae0cbed477560e8a1e8c7ef5f29b7e75c1f273c8a091\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:52.349140 containerd[1457]: time="2025-09-09T00:28:52.349087576Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:8788ccd28ceed9e2e5f8fc31375ef5771df8ea6e518b362c9a06f3cc709cd6c7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:52.350416 containerd[1457]: time="2025-09-09T00:28:52.350178674Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.8\" with image id \"sha256:b248d0b0c74ad8230e0bae0cbed477560e8a1e8c7ef5f29b7e75c1f273c8a091\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:8788ccd28ceed9e2e5f8fc31375ef5771df8ea6e518b362c9a06f3cc709cd6c7\", size \"26387322\" in 1.922641268s" Sep 9 00:28:52.350416 containerd[1457]: time="2025-09-09T00:28:52.350233148Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.8\" returns image reference \"sha256:b248d0b0c74ad8230e0bae0cbed477560e8a1e8c7ef5f29b7e75c1f273c8a091\"" Sep 9 00:28:52.351372 containerd[1457]: time="2025-09-09T00:28:52.351061702Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.8\"" Sep 9 00:28:52.371031 kubelet[1891]: E0909 00:28:52.370909 1891 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 9 00:28:52.375973 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 9 00:28:52.376287 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 9 00:28:55.577954 containerd[1457]: time="2025-09-09T00:28:55.577845955Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:55.579468 containerd[1457]: time="2025-09-09T00:28:55.579403909Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.8: active requests=0, bytes read=19175036" Sep 9 00:28:55.581467 containerd[1457]: time="2025-09-09T00:28:55.581417804Z" level=info msg="ImageCreate event name:\"sha256:2ac266f06c9a5a3d0d20ae482dbccb54d3be454d5ca49f48b528bdf5bae3e908\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:55.585690 containerd[1457]: time="2025-09-09T00:28:55.585640200Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:43c58bcbd1c7812dd19f8bfa5ae11093ebefd28699453ce86fc710869e155cd4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:55.587349 containerd[1457]: time="2025-09-09T00:28:55.587023731Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.8\" with image id \"sha256:2ac266f06c9a5a3d0d20ae482dbccb54d3be454d5ca49f48b528bdf5bae3e908\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:43c58bcbd1c7812dd19f8bfa5ae11093ebefd28699453ce86fc710869e155cd4\", size \"20778248\" in 3.235906297s" Sep 9 00:28:55.587349 containerd[1457]: time="2025-09-09T00:28:55.587110071Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.8\" returns image reference \"sha256:2ac266f06c9a5a3d0d20ae482dbccb54d3be454d5ca49f48b528bdf5bae3e908\"" Sep 9 00:28:55.587830 containerd[1457]: time="2025-09-09T00:28:55.587779721Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.8\"" Sep 9 00:28:57.361795 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1859958605.mount: Deactivated successfully. Sep 9 00:28:57.788720 containerd[1457]: time="2025-09-09T00:28:57.788304851Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:57.789837 containerd[1457]: time="2025-09-09T00:28:57.789785031Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.8: active requests=0, bytes read=30897170" Sep 9 00:28:57.791485 containerd[1457]: time="2025-09-09T00:28:57.791393864Z" level=info msg="ImageCreate event name:\"sha256:d7b94972d43c5d6ce8088a8bcd08614a5ecf2bf04166232c688adcd0b8ed4b12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:57.795371 containerd[1457]: time="2025-09-09T00:28:57.795300713Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:adc1335b480ddd833aac3b0bd20f68ff0f3c3cf7a0bd337933b006d9f5cec40a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:57.795995 containerd[1457]: time="2025-09-09T00:28:57.795943659Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.8\" with image id \"sha256:d7b94972d43c5d6ce8088a8bcd08614a5ecf2bf04166232c688adcd0b8ed4b12\", repo tag \"registry.k8s.io/kube-proxy:v1.32.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:adc1335b480ddd833aac3b0bd20f68ff0f3c3cf7a0bd337933b006d9f5cec40a\", size \"30896189\" in 2.208127754s" Sep 9 00:28:57.795995 containerd[1457]: time="2025-09-09T00:28:57.795982491Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.8\" returns image reference \"sha256:d7b94972d43c5d6ce8088a8bcd08614a5ecf2bf04166232c688adcd0b8ed4b12\"" Sep 9 00:28:57.796556 containerd[1457]: time="2025-09-09T00:28:57.796517863Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 9 00:28:58.601829 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3277363468.mount: Deactivated successfully. Sep 9 00:28:59.957534 containerd[1457]: time="2025-09-09T00:28:59.957323248Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:59.961547 containerd[1457]: time="2025-09-09T00:28:59.961027481Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Sep 9 00:28:59.962541 containerd[1457]: time="2025-09-09T00:28:59.962482294Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:59.966822 containerd[1457]: time="2025-09-09T00:28:59.966730201Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:28:59.968439 containerd[1457]: time="2025-09-09T00:28:59.968370149Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 2.171816579s" Sep 9 00:28:59.968439 containerd[1457]: time="2025-09-09T00:28:59.968417473Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 9 00:28:59.969376 containerd[1457]: time="2025-09-09T00:28:59.969110742Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 9 00:29:00.510860 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount618142234.mount: Deactivated successfully. Sep 9 00:29:00.517430 containerd[1457]: time="2025-09-09T00:29:00.517372268Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:00.518386 containerd[1457]: time="2025-09-09T00:29:00.518299093Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 9 00:29:00.519572 containerd[1457]: time="2025-09-09T00:29:00.519529095Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:00.522137 containerd[1457]: time="2025-09-09T00:29:00.522088654Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:00.523149 containerd[1457]: time="2025-09-09T00:29:00.523107791Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 553.943814ms" Sep 9 00:29:00.523197 containerd[1457]: time="2025-09-09T00:29:00.523155194Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 9 00:29:00.523984 containerd[1457]: time="2025-09-09T00:29:00.523922955Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Sep 9 00:29:02.092842 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3690341161.mount: Deactivated successfully. Sep 9 00:29:02.626667 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Sep 9 00:29:02.635578 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:29:02.883735 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:29:02.889652 (kubelet)[1988]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 9 00:29:03.062217 kubelet[1988]: E0909 00:29:03.062054 1988 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 9 00:29:03.066644 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 9 00:29:03.066935 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 9 00:29:05.496750 containerd[1457]: time="2025-09-09T00:29:05.496673005Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:05.498040 containerd[1457]: time="2025-09-09T00:29:05.497958918Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Sep 9 00:29:05.499870 containerd[1457]: time="2025-09-09T00:29:05.499814168Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:05.504058 containerd[1457]: time="2025-09-09T00:29:05.504017568Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:05.505745 containerd[1457]: time="2025-09-09T00:29:05.505700734Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 4.981730915s" Sep 9 00:29:05.505745 containerd[1457]: time="2025-09-09T00:29:05.505735711Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Sep 9 00:29:08.327028 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:29:08.337447 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:29:08.369886 systemd[1]: Reloading requested from client PID 2067 ('systemctl') (unit session-7.scope)... Sep 9 00:29:08.369903 systemd[1]: Reloading... Sep 9 00:29:08.485131 zram_generator::config[2112]: No configuration found. Sep 9 00:29:08.874338 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 9 00:29:08.961108 systemd[1]: Reloading finished in 590 ms. Sep 9 00:29:09.013791 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 9 00:29:09.013917 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 9 00:29:09.014301 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:29:09.017994 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:29:09.220135 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:29:09.226396 (kubelet)[2155]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 9 00:29:09.272618 kubelet[2155]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 00:29:09.272618 kubelet[2155]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 9 00:29:09.272618 kubelet[2155]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 00:29:09.272618 kubelet[2155]: I0909 00:29:09.272287 2155 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 9 00:29:09.537810 kubelet[2155]: I0909 00:29:09.537684 2155 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 9 00:29:09.537810 kubelet[2155]: I0909 00:29:09.537717 2155 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 9 00:29:09.538000 kubelet[2155]: I0909 00:29:09.537978 2155 server.go:954] "Client rotation is on, will bootstrap in background" Sep 9 00:29:09.562431 kubelet[2155]: E0909 00:29:09.562366 2155 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.113:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:09.562820 kubelet[2155]: I0909 00:29:09.562792 2155 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 9 00:29:09.570587 kubelet[2155]: E0909 00:29:09.570545 2155 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 9 00:29:09.570587 kubelet[2155]: I0909 00:29:09.570581 2155 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 9 00:29:09.576952 kubelet[2155]: I0909 00:29:09.576907 2155 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 9 00:29:09.578366 kubelet[2155]: I0909 00:29:09.578314 2155 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 9 00:29:09.578561 kubelet[2155]: I0909 00:29:09.578360 2155 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 9 00:29:09.578698 kubelet[2155]: I0909 00:29:09.578569 2155 topology_manager.go:138] "Creating topology manager with none policy" Sep 9 00:29:09.578698 kubelet[2155]: I0909 00:29:09.578579 2155 container_manager_linux.go:304] "Creating device plugin manager" Sep 9 00:29:09.578761 kubelet[2155]: I0909 00:29:09.578734 2155 state_mem.go:36] "Initialized new in-memory state store" Sep 9 00:29:09.582195 kubelet[2155]: I0909 00:29:09.582043 2155 kubelet.go:446] "Attempting to sync node with API server" Sep 9 00:29:09.582195 kubelet[2155]: I0909 00:29:09.582083 2155 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 9 00:29:09.582195 kubelet[2155]: I0909 00:29:09.582106 2155 kubelet.go:352] "Adding apiserver pod source" Sep 9 00:29:09.582195 kubelet[2155]: I0909 00:29:09.582117 2155 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 9 00:29:09.584948 kubelet[2155]: I0909 00:29:09.584908 2155 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 9 00:29:09.585934 kubelet[2155]: I0909 00:29:09.585681 2155 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 9 00:29:09.586468 kubelet[2155]: W0909 00:29:09.586398 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.113:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:09.586529 kubelet[2155]: E0909 00:29:09.586485 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.113:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:09.588040 kubelet[2155]: W0909 00:29:09.587986 2155 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 9 00:29:09.588040 kubelet[2155]: W0909 00:29:09.587982 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.113:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:09.588040 kubelet[2155]: E0909 00:29:09.588045 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.113:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:09.589959 kubelet[2155]: I0909 00:29:09.589936 2155 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 9 00:29:09.590032 kubelet[2155]: I0909 00:29:09.589981 2155 server.go:1287] "Started kubelet" Sep 9 00:29:09.590603 kubelet[2155]: I0909 00:29:09.590546 2155 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 9 00:29:09.591691 kubelet[2155]: I0909 00:29:09.590967 2155 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 9 00:29:09.591691 kubelet[2155]: I0909 00:29:09.590980 2155 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 9 00:29:09.591691 kubelet[2155]: I0909 00:29:09.591628 2155 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 9 00:29:09.591861 kubelet[2155]: I0909 00:29:09.591714 2155 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 9 00:29:09.591861 kubelet[2155]: I0909 00:29:09.591793 2155 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 9 00:29:09.591861 kubelet[2155]: I0909 00:29:09.591845 2155 server.go:479] "Adding debug handlers to kubelet server" Sep 9 00:29:09.594385 kubelet[2155]: E0909 00:29:09.594007 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:09.594464 kubelet[2155]: W0909 00:29:09.594429 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.113:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:09.594512 kubelet[2155]: E0909 00:29:09.594467 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.113:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:09.595097 kubelet[2155]: I0909 00:29:09.595052 2155 reconciler.go:26] "Reconciler: start to sync state" Sep 9 00:29:09.595097 kubelet[2155]: I0909 00:29:09.595090 2155 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 9 00:29:09.595365 kubelet[2155]: E0909 00:29:09.595325 2155 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.113:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.113:6443: connect: connection refused" interval="200ms" Sep 9 00:29:09.596680 kubelet[2155]: I0909 00:29:09.595771 2155 factory.go:221] Registration of the containerd container factory successfully Sep 9 00:29:09.596680 kubelet[2155]: I0909 00:29:09.595788 2155 factory.go:221] Registration of the systemd container factory successfully Sep 9 00:29:09.596680 kubelet[2155]: I0909 00:29:09.595844 2155 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 9 00:29:09.596680 kubelet[2155]: E0909 00:29:09.594966 2155 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.113:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.113:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.186375c0947c1ae9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-09 00:29:09.589957353 +0000 UTC m=+0.358978969,LastTimestamp:2025-09-09 00:29:09.589957353 +0000 UTC m=+0.358978969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 9 00:29:09.596899 kubelet[2155]: E0909 00:29:09.596878 2155 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 9 00:29:09.609161 kubelet[2155]: I0909 00:29:09.609055 2155 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 9 00:29:09.609161 kubelet[2155]: I0909 00:29:09.609102 2155 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 9 00:29:09.609161 kubelet[2155]: I0909 00:29:09.609123 2155 state_mem.go:36] "Initialized new in-memory state store" Sep 9 00:29:09.694392 kubelet[2155]: E0909 00:29:09.694305 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:09.795117 kubelet[2155]: E0909 00:29:09.794961 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:09.796552 kubelet[2155]: E0909 00:29:09.796521 2155 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.113:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.113:6443: connect: connection refused" interval="400ms" Sep 9 00:29:09.895841 kubelet[2155]: E0909 00:29:09.895781 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:09.996415 kubelet[2155]: E0909 00:29:09.996340 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:10.096672 kubelet[2155]: E0909 00:29:10.096609 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:10.197374 kubelet[2155]: E0909 00:29:10.197278 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:10.197724 kubelet[2155]: E0909 00:29:10.197637 2155 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.113:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.113:6443: connect: connection refused" interval="800ms" Sep 9 00:29:10.298409 kubelet[2155]: E0909 00:29:10.298300 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:10.399309 kubelet[2155]: E0909 00:29:10.399140 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:10.457698 kubelet[2155]: I0909 00:29:10.457653 2155 policy_none.go:49] "None policy: Start" Sep 9 00:29:10.457698 kubelet[2155]: I0909 00:29:10.457695 2155 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 9 00:29:10.457908 kubelet[2155]: I0909 00:29:10.457721 2155 state_mem.go:35] "Initializing new in-memory state store" Sep 9 00:29:10.463595 kubelet[2155]: I0909 00:29:10.463546 2155 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 9 00:29:10.465301 kubelet[2155]: I0909 00:29:10.465123 2155 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 9 00:29:10.465301 kubelet[2155]: I0909 00:29:10.465184 2155 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 9 00:29:10.465301 kubelet[2155]: I0909 00:29:10.465215 2155 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 9 00:29:10.465301 kubelet[2155]: I0909 00:29:10.465225 2155 kubelet.go:2382] "Starting kubelet main sync loop" Sep 9 00:29:10.465500 kubelet[2155]: E0909 00:29:10.465316 2155 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 9 00:29:10.467320 kubelet[2155]: W0909 00:29:10.467264 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.113:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:10.467576 kubelet[2155]: E0909 00:29:10.467335 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.113:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:10.471848 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 9 00:29:10.486915 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 9 00:29:10.490677 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 9 00:29:10.499527 kubelet[2155]: E0909 00:29:10.499483 2155 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:10.505554 kubelet[2155]: I0909 00:29:10.505486 2155 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 9 00:29:10.505843 kubelet[2155]: I0909 00:29:10.505822 2155 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 9 00:29:10.505907 kubelet[2155]: I0909 00:29:10.505843 2155 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 9 00:29:10.506385 kubelet[2155]: I0909 00:29:10.506207 2155 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 9 00:29:10.507327 kubelet[2155]: E0909 00:29:10.507289 2155 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 9 00:29:10.507327 kubelet[2155]: E0909 00:29:10.507341 2155 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 9 00:29:10.577221 systemd[1]: Created slice kubepods-burstable-poda9176403b596d0b29ae8ad12d635226d.slice - libcontainer container kubepods-burstable-poda9176403b596d0b29ae8ad12d635226d.slice. Sep 9 00:29:10.596133 kubelet[2155]: E0909 00:29:10.596003 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:10.600783 kubelet[2155]: I0909 00:29:10.600736 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:10.600783 kubelet[2155]: I0909 00:29:10.600779 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:10.601027 kubelet[2155]: I0909 00:29:10.600808 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a9176403b596d0b29ae8ad12d635226d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a9176403b596d0b29ae8ad12d635226d\") " pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:10.601027 kubelet[2155]: I0909 00:29:10.600831 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8caf8dcf530eddbba5e07f694f7fc047-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"8caf8dcf530eddbba5e07f694f7fc047\") " pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:10.601027 kubelet[2155]: I0909 00:29:10.600856 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8caf8dcf530eddbba5e07f694f7fc047-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"8caf8dcf530eddbba5e07f694f7fc047\") " pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:10.601027 kubelet[2155]: I0909 00:29:10.600878 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8caf8dcf530eddbba5e07f694f7fc047-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"8caf8dcf530eddbba5e07f694f7fc047\") " pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:10.601027 kubelet[2155]: I0909 00:29:10.600902 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:10.600944 systemd[1]: Created slice kubepods-burstable-pod8caf8dcf530eddbba5e07f694f7fc047.slice - libcontainer container kubepods-burstable-pod8caf8dcf530eddbba5e07f694f7fc047.slice. Sep 9 00:29:10.601316 kubelet[2155]: I0909 00:29:10.600945 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:10.601316 kubelet[2155]: I0909 00:29:10.600973 2155 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:10.603012 kubelet[2155]: E0909 00:29:10.602989 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:10.604742 systemd[1]: Created slice kubepods-burstable-poda88c9297c136b0f15880bf567e89a977.slice - libcontainer container kubepods-burstable-poda88c9297c136b0f15880bf567e89a977.slice. Sep 9 00:29:10.606898 kubelet[2155]: E0909 00:29:10.606866 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:10.607419 kubelet[2155]: I0909 00:29:10.607364 2155 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 00:29:10.607886 kubelet[2155]: E0909 00:29:10.607849 2155 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.113:6443/api/v1/nodes\": dial tcp 10.0.0.113:6443: connect: connection refused" node="localhost" Sep 9 00:29:10.632742 kubelet[2155]: W0909 00:29:10.632652 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.113:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:10.632742 kubelet[2155]: E0909 00:29:10.632750 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.113:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:10.794451 kubelet[2155]: W0909 00:29:10.794240 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.113:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:10.794451 kubelet[2155]: E0909 00:29:10.794316 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.113:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:10.810013 kubelet[2155]: I0909 00:29:10.809988 2155 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 00:29:10.810358 kubelet[2155]: E0909 00:29:10.810328 2155 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.113:6443/api/v1/nodes\": dial tcp 10.0.0.113:6443: connect: connection refused" node="localhost" Sep 9 00:29:10.897108 kubelet[2155]: E0909 00:29:10.897023 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:10.897967 containerd[1457]: time="2025-09-09T00:29:10.897906608Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a9176403b596d0b29ae8ad12d635226d,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:10.904216 kubelet[2155]: E0909 00:29:10.904180 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:10.904699 containerd[1457]: time="2025-09-09T00:29:10.904661803Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:8caf8dcf530eddbba5e07f694f7fc047,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:10.907941 kubelet[2155]: E0909 00:29:10.907901 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:10.908412 containerd[1457]: time="2025-09-09T00:29:10.908360579Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:a88c9297c136b0f15880bf567e89a977,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:10.998590 kubelet[2155]: E0909 00:29:10.998518 2155 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.113:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.113:6443: connect: connection refused" interval="1.6s" Sep 9 00:29:11.142787 kubelet[2155]: W0909 00:29:11.142683 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.113:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:11.142787 kubelet[2155]: E0909 00:29:11.142771 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.113:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:11.212093 kubelet[2155]: I0909 00:29:11.212039 2155 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 00:29:11.212508 kubelet[2155]: E0909 00:29:11.212471 2155 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.113:6443/api/v1/nodes\": dial tcp 10.0.0.113:6443: connect: connection refused" node="localhost" Sep 9 00:29:11.442791 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount169899248.mount: Deactivated successfully. Sep 9 00:29:11.450069 containerd[1457]: time="2025-09-09T00:29:11.449987321Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 00:29:11.451345 containerd[1457]: time="2025-09-09T00:29:11.451275213Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 00:29:11.452050 containerd[1457]: time="2025-09-09T00:29:11.452002374Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Sep 9 00:29:11.453256 containerd[1457]: time="2025-09-09T00:29:11.453208869Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 00:29:11.454155 containerd[1457]: time="2025-09-09T00:29:11.454117159Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 9 00:29:11.455244 containerd[1457]: time="2025-09-09T00:29:11.455190577Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 00:29:11.456405 containerd[1457]: time="2025-09-09T00:29:11.456341745Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 9 00:29:11.459800 containerd[1457]: time="2025-09-09T00:29:11.459745203Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 9 00:29:11.462540 containerd[1457]: time="2025-09-09T00:29:11.462477869Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 554.020682ms" Sep 9 00:29:11.463441 containerd[1457]: time="2025-09-09T00:29:11.463370589Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 558.631857ms" Sep 9 00:29:11.464170 containerd[1457]: time="2025-09-09T00:29:11.464126034Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 566.113743ms" Sep 9 00:29:11.595571 update_engine[1445]: I20250909 00:29:11.595175 1445 update_attempter.cc:509] Updating boot flags... Sep 9 00:29:11.623647 containerd[1457]: time="2025-09-09T00:29:11.622504750Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:11.623647 containerd[1457]: time="2025-09-09T00:29:11.622674577Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:11.623647 containerd[1457]: time="2025-09-09T00:29:11.622760031Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:11.623647 containerd[1457]: time="2025-09-09T00:29:11.623042927Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:11.629114 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (2234) Sep 9 00:29:11.641850 containerd[1457]: time="2025-09-09T00:29:11.639343915Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:11.641850 containerd[1457]: time="2025-09-09T00:29:11.639425502Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:11.641850 containerd[1457]: time="2025-09-09T00:29:11.639450320Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:11.641850 containerd[1457]: time="2025-09-09T00:29:11.639785386Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:11.655899 containerd[1457]: time="2025-09-09T00:29:11.655546144Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:11.655899 containerd[1457]: time="2025-09-09T00:29:11.655615789Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:11.655899 containerd[1457]: time="2025-09-09T00:29:11.655636518Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:11.655899 containerd[1457]: time="2025-09-09T00:29:11.655750438Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:11.678108 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (2185) Sep 9 00:29:11.703395 kubelet[2155]: E0909 00:29:11.703248 2155 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.113:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:11.756363 systemd[1]: Started cri-containerd-3a241a5d53923cb9cbab1f62624f7e42539f546c407b67ada63decd72d3d3984.scope - libcontainer container 3a241a5d53923cb9cbab1f62624f7e42539f546c407b67ada63decd72d3d3984. Sep 9 00:29:11.758909 systemd[1]: Started cri-containerd-46bb25226b5a0e001f9fc44dbc0b04674a8d01f2f1d48e59e8cf66a34f5f7069.scope - libcontainer container 46bb25226b5a0e001f9fc44dbc0b04674a8d01f2f1d48e59e8cf66a34f5f7069. Sep 9 00:29:11.761334 systemd[1]: Started cri-containerd-53654b4e06b02d6ee3a2e5cc565770a83a9149d66c005fbb7098edf563e5bc6a.scope - libcontainer container 53654b4e06b02d6ee3a2e5cc565770a83a9149d66c005fbb7098edf563e5bc6a. Sep 9 00:29:11.809116 containerd[1457]: time="2025-09-09T00:29:11.809026400Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:a88c9297c136b0f15880bf567e89a977,Namespace:kube-system,Attempt:0,} returns sandbox id \"46bb25226b5a0e001f9fc44dbc0b04674a8d01f2f1d48e59e8cf66a34f5f7069\"" Sep 9 00:29:11.812096 kubelet[2155]: E0909 00:29:11.811468 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:11.815957 containerd[1457]: time="2025-09-09T00:29:11.815922789Z" level=info msg="CreateContainer within sandbox \"46bb25226b5a0e001f9fc44dbc0b04674a8d01f2f1d48e59e8cf66a34f5f7069\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 9 00:29:11.816601 containerd[1457]: time="2025-09-09T00:29:11.816566880Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a9176403b596d0b29ae8ad12d635226d,Namespace:kube-system,Attempt:0,} returns sandbox id \"53654b4e06b02d6ee3a2e5cc565770a83a9149d66c005fbb7098edf563e5bc6a\"" Sep 9 00:29:11.819458 kubelet[2155]: E0909 00:29:11.819397 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:11.821758 containerd[1457]: time="2025-09-09T00:29:11.821698377Z" level=info msg="CreateContainer within sandbox \"53654b4e06b02d6ee3a2e5cc565770a83a9149d66c005fbb7098edf563e5bc6a\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 9 00:29:11.822216 containerd[1457]: time="2025-09-09T00:29:11.822185335Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:8caf8dcf530eddbba5e07f694f7fc047,Namespace:kube-system,Attempt:0,} returns sandbox id \"3a241a5d53923cb9cbab1f62624f7e42539f546c407b67ada63decd72d3d3984\"" Sep 9 00:29:11.833768 kubelet[2155]: W0909 00:29:11.833717 2155 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.113:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.113:6443: connect: connection refused Sep 9 00:29:11.833768 kubelet[2155]: E0909 00:29:11.833771 2155 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.113:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.113:6443: connect: connection refused" logger="UnhandledError" Sep 9 00:29:11.834000 kubelet[2155]: E0909 00:29:11.833951 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:11.836508 containerd[1457]: time="2025-09-09T00:29:11.836444881Z" level=info msg="CreateContainer within sandbox \"3a241a5d53923cb9cbab1f62624f7e42539f546c407b67ada63decd72d3d3984\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 9 00:29:11.842715 containerd[1457]: time="2025-09-09T00:29:11.842644176Z" level=info msg="CreateContainer within sandbox \"46bb25226b5a0e001f9fc44dbc0b04674a8d01f2f1d48e59e8cf66a34f5f7069\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"4a2c0fc54b62d39d2f1e28f5ad0f0872733ee9db37c3f81ae640a7bfee444a7c\"" Sep 9 00:29:11.843489 containerd[1457]: time="2025-09-09T00:29:11.843451431Z" level=info msg="StartContainer for \"4a2c0fc54b62d39d2f1e28f5ad0f0872733ee9db37c3f81ae640a7bfee444a7c\"" Sep 9 00:29:11.856570 containerd[1457]: time="2025-09-09T00:29:11.856495865Z" level=info msg="CreateContainer within sandbox \"53654b4e06b02d6ee3a2e5cc565770a83a9149d66c005fbb7098edf563e5bc6a\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f9bd9a7a9af57b45b1d0a9e819be9c73a44619a558c65a8dc895d96ede015cde\"" Sep 9 00:29:11.857590 containerd[1457]: time="2025-09-09T00:29:11.857526792Z" level=info msg="StartContainer for \"f9bd9a7a9af57b45b1d0a9e819be9c73a44619a558c65a8dc895d96ede015cde\"" Sep 9 00:29:11.860128 containerd[1457]: time="2025-09-09T00:29:11.859262787Z" level=info msg="CreateContainer within sandbox \"3a241a5d53923cb9cbab1f62624f7e42539f546c407b67ada63decd72d3d3984\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"ff6fde59966932495e92aec07e810323cbc43ac8ab7cf82d576d7eba9d9f78e9\"" Sep 9 00:29:11.861469 containerd[1457]: time="2025-09-09T00:29:11.861419723Z" level=info msg="StartContainer for \"ff6fde59966932495e92aec07e810323cbc43ac8ab7cf82d576d7eba9d9f78e9\"" Sep 9 00:29:11.881255 systemd[1]: Started cri-containerd-4a2c0fc54b62d39d2f1e28f5ad0f0872733ee9db37c3f81ae640a7bfee444a7c.scope - libcontainer container 4a2c0fc54b62d39d2f1e28f5ad0f0872733ee9db37c3f81ae640a7bfee444a7c. Sep 9 00:29:11.898347 systemd[1]: Started cri-containerd-f9bd9a7a9af57b45b1d0a9e819be9c73a44619a558c65a8dc895d96ede015cde.scope - libcontainer container f9bd9a7a9af57b45b1d0a9e819be9c73a44619a558c65a8dc895d96ede015cde. Sep 9 00:29:11.910450 systemd[1]: Started cri-containerd-ff6fde59966932495e92aec07e810323cbc43ac8ab7cf82d576d7eba9d9f78e9.scope - libcontainer container ff6fde59966932495e92aec07e810323cbc43ac8ab7cf82d576d7eba9d9f78e9. Sep 9 00:29:11.959980 containerd[1457]: time="2025-09-09T00:29:11.959733855Z" level=info msg="StartContainer for \"4a2c0fc54b62d39d2f1e28f5ad0f0872733ee9db37c3f81ae640a7bfee444a7c\" returns successfully" Sep 9 00:29:12.021206 kubelet[2155]: I0909 00:29:12.021154 2155 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 00:29:12.022092 kubelet[2155]: E0909 00:29:12.022042 2155 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.113:6443/api/v1/nodes\": dial tcp 10.0.0.113:6443: connect: connection refused" node="localhost" Sep 9 00:29:12.193341 containerd[1457]: time="2025-09-09T00:29:12.193145936Z" level=info msg="StartContainer for \"ff6fde59966932495e92aec07e810323cbc43ac8ab7cf82d576d7eba9d9f78e9\" returns successfully" Sep 9 00:29:12.193496 containerd[1457]: time="2025-09-09T00:29:12.193230388Z" level=info msg="StartContainer for \"f9bd9a7a9af57b45b1d0a9e819be9c73a44619a558c65a8dc895d96ede015cde\" returns successfully" Sep 9 00:29:12.474925 kubelet[2155]: E0909 00:29:12.474879 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:12.475113 kubelet[2155]: E0909 00:29:12.475015 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:12.476601 kubelet[2155]: E0909 00:29:12.476554 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:12.476689 kubelet[2155]: E0909 00:29:12.476682 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:12.478121 kubelet[2155]: E0909 00:29:12.478065 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:12.478246 kubelet[2155]: E0909 00:29:12.478221 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:13.207107 kubelet[2155]: E0909 00:29:13.204244 2155 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 9 00:29:13.343104 kubelet[2155]: E0909 00:29:13.342916 2155 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.186375c0947c1ae9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-09 00:29:09.589957353 +0000 UTC m=+0.358978969,LastTimestamp:2025-09-09 00:29:09.589957353 +0000 UTC m=+0.358978969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 9 00:29:13.481142 kubelet[2155]: E0909 00:29:13.480985 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:13.481266 kubelet[2155]: E0909 00:29:13.481154 2155 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Sep 9 00:29:13.481266 kubelet[2155]: E0909 00:29:13.481184 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:13.481550 kubelet[2155]: E0909 00:29:13.481324 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:13.585908 kubelet[2155]: I0909 00:29:13.585835 2155 apiserver.go:52] "Watching apiserver" Sep 9 00:29:13.595615 kubelet[2155]: I0909 00:29:13.595552 2155 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 9 00:29:13.624702 kubelet[2155]: I0909 00:29:13.624654 2155 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 00:29:13.753909 kubelet[2155]: I0909 00:29:13.753767 2155 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 9 00:29:13.753909 kubelet[2155]: E0909 00:29:13.753809 2155 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 9 00:29:13.795855 kubelet[2155]: I0909 00:29:13.795795 2155 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:13.915700 kubelet[2155]: E0909 00:29:13.915638 2155 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:13.915700 kubelet[2155]: I0909 00:29:13.915674 2155 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:13.917998 kubelet[2155]: E0909 00:29:13.917902 2155 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:13.917998 kubelet[2155]: I0909 00:29:13.917933 2155 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:13.920009 kubelet[2155]: E0909 00:29:13.919978 2155 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:14.481923 kubelet[2155]: I0909 00:29:14.481880 2155 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:14.490327 kubelet[2155]: E0909 00:29:14.490132 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:15.195588 systemd[1]: Reloading requested from client PID 2449 ('systemctl') (unit session-7.scope)... Sep 9 00:29:15.195606 systemd[1]: Reloading... Sep 9 00:29:15.286134 zram_generator::config[2488]: No configuration found. Sep 9 00:29:15.483722 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 9 00:29:15.484140 kubelet[2155]: E0909 00:29:15.483959 2155 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:15.599730 systemd[1]: Reloading finished in 403 ms. Sep 9 00:29:15.656013 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:29:15.676458 systemd[1]: kubelet.service: Deactivated successfully. Sep 9 00:29:15.676894 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:29:15.686498 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 9 00:29:15.879903 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 9 00:29:15.894673 (kubelet)[2533]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 9 00:29:15.949812 kubelet[2533]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 00:29:15.949812 kubelet[2533]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 9 00:29:15.949812 kubelet[2533]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 9 00:29:15.950390 kubelet[2533]: I0909 00:29:15.949934 2533 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 9 00:29:15.959679 kubelet[2533]: I0909 00:29:15.959613 2533 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 9 00:29:15.959679 kubelet[2533]: I0909 00:29:15.959649 2533 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 9 00:29:15.959945 kubelet[2533]: I0909 00:29:15.959911 2533 server.go:954] "Client rotation is on, will bootstrap in background" Sep 9 00:29:15.961451 kubelet[2533]: I0909 00:29:15.961414 2533 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 9 00:29:15.964397 kubelet[2533]: I0909 00:29:15.964348 2533 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 9 00:29:15.969163 kubelet[2533]: E0909 00:29:15.969115 2533 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 9 00:29:15.969163 kubelet[2533]: I0909 00:29:15.969155 2533 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 9 00:29:15.975401 kubelet[2533]: I0909 00:29:15.975359 2533 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 9 00:29:15.975782 kubelet[2533]: I0909 00:29:15.975728 2533 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 9 00:29:15.976015 kubelet[2533]: I0909 00:29:15.975772 2533 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 9 00:29:15.976115 kubelet[2533]: I0909 00:29:15.976021 2533 topology_manager.go:138] "Creating topology manager with none policy" Sep 9 00:29:15.976115 kubelet[2533]: I0909 00:29:15.976032 2533 container_manager_linux.go:304] "Creating device plugin manager" Sep 9 00:29:15.976115 kubelet[2533]: I0909 00:29:15.976113 2533 state_mem.go:36] "Initialized new in-memory state store" Sep 9 00:29:15.976341 kubelet[2533]: I0909 00:29:15.976321 2533 kubelet.go:446] "Attempting to sync node with API server" Sep 9 00:29:15.976370 kubelet[2533]: I0909 00:29:15.976349 2533 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 9 00:29:15.976370 kubelet[2533]: I0909 00:29:15.976369 2533 kubelet.go:352] "Adding apiserver pod source" Sep 9 00:29:15.976407 kubelet[2533]: I0909 00:29:15.976380 2533 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 9 00:29:15.977613 kubelet[2533]: I0909 00:29:15.977577 2533 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 9 00:29:15.978146 kubelet[2533]: I0909 00:29:15.978056 2533 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 9 00:29:15.978747 kubelet[2533]: I0909 00:29:15.978672 2533 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 9 00:29:15.978747 kubelet[2533]: I0909 00:29:15.978710 2533 server.go:1287] "Started kubelet" Sep 9 00:29:15.979627 kubelet[2533]: I0909 00:29:15.979518 2533 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 9 00:29:15.980894 kubelet[2533]: I0909 00:29:15.980861 2533 server.go:479] "Adding debug handlers to kubelet server" Sep 9 00:29:15.982386 kubelet[2533]: I0909 00:29:15.982329 2533 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 9 00:29:15.982649 kubelet[2533]: I0909 00:29:15.982629 2533 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 9 00:29:15.984618 kubelet[2533]: I0909 00:29:15.984589 2533 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 9 00:29:15.985037 kubelet[2533]: I0909 00:29:15.984996 2533 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 9 00:29:15.992100 kubelet[2533]: I0909 00:29:15.991252 2533 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 9 00:29:15.992100 kubelet[2533]: E0909 00:29:15.991454 2533 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 9 00:29:15.992100 kubelet[2533]: I0909 00:29:15.991606 2533 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 9 00:29:15.992100 kubelet[2533]: I0909 00:29:15.991758 2533 reconciler.go:26] "Reconciler: start to sync state" Sep 9 00:29:15.993316 kubelet[2533]: I0909 00:29:15.993293 2533 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 9 00:29:15.995922 kubelet[2533]: E0909 00:29:15.995585 2533 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 9 00:29:15.996353 kubelet[2533]: I0909 00:29:15.996335 2533 factory.go:221] Registration of the containerd container factory successfully Sep 9 00:29:15.996433 kubelet[2533]: I0909 00:29:15.996422 2533 factory.go:221] Registration of the systemd container factory successfully Sep 9 00:29:16.014804 kubelet[2533]: I0909 00:29:16.014731 2533 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 9 00:29:16.023663 kubelet[2533]: I0909 00:29:16.023206 2533 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 9 00:29:16.023663 kubelet[2533]: I0909 00:29:16.023248 2533 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 9 00:29:16.023663 kubelet[2533]: I0909 00:29:16.023278 2533 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 9 00:29:16.023663 kubelet[2533]: I0909 00:29:16.023286 2533 kubelet.go:2382] "Starting kubelet main sync loop" Sep 9 00:29:16.023663 kubelet[2533]: E0909 00:29:16.023359 2533 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 9 00:29:16.058087 kubelet[2533]: I0909 00:29:16.057981 2533 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 9 00:29:16.058087 kubelet[2533]: I0909 00:29:16.058047 2533 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 9 00:29:16.058257 kubelet[2533]: I0909 00:29:16.058105 2533 state_mem.go:36] "Initialized new in-memory state store" Sep 9 00:29:16.058473 kubelet[2533]: I0909 00:29:16.058405 2533 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 9 00:29:16.058473 kubelet[2533]: I0909 00:29:16.058427 2533 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 9 00:29:16.058473 kubelet[2533]: I0909 00:29:16.058453 2533 policy_none.go:49] "None policy: Start" Sep 9 00:29:16.058473 kubelet[2533]: I0909 00:29:16.058466 2533 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 9 00:29:16.058580 kubelet[2533]: I0909 00:29:16.058479 2533 state_mem.go:35] "Initializing new in-memory state store" Sep 9 00:29:16.058643 kubelet[2533]: I0909 00:29:16.058625 2533 state_mem.go:75] "Updated machine memory state" Sep 9 00:29:16.063807 kubelet[2533]: I0909 00:29:16.063665 2533 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 9 00:29:16.064015 kubelet[2533]: I0909 00:29:16.063904 2533 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 9 00:29:16.064015 kubelet[2533]: I0909 00:29:16.063918 2533 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 9 00:29:16.064426 kubelet[2533]: I0909 00:29:16.064205 2533 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 9 00:29:16.065548 kubelet[2533]: E0909 00:29:16.065515 2533 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 9 00:29:16.125173 kubelet[2533]: I0909 00:29:16.125009 2533 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:16.125173 kubelet[2533]: I0909 00:29:16.125154 2533 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:16.125411 kubelet[2533]: I0909 00:29:16.125293 2533 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:16.136486 kubelet[2533]: E0909 00:29:16.136293 2533 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:16.169042 kubelet[2533]: I0909 00:29:16.168987 2533 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Sep 9 00:29:16.178689 kubelet[2533]: I0909 00:29:16.178651 2533 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Sep 9 00:29:16.178844 kubelet[2533]: I0909 00:29:16.178772 2533 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Sep 9 00:29:16.183201 sudo[2572]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 9 00:29:16.183699 sudo[2572]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 9 00:29:16.192177 kubelet[2533]: I0909 00:29:16.192119 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:16.192177 kubelet[2533]: I0909 00:29:16.192166 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a9176403b596d0b29ae8ad12d635226d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a9176403b596d0b29ae8ad12d635226d\") " pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:16.192177 kubelet[2533]: I0909 00:29:16.192185 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8caf8dcf530eddbba5e07f694f7fc047-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"8caf8dcf530eddbba5e07f694f7fc047\") " pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:16.192421 kubelet[2533]: I0909 00:29:16.192199 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8caf8dcf530eddbba5e07f694f7fc047-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"8caf8dcf530eddbba5e07f694f7fc047\") " pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:16.192421 kubelet[2533]: I0909 00:29:16.192224 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:16.192421 kubelet[2533]: I0909 00:29:16.192237 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:16.192421 kubelet[2533]: I0909 00:29:16.192254 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:16.192421 kubelet[2533]: I0909 00:29:16.192268 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8caf8dcf530eddbba5e07f694f7fc047-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"8caf8dcf530eddbba5e07f694f7fc047\") " pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:16.192652 kubelet[2533]: I0909 00:29:16.192283 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a88c9297c136b0f15880bf567e89a977-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"a88c9297c136b0f15880bf567e89a977\") " pod="kube-system/kube-controller-manager-localhost" Sep 9 00:29:16.434772 kubelet[2533]: E0909 00:29:16.434344 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:16.435993 kubelet[2533]: E0909 00:29:16.435908 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:16.438125 kubelet[2533]: E0909 00:29:16.437463 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:16.835518 sudo[2572]: pam_unix(sudo:session): session closed for user root Sep 9 00:29:16.977359 kubelet[2533]: I0909 00:29:16.977303 2533 apiserver.go:52] "Watching apiserver" Sep 9 00:29:16.992864 kubelet[2533]: I0909 00:29:16.992785 2533 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 9 00:29:17.043215 kubelet[2533]: E0909 00:29:17.043164 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:17.043520 kubelet[2533]: I0909 00:29:17.043452 2533 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:17.044062 kubelet[2533]: I0909 00:29:17.043991 2533 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:17.199239 kubelet[2533]: E0909 00:29:17.198496 2533 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 9 00:29:17.199239 kubelet[2533]: E0909 00:29:17.198638 2533 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 9 00:29:17.199239 kubelet[2533]: E0909 00:29:17.198857 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:17.199239 kubelet[2533]: E0909 00:29:17.198867 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:17.259730 kubelet[2533]: I0909 00:29:17.259572 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.2595367149999999 podStartE2EDuration="1.259536715s" podCreationTimestamp="2025-09-09 00:29:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:29:17.258029972 +0000 UTC m=+1.357784394" watchObservedRunningTime="2025-09-09 00:29:17.259536715 +0000 UTC m=+1.359291137" Sep 9 00:29:17.260001 kubelet[2533]: I0909 00:29:17.259797 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.259782977 podStartE2EDuration="1.259782977s" podCreationTimestamp="2025-09-09 00:29:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:29:17.241753463 +0000 UTC m=+1.341507885" watchObservedRunningTime="2025-09-09 00:29:17.259782977 +0000 UTC m=+1.359537399" Sep 9 00:29:17.273449 kubelet[2533]: I0909 00:29:17.273351 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.273315153 podStartE2EDuration="3.273315153s" podCreationTimestamp="2025-09-09 00:29:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:29:17.27327161 +0000 UTC m=+1.373026032" watchObservedRunningTime="2025-09-09 00:29:17.273315153 +0000 UTC m=+1.373069575" Sep 9 00:29:18.045739 kubelet[2533]: E0909 00:29:18.045438 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:18.045739 kubelet[2533]: E0909 00:29:18.045533 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:19.568893 sudo[1643]: pam_unix(sudo:session): session closed for user root Sep 9 00:29:19.581321 sshd[1640]: pam_unix(sshd:session): session closed for user core Sep 9 00:29:19.596129 systemd[1]: sshd@7-10.0.0.113:22-10.0.0.1:51856.service: Deactivated successfully. Sep 9 00:29:19.602255 systemd[1]: session-7.scope: Deactivated successfully. Sep 9 00:29:19.602712 systemd[1]: session-7.scope: Consumed 7.188s CPU time, 160.9M memory peak, 0B memory swap peak. Sep 9 00:29:19.609214 systemd-logind[1441]: Session 7 logged out. Waiting for processes to exit. Sep 9 00:29:19.616129 systemd-logind[1441]: Removed session 7. Sep 9 00:29:21.633608 kubelet[2533]: I0909 00:29:21.633523 2533 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 9 00:29:21.635022 containerd[1457]: time="2025-09-09T00:29:21.634917370Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 9 00:29:21.635512 kubelet[2533]: I0909 00:29:21.635250 2533 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 9 00:29:21.961157 systemd[1]: Created slice kubepods-besteffort-pod2ddabe0b_a412_4a54_b30b_5f1ad0fd68a7.slice - libcontainer container kubepods-besteffort-pod2ddabe0b_a412_4a54_b30b_5f1ad0fd68a7.slice. Sep 9 00:29:21.980325 systemd[1]: Created slice kubepods-burstable-pod8dfa58a8_5c75_49b7_bb91_bfbce0f73f4a.slice - libcontainer container kubepods-burstable-pod8dfa58a8_5c75_49b7_bb91_bfbce0f73f4a.slice. Sep 9 00:29:22.136719 kubelet[2533]: I0909 00:29:22.136626 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-config-path\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.136719 kubelet[2533]: I0909 00:29:22.136694 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-bpf-maps\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.136719 kubelet[2533]: I0909 00:29:22.136721 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-net\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.136719 kubelet[2533]: I0909 00:29:22.136745 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nxvs\" (UniqueName: \"kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-kube-api-access-8nxvs\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137014 kubelet[2533]: I0909 00:29:22.136765 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-clustermesh-secrets\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137014 kubelet[2533]: I0909 00:29:22.136784 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7-kube-proxy\") pod \"kube-proxy-lkjz2\" (UID: \"2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7\") " pod="kube-system/kube-proxy-lkjz2" Sep 9 00:29:22.137014 kubelet[2533]: I0909 00:29:22.136805 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7-xtables-lock\") pod \"kube-proxy-lkjz2\" (UID: \"2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7\") " pod="kube-system/kube-proxy-lkjz2" Sep 9 00:29:22.137014 kubelet[2533]: I0909 00:29:22.136830 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7-lib-modules\") pod \"kube-proxy-lkjz2\" (UID: \"2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7\") " pod="kube-system/kube-proxy-lkjz2" Sep 9 00:29:22.137014 kubelet[2533]: I0909 00:29:22.136870 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-run\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137014 kubelet[2533]: I0909 00:29:22.136916 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-etc-cni-netd\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137248 kubelet[2533]: I0909 00:29:22.136949 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-kernel\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137248 kubelet[2533]: I0909 00:29:22.136975 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hubble-tls\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137248 kubelet[2533]: I0909 00:29:22.137003 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hostproc\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137248 kubelet[2533]: I0909 00:29:22.137022 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-cgroup\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137248 kubelet[2533]: I0909 00:29:22.137044 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cni-path\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137248 kubelet[2533]: I0909 00:29:22.137065 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-lib-modules\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.137443 kubelet[2533]: I0909 00:29:22.137115 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w292f\" (UniqueName: \"kubernetes.io/projected/2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7-kube-api-access-w292f\") pod \"kube-proxy-lkjz2\" (UID: \"2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7\") " pod="kube-system/kube-proxy-lkjz2" Sep 9 00:29:22.137443 kubelet[2533]: I0909 00:29:22.137138 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-xtables-lock\") pod \"cilium-6cdmj\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " pod="kube-system/cilium-6cdmj" Sep 9 00:29:22.253758 kubelet[2533]: E0909 00:29:22.253241 2533 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 9 00:29:22.253758 kubelet[2533]: E0909 00:29:22.253297 2533 projected.go:194] Error preparing data for projected volume kube-api-access-8nxvs for pod kube-system/cilium-6cdmj: configmap "kube-root-ca.crt" not found Sep 9 00:29:22.253758 kubelet[2533]: E0909 00:29:22.253366 2533 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-kube-api-access-8nxvs podName:8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a nodeName:}" failed. No retries permitted until 2025-09-09 00:29:22.753337114 +0000 UTC m=+6.853091536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8nxvs" (UniqueName: "kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-kube-api-access-8nxvs") pod "cilium-6cdmj" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a") : configmap "kube-root-ca.crt" not found Sep 9 00:29:22.261754 kubelet[2533]: E0909 00:29:22.261716 2533 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 9 00:29:22.266230 kubelet[2533]: E0909 00:29:22.266184 2533 projected.go:194] Error preparing data for projected volume kube-api-access-w292f for pod kube-system/kube-proxy-lkjz2: configmap "kube-root-ca.crt" not found Sep 9 00:29:22.266397 kubelet[2533]: E0909 00:29:22.266286 2533 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7-kube-api-access-w292f podName:2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7 nodeName:}" failed. No retries permitted until 2025-09-09 00:29:22.766260737 +0000 UTC m=+6.866015159 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-w292f" (UniqueName: "kubernetes.io/projected/2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7-kube-api-access-w292f") pod "kube-proxy-lkjz2" (UID: "2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7") : configmap "kube-root-ca.crt" not found Sep 9 00:29:22.560336 systemd[1]: Created slice kubepods-besteffort-pod26b20ccf_b4d8_42dc_8854_c6374a65c9b4.slice - libcontainer container kubepods-besteffort-pod26b20ccf_b4d8_42dc_8854_c6374a65c9b4.slice. Sep 9 00:29:22.641222 kubelet[2533]: I0909 00:29:22.641119 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-rb6db\" (UID: \"26b20ccf-b4d8-42dc-8854-c6374a65c9b4\") " pod="kube-system/cilium-operator-6c4d7847fc-rb6db" Sep 9 00:29:22.641222 kubelet[2533]: I0909 00:29:22.641218 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mzz8\" (UniqueName: \"kubernetes.io/projected/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-kube-api-access-9mzz8\") pod \"cilium-operator-6c4d7847fc-rb6db\" (UID: \"26b20ccf-b4d8-42dc-8854-c6374a65c9b4\") " pod="kube-system/cilium-operator-6c4d7847fc-rb6db" Sep 9 00:29:22.863530 kubelet[2533]: E0909 00:29:22.863330 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:22.864372 containerd[1457]: time="2025-09-09T00:29:22.864308452Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-rb6db,Uid:26b20ccf-b4d8-42dc-8854-c6374a65c9b4,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:22.873645 kubelet[2533]: E0909 00:29:22.873416 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:22.874249 containerd[1457]: time="2025-09-09T00:29:22.874195337Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-lkjz2,Uid:2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:22.887385 kubelet[2533]: E0909 00:29:22.887315 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:22.888144 containerd[1457]: time="2025-09-09T00:29:22.887950285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6cdmj,Uid:8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:22.982918 containerd[1457]: time="2025-09-09T00:29:22.982755964Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:22.989144 containerd[1457]: time="2025-09-09T00:29:22.986306290Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:22.989144 containerd[1457]: time="2025-09-09T00:29:22.986350345Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:22.989144 containerd[1457]: time="2025-09-09T00:29:22.986476846Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:23.000887 kubelet[2533]: E0909 00:29:23.000774 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:23.021345 containerd[1457]: time="2025-09-09T00:29:23.020807478Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:23.021345 containerd[1457]: time="2025-09-09T00:29:23.020911626Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:23.021345 containerd[1457]: time="2025-09-09T00:29:23.020939149Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:23.021345 containerd[1457]: time="2025-09-09T00:29:23.021054880Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:23.024216 containerd[1457]: time="2025-09-09T00:29:23.022326521Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:23.024216 containerd[1457]: time="2025-09-09T00:29:23.022632375Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:23.024216 containerd[1457]: time="2025-09-09T00:29:23.022653314Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:23.024216 containerd[1457]: time="2025-09-09T00:29:23.022751070Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:23.023730 systemd[1]: Started cri-containerd-e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3.scope - libcontainer container e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3. Sep 9 00:29:23.059362 systemd[1]: Started cri-containerd-bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161.scope - libcontainer container bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161. Sep 9 00:29:23.059653 kubelet[2533]: E0909 00:29:23.059411 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:23.067729 systemd[1]: Started cri-containerd-64d35b2352f0b906c668e2e0c3833fe399baf889e608f1c9921f36a861b35273.scope - libcontainer container 64d35b2352f0b906c668e2e0c3833fe399baf889e608f1c9921f36a861b35273. Sep 9 00:29:23.113394 containerd[1457]: time="2025-09-09T00:29:23.113294040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6cdmj,Uid:8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a,Namespace:kube-system,Attempt:0,} returns sandbox id \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\"" Sep 9 00:29:23.115765 containerd[1457]: time="2025-09-09T00:29:23.115148983Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-lkjz2,Uid:2ddabe0b-a412-4a54-b30b-5f1ad0fd68a7,Namespace:kube-system,Attempt:0,} returns sandbox id \"64d35b2352f0b906c668e2e0c3833fe399baf889e608f1c9921f36a861b35273\"" Sep 9 00:29:23.115830 kubelet[2533]: E0909 00:29:23.115581 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:23.115930 kubelet[2533]: E0909 00:29:23.115908 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:23.120035 containerd[1457]: time="2025-09-09T00:29:23.119978593Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-rb6db,Uid:26b20ccf-b4d8-42dc-8854-c6374a65c9b4,Namespace:kube-system,Attempt:0,} returns sandbox id \"e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3\"" Sep 9 00:29:23.121494 kubelet[2533]: E0909 00:29:23.121447 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:23.121662 containerd[1457]: time="2025-09-09T00:29:23.121457881Z" level=info msg="CreateContainer within sandbox \"64d35b2352f0b906c668e2e0c3833fe399baf889e608f1c9921f36a861b35273\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 9 00:29:23.121824 containerd[1457]: time="2025-09-09T00:29:23.121465836Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 9 00:29:23.157607 containerd[1457]: time="2025-09-09T00:29:23.157523430Z" level=info msg="CreateContainer within sandbox \"64d35b2352f0b906c668e2e0c3833fe399baf889e608f1c9921f36a861b35273\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"207d44c4ac0f855e18633ff8837f8e96786910565e6b5fda87f23f5cd365be66\"" Sep 9 00:29:23.160115 containerd[1457]: time="2025-09-09T00:29:23.158508276Z" level=info msg="StartContainer for \"207d44c4ac0f855e18633ff8837f8e96786910565e6b5fda87f23f5cd365be66\"" Sep 9 00:29:23.195517 systemd[1]: Started cri-containerd-207d44c4ac0f855e18633ff8837f8e96786910565e6b5fda87f23f5cd365be66.scope - libcontainer container 207d44c4ac0f855e18633ff8837f8e96786910565e6b5fda87f23f5cd365be66. Sep 9 00:29:23.234953 containerd[1457]: time="2025-09-09T00:29:23.234886705Z" level=info msg="StartContainer for \"207d44c4ac0f855e18633ff8837f8e96786910565e6b5fda87f23f5cd365be66\" returns successfully" Sep 9 00:29:24.064288 kubelet[2533]: E0909 00:29:24.064245 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:24.074609 kubelet[2533]: I0909 00:29:24.074415 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-lkjz2" podStartSLOduration=3.074378486 podStartE2EDuration="3.074378486s" podCreationTimestamp="2025-09-09 00:29:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:29:24.074234993 +0000 UTC m=+8.173989415" watchObservedRunningTime="2025-09-09 00:29:24.074378486 +0000 UTC m=+8.174132908" Sep 9 00:29:24.104475 kubelet[2533]: E0909 00:29:24.104343 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:25.068324 kubelet[2533]: E0909 00:29:25.068262 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:25.620751 kubelet[2533]: E0909 00:29:25.620690 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:26.069268 kubelet[2533]: E0909 00:29:26.069221 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:27.072826 kubelet[2533]: E0909 00:29:27.072765 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:38.016277 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2573885575.mount: Deactivated successfully. Sep 9 00:29:45.691593 containerd[1457]: time="2025-09-09T00:29:45.691484410Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:45.766100 containerd[1457]: time="2025-09-09T00:29:45.765984079Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Sep 9 00:29:45.796447 containerd[1457]: time="2025-09-09T00:29:45.796392659Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:45.798629 containerd[1457]: time="2025-09-09T00:29:45.798584863Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 22.676787406s" Sep 9 00:29:45.798629 containerd[1457]: time="2025-09-09T00:29:45.798621282Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 9 00:29:45.806201 containerd[1457]: time="2025-09-09T00:29:45.806166777Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 9 00:29:45.829000 containerd[1457]: time="2025-09-09T00:29:45.828933781Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 9 00:29:46.989491 containerd[1457]: time="2025-09-09T00:29:46.989360155Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\"" Sep 9 00:29:46.990175 containerd[1457]: time="2025-09-09T00:29:46.990138517Z" level=info msg="StartContainer for \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\"" Sep 9 00:29:47.031380 systemd[1]: Started cri-containerd-e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f.scope - libcontainer container e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f. Sep 9 00:29:47.085201 systemd[1]: cri-containerd-e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f.scope: Deactivated successfully. Sep 9 00:29:47.127234 containerd[1457]: time="2025-09-09T00:29:47.127158784Z" level=info msg="StartContainer for \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\" returns successfully" Sep 9 00:29:47.788778 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f-rootfs.mount: Deactivated successfully. Sep 9 00:29:47.874734 kubelet[2533]: E0909 00:29:47.874691 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:47.893466 containerd[1457]: time="2025-09-09T00:29:47.893166723Z" level=info msg="shim disconnected" id=e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f namespace=k8s.io Sep 9 00:29:47.893466 containerd[1457]: time="2025-09-09T00:29:47.893247215Z" level=warning msg="cleaning up after shim disconnected" id=e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f namespace=k8s.io Sep 9 00:29:47.893466 containerd[1457]: time="2025-09-09T00:29:47.893260801Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:29:48.877937 kubelet[2533]: E0909 00:29:48.877886 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:48.880228 containerd[1457]: time="2025-09-09T00:29:48.880102637Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 9 00:29:48.992859 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2804062266.mount: Deactivated successfully. Sep 9 00:29:49.416343 containerd[1457]: time="2025-09-09T00:29:49.416125563Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\"" Sep 9 00:29:49.417121 containerd[1457]: time="2025-09-09T00:29:49.417071491Z" level=info msg="StartContainer for \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\"" Sep 9 00:29:49.452329 systemd[1]: Started cri-containerd-e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62.scope - libcontainer container e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62. Sep 9 00:29:49.536804 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 9 00:29:49.537088 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 9 00:29:49.537171 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 9 00:29:49.547576 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 9 00:29:49.547970 systemd[1]: cri-containerd-e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62.scope: Deactivated successfully. Sep 9 00:29:49.567126 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 9 00:29:49.625115 containerd[1457]: time="2025-09-09T00:29:49.624994013Z" level=info msg="StartContainer for \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\" returns successfully" Sep 9 00:29:49.685005 containerd[1457]: time="2025-09-09T00:29:49.684807264Z" level=info msg="shim disconnected" id=e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62 namespace=k8s.io Sep 9 00:29:49.685005 containerd[1457]: time="2025-09-09T00:29:49.684888377Z" level=warning msg="cleaning up after shim disconnected" id=e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62 namespace=k8s.io Sep 9 00:29:49.685005 containerd[1457]: time="2025-09-09T00:29:49.684905740Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:29:49.884816 kubelet[2533]: E0909 00:29:49.884769 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:49.887057 containerd[1457]: time="2025-09-09T00:29:49.887022401Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 9 00:29:49.916273 containerd[1457]: time="2025-09-09T00:29:49.916188282Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\"" Sep 9 00:29:49.916903 containerd[1457]: time="2025-09-09T00:29:49.916874970Z" level=info msg="StartContainer for \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\"" Sep 9 00:29:49.948331 systemd[1]: Started cri-containerd-110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364.scope - libcontainer container 110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364. Sep 9 00:29:49.983226 systemd[1]: cri-containerd-110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364.scope: Deactivated successfully. Sep 9 00:29:49.991370 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62-rootfs.mount: Deactivated successfully. Sep 9 00:29:50.048063 containerd[1457]: time="2025-09-09T00:29:50.047974199Z" level=info msg="StartContainer for \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\" returns successfully" Sep 9 00:29:50.069135 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364-rootfs.mount: Deactivated successfully. Sep 9 00:29:50.492099 containerd[1457]: time="2025-09-09T00:29:50.491992501Z" level=info msg="shim disconnected" id=110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364 namespace=k8s.io Sep 9 00:29:50.492099 containerd[1457]: time="2025-09-09T00:29:50.492068154Z" level=warning msg="cleaning up after shim disconnected" id=110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364 namespace=k8s.io Sep 9 00:29:50.492099 containerd[1457]: time="2025-09-09T00:29:50.492096949Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:29:50.497732 containerd[1457]: time="2025-09-09T00:29:50.497684065Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:50.498992 containerd[1457]: time="2025-09-09T00:29:50.498907406Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Sep 9 00:29:50.500999 containerd[1457]: time="2025-09-09T00:29:50.500651061Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 9 00:29:50.502677 containerd[1457]: time="2025-09-09T00:29:50.502645268Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 4.696443354s" Sep 9 00:29:50.503138 containerd[1457]: time="2025-09-09T00:29:50.503111148Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 9 00:29:50.506120 containerd[1457]: time="2025-09-09T00:29:50.506026957Z" level=info msg="CreateContainer within sandbox \"e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 9 00:29:50.523027 containerd[1457]: time="2025-09-09T00:29:50.522982332Z" level=info msg="CreateContainer within sandbox \"e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\"" Sep 9 00:29:50.523633 containerd[1457]: time="2025-09-09T00:29:50.523604207Z" level=info msg="StartContainer for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\"" Sep 9 00:29:50.555302 systemd[1]: Started cri-containerd-347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2.scope - libcontainer container 347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2. Sep 9 00:29:50.586827 containerd[1457]: time="2025-09-09T00:29:50.586765740Z" level=info msg="StartContainer for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" returns successfully" Sep 9 00:29:50.890015 kubelet[2533]: E0909 00:29:50.889807 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:50.892563 kubelet[2533]: E0909 00:29:50.892537 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:50.892808 containerd[1457]: time="2025-09-09T00:29:50.892758496Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 9 00:29:50.984415 containerd[1457]: time="2025-09-09T00:29:50.984035980Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\"" Sep 9 00:29:50.986726 containerd[1457]: time="2025-09-09T00:29:50.986335435Z" level=info msg="StartContainer for \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\"" Sep 9 00:29:51.082394 systemd[1]: Started cri-containerd-3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea.scope - libcontainer container 3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea. Sep 9 00:29:51.116727 systemd[1]: cri-containerd-3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea.scope: Deactivated successfully. Sep 9 00:29:51.140219 containerd[1457]: time="2025-09-09T00:29:51.139986716Z" level=info msg="StartContainer for \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\" returns successfully" Sep 9 00:29:51.168527 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea-rootfs.mount: Deactivated successfully. Sep 9 00:29:51.194896 containerd[1457]: time="2025-09-09T00:29:51.194791960Z" level=info msg="shim disconnected" id=3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea namespace=k8s.io Sep 9 00:29:51.194896 containerd[1457]: time="2025-09-09T00:29:51.194874075Z" level=warning msg="cleaning up after shim disconnected" id=3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea namespace=k8s.io Sep 9 00:29:51.194896 containerd[1457]: time="2025-09-09T00:29:51.194886328Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:29:51.898034 kubelet[2533]: E0909 00:29:51.897789 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:51.898034 kubelet[2533]: E0909 00:29:51.897789 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:51.900796 containerd[1457]: time="2025-09-09T00:29:51.900630314Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 9 00:29:51.919187 kubelet[2533]: I0909 00:29:51.919094 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-rb6db" podStartSLOduration=2.536835232 podStartE2EDuration="29.9190385s" podCreationTimestamp="2025-09-09 00:29:22 +0000 UTC" firstStartedPulling="2025-09-09 00:29:23.122068043 +0000 UTC m=+7.221822465" lastFinishedPulling="2025-09-09 00:29:50.504271301 +0000 UTC m=+34.604025733" observedRunningTime="2025-09-09 00:29:51.01740918 +0000 UTC m=+35.117163602" watchObservedRunningTime="2025-09-09 00:29:51.9190385 +0000 UTC m=+36.018792922" Sep 9 00:29:52.018317 containerd[1457]: time="2025-09-09T00:29:52.018209504Z" level=info msg="CreateContainer within sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\"" Sep 9 00:29:52.019054 containerd[1457]: time="2025-09-09T00:29:52.019021128Z" level=info msg="StartContainer for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\"" Sep 9 00:29:52.065453 systemd[1]: Started cri-containerd-de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87.scope - libcontainer container de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87. Sep 9 00:29:52.132008 containerd[1457]: time="2025-09-09T00:29:52.131604892Z" level=info msg="StartContainer for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" returns successfully" Sep 9 00:29:52.259260 kubelet[2533]: I0909 00:29:52.258872 2533 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 9 00:29:52.371977 systemd[1]: Created slice kubepods-burstable-pod40b3767c_704e_4b8d_b44c_45ff00055f45.slice - libcontainer container kubepods-burstable-pod40b3767c_704e_4b8d_b44c_45ff00055f45.slice. Sep 9 00:29:52.382441 systemd[1]: Created slice kubepods-burstable-poded8eee57_e06b_44ee_a87b_c89baf525c58.slice - libcontainer container kubepods-burstable-poded8eee57_e06b_44ee_a87b_c89baf525c58.slice. Sep 9 00:29:52.433382 kubelet[2533]: I0909 00:29:52.433299 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz5jq\" (UniqueName: \"kubernetes.io/projected/40b3767c-704e-4b8d-b44c-45ff00055f45-kube-api-access-gz5jq\") pod \"coredns-668d6bf9bc-ppvzb\" (UID: \"40b3767c-704e-4b8d-b44c-45ff00055f45\") " pod="kube-system/coredns-668d6bf9bc-ppvzb" Sep 9 00:29:52.433382 kubelet[2533]: I0909 00:29:52.433380 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed8eee57-e06b-44ee-a87b-c89baf525c58-config-volume\") pod \"coredns-668d6bf9bc-hnffl\" (UID: \"ed8eee57-e06b-44ee-a87b-c89baf525c58\") " pod="kube-system/coredns-668d6bf9bc-hnffl" Sep 9 00:29:52.433382 kubelet[2533]: I0909 00:29:52.433421 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40b3767c-704e-4b8d-b44c-45ff00055f45-config-volume\") pod \"coredns-668d6bf9bc-ppvzb\" (UID: \"40b3767c-704e-4b8d-b44c-45ff00055f45\") " pod="kube-system/coredns-668d6bf9bc-ppvzb" Sep 9 00:29:52.434265 kubelet[2533]: I0909 00:29:52.433443 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp2m8\" (UniqueName: \"kubernetes.io/projected/ed8eee57-e06b-44ee-a87b-c89baf525c58-kube-api-access-xp2m8\") pod \"coredns-668d6bf9bc-hnffl\" (UID: \"ed8eee57-e06b-44ee-a87b-c89baf525c58\") " pod="kube-system/coredns-668d6bf9bc-hnffl" Sep 9 00:29:52.678703 kubelet[2533]: E0909 00:29:52.678622 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:52.688381 kubelet[2533]: E0909 00:29:52.688338 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:52.695277 containerd[1457]: time="2025-09-09T00:29:52.695214518Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hnffl,Uid:ed8eee57-e06b-44ee-a87b-c89baf525c58,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:52.696195 containerd[1457]: time="2025-09-09T00:29:52.696139185Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-ppvzb,Uid:40b3767c-704e-4b8d-b44c-45ff00055f45,Namespace:kube-system,Attempt:0,}" Sep 9 00:29:52.903971 kubelet[2533]: E0909 00:29:52.903930 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:53.004026 kubelet[2533]: I0909 00:29:53.003819 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-6cdmj" podStartSLOduration=9.318808148 podStartE2EDuration="32.003789328s" podCreationTimestamp="2025-09-09 00:29:21 +0000 UTC" firstStartedPulling="2025-09-09 00:29:23.120841127 +0000 UTC m=+7.220595549" lastFinishedPulling="2025-09-09 00:29:45.805822306 +0000 UTC m=+29.905576729" observedRunningTime="2025-09-09 00:29:53.003204103 +0000 UTC m=+37.102958555" watchObservedRunningTime="2025-09-09 00:29:53.003789328 +0000 UTC m=+37.103543750" Sep 9 00:29:53.905654 kubelet[2533]: E0909 00:29:53.905606 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:54.442836 systemd[1]: Started sshd@8-10.0.0.113:22-10.0.0.1:54890.service - OpenSSH per-connection server daemon (10.0.0.1:54890). Sep 9 00:29:54.489749 sshd[3372]: Accepted publickey for core from 10.0.0.1 port 54890 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:29:54.491933 sshd[3372]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:29:54.497434 systemd-logind[1441]: New session 8 of user core. Sep 9 00:29:54.505359 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 9 00:29:54.561431 systemd-networkd[1384]: cilium_host: Link UP Sep 9 00:29:54.561705 systemd-networkd[1384]: cilium_net: Link UP Sep 9 00:29:54.561711 systemd-networkd[1384]: cilium_net: Gained carrier Sep 9 00:29:54.562038 systemd-networkd[1384]: cilium_host: Gained carrier Sep 9 00:29:54.563413 systemd-networkd[1384]: cilium_host: Gained IPv6LL Sep 9 00:29:54.667132 sshd[3372]: pam_unix(sshd:session): session closed for user core Sep 9 00:29:54.672029 systemd[1]: sshd@8-10.0.0.113:22-10.0.0.1:54890.service: Deactivated successfully. Sep 9 00:29:54.674634 systemd[1]: session-8.scope: Deactivated successfully. Sep 9 00:29:54.675279 systemd-logind[1441]: Session 8 logged out. Waiting for processes to exit. Sep 9 00:29:54.676454 systemd-logind[1441]: Removed session 8. Sep 9 00:29:54.697738 systemd-networkd[1384]: cilium_vxlan: Link UP Sep 9 00:29:54.697749 systemd-networkd[1384]: cilium_vxlan: Gained carrier Sep 9 00:29:54.908443 kubelet[2533]: E0909 00:29:54.908404 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:54.951116 kernel: NET: Registered PF_ALG protocol family Sep 9 00:29:55.416194 systemd-networkd[1384]: cilium_net: Gained IPv6LL Sep 9 00:29:55.674776 systemd-networkd[1384]: lxc_health: Link UP Sep 9 00:29:55.684218 systemd-networkd[1384]: lxc_health: Gained carrier Sep 9 00:29:55.864532 systemd-networkd[1384]: cilium_vxlan: Gained IPv6LL Sep 9 00:29:56.214766 systemd-networkd[1384]: lxcff8134d0ee7f: Link UP Sep 9 00:29:56.225315 kernel: eth0: renamed from tmpc9f0d Sep 9 00:29:56.238257 kernel: eth0: renamed from tmpd183d Sep 9 00:29:56.242515 systemd-networkd[1384]: tmpd183d: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 9 00:29:56.242634 systemd-networkd[1384]: tmpd183d: Cannot enable IPv6, ignoring: No such file or directory Sep 9 00:29:56.242684 systemd-networkd[1384]: tmpd183d: Cannot configure IPv6 privacy extensions for interface, ignoring: No such file or directory Sep 9 00:29:56.242705 systemd-networkd[1384]: tmpd183d: Cannot disable kernel IPv6 accept_ra for interface, ignoring: No such file or directory Sep 9 00:29:56.242729 systemd-networkd[1384]: tmpd183d: Cannot set IPv6 proxy NDP, ignoring: No such file or directory Sep 9 00:29:56.242758 systemd-networkd[1384]: tmpd183d: Cannot enable promote_secondaries for interface, ignoring: No such file or directory Sep 9 00:29:56.243191 systemd-networkd[1384]: lxcff8134d0ee7f: Gained carrier Sep 9 00:29:56.243391 systemd-networkd[1384]: lxcd57196de31cc: Link UP Sep 9 00:29:56.245850 systemd-networkd[1384]: lxcd57196de31cc: Gained carrier Sep 9 00:29:56.889616 kubelet[2533]: E0909 00:29:56.889471 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:56.912108 kubelet[2533]: E0909 00:29:56.912053 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:57.016476 systemd-networkd[1384]: lxc_health: Gained IPv6LL Sep 9 00:29:57.336328 systemd-networkd[1384]: lxcff8134d0ee7f: Gained IPv6LL Sep 9 00:29:57.656372 systemd-networkd[1384]: lxcd57196de31cc: Gained IPv6LL Sep 9 00:29:57.914574 kubelet[2533]: E0909 00:29:57.914428 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:29:59.680966 systemd[1]: Started sshd@9-10.0.0.113:22-10.0.0.1:54898.service - OpenSSH per-connection server daemon (10.0.0.1:54898). Sep 9 00:29:59.731873 sshd[3773]: Accepted publickey for core from 10.0.0.1 port 54898 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:29:59.733947 sshd[3773]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:29:59.740029 systemd-logind[1441]: New session 9 of user core. Sep 9 00:29:59.754594 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 9 00:29:59.924195 containerd[1457]: time="2025-09-09T00:29:59.922700827Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:59.924195 containerd[1457]: time="2025-09-09T00:29:59.922874505Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:59.924195 containerd[1457]: time="2025-09-09T00:29:59.922889995Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:59.924195 containerd[1457]: time="2025-09-09T00:29:59.923004621Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:59.937336 containerd[1457]: time="2025-09-09T00:29:59.936939368Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:29:59.937336 containerd[1457]: time="2025-09-09T00:29:59.937008268Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:29:59.937336 containerd[1457]: time="2025-09-09T00:29:59.937032985Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:59.937336 containerd[1457]: time="2025-09-09T00:29:59.937145357Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:29:59.945427 systemd[1]: run-containerd-runc-k8s.io-c9f0d446ea3e6bdefa6333936616b9e33142896f7910dd7bcba6ab3b7b2c6ebf-runc.y0nFXK.mount: Deactivated successfully. Sep 9 00:29:59.954341 systemd[1]: Started cri-containerd-c9f0d446ea3e6bdefa6333936616b9e33142896f7910dd7bcba6ab3b7b2c6ebf.scope - libcontainer container c9f0d446ea3e6bdefa6333936616b9e33142896f7910dd7bcba6ab3b7b2c6ebf. Sep 9 00:29:59.969251 systemd[1]: Started cri-containerd-d183dbd92f795d786e5e7f65969c6dfa48845273b1b76a00ccb1e53d0e3e7eaf.scope - libcontainer container d183dbd92f795d786e5e7f65969c6dfa48845273b1b76a00ccb1e53d0e3e7eaf. Sep 9 00:29:59.974692 systemd-resolved[1329]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 00:29:59.988333 sshd[3773]: pam_unix(sshd:session): session closed for user core Sep 9 00:29:59.991678 systemd-resolved[1329]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 9 00:29:59.996524 systemd[1]: sshd@9-10.0.0.113:22-10.0.0.1:54898.service: Deactivated successfully. Sep 9 00:29:59.999910 systemd[1]: session-9.scope: Deactivated successfully. Sep 9 00:30:00.001399 systemd-logind[1441]: Session 9 logged out. Waiting for processes to exit. Sep 9 00:30:00.002696 systemd-logind[1441]: Removed session 9. Sep 9 00:30:00.017993 containerd[1457]: time="2025-09-09T00:30:00.017948283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-ppvzb,Uid:40b3767c-704e-4b8d-b44c-45ff00055f45,Namespace:kube-system,Attempt:0,} returns sandbox id \"c9f0d446ea3e6bdefa6333936616b9e33142896f7910dd7bcba6ab3b7b2c6ebf\"" Sep 9 00:30:00.021109 kubelet[2533]: E0909 00:30:00.020599 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:00.030040 containerd[1457]: time="2025-09-09T00:30:00.029977431Z" level=info msg="CreateContainer within sandbox \"c9f0d446ea3e6bdefa6333936616b9e33142896f7910dd7bcba6ab3b7b2c6ebf\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 9 00:30:00.032382 containerd[1457]: time="2025-09-09T00:30:00.032323290Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hnffl,Uid:ed8eee57-e06b-44ee-a87b-c89baf525c58,Namespace:kube-system,Attempt:0,} returns sandbox id \"d183dbd92f795d786e5e7f65969c6dfa48845273b1b76a00ccb1e53d0e3e7eaf\"" Sep 9 00:30:00.033351 kubelet[2533]: E0909 00:30:00.033304 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:00.035178 containerd[1457]: time="2025-09-09T00:30:00.035132503Z" level=info msg="CreateContainer within sandbox \"d183dbd92f795d786e5e7f65969c6dfa48845273b1b76a00ccb1e53d0e3e7eaf\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 9 00:30:00.078205 containerd[1457]: time="2025-09-09T00:30:00.078127668Z" level=info msg="CreateContainer within sandbox \"d183dbd92f795d786e5e7f65969c6dfa48845273b1b76a00ccb1e53d0e3e7eaf\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"15a5f24f4a1d21dc8e8cb4ed377bc06fba2ae7f33ac2039611ed6097b258e7e4\"" Sep 9 00:30:00.078845 containerd[1457]: time="2025-09-09T00:30:00.078800829Z" level=info msg="StartContainer for \"15a5f24f4a1d21dc8e8cb4ed377bc06fba2ae7f33ac2039611ed6097b258e7e4\"" Sep 9 00:30:00.079139 containerd[1457]: time="2025-09-09T00:30:00.079054568Z" level=info msg="CreateContainer within sandbox \"c9f0d446ea3e6bdefa6333936616b9e33142896f7910dd7bcba6ab3b7b2c6ebf\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"331193e9aaee0937333e5da2d9daae3043df7f769d27dd7b42f7e99d7f99551b\"" Sep 9 00:30:00.081033 containerd[1457]: time="2025-09-09T00:30:00.080993538Z" level=info msg="StartContainer for \"331193e9aaee0937333e5da2d9daae3043df7f769d27dd7b42f7e99d7f99551b\"" Sep 9 00:30:00.114251 systemd[1]: Started cri-containerd-15a5f24f4a1d21dc8e8cb4ed377bc06fba2ae7f33ac2039611ed6097b258e7e4.scope - libcontainer container 15a5f24f4a1d21dc8e8cb4ed377bc06fba2ae7f33ac2039611ed6097b258e7e4. Sep 9 00:30:00.115586 systemd[1]: Started cri-containerd-331193e9aaee0937333e5da2d9daae3043df7f769d27dd7b42f7e99d7f99551b.scope - libcontainer container 331193e9aaee0937333e5da2d9daae3043df7f769d27dd7b42f7e99d7f99551b. Sep 9 00:30:00.176294 containerd[1457]: time="2025-09-09T00:30:00.176007233Z" level=info msg="StartContainer for \"331193e9aaee0937333e5da2d9daae3043df7f769d27dd7b42f7e99d7f99551b\" returns successfully" Sep 9 00:30:00.176294 containerd[1457]: time="2025-09-09T00:30:00.176064581Z" level=info msg="StartContainer for \"15a5f24f4a1d21dc8e8cb4ed377bc06fba2ae7f33ac2039611ed6097b258e7e4\" returns successfully" Sep 9 00:30:00.924666 kubelet[2533]: E0909 00:30:00.924601 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:00.928181 kubelet[2533]: E0909 00:30:00.928093 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:00.948202 kubelet[2533]: I0909 00:30:00.948117 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-ppvzb" podStartSLOduration=38.94808829 podStartE2EDuration="38.94808829s" podCreationTimestamp="2025-09-09 00:29:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:30:00.93754145 +0000 UTC m=+45.037295882" watchObservedRunningTime="2025-09-09 00:30:00.94808829 +0000 UTC m=+45.047842732" Sep 9 00:30:00.965859 kubelet[2533]: I0909 00:30:00.965594 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-hnffl" podStartSLOduration=38.965570923 podStartE2EDuration="38.965570923s" podCreationTimestamp="2025-09-09 00:29:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:30:00.964152025 +0000 UTC m=+45.063906467" watchObservedRunningTime="2025-09-09 00:30:00.965570923 +0000 UTC m=+45.065325345" Sep 9 00:30:01.929737 kubelet[2533]: E0909 00:30:01.929697 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:01.930218 kubelet[2533]: E0909 00:30:01.929871 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:02.931611 kubelet[2533]: E0909 00:30:02.931568 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:02.932215 kubelet[2533]: E0909 00:30:02.931696 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:05.003747 systemd[1]: Started sshd@10-10.0.0.113:22-10.0.0.1:40728.service - OpenSSH per-connection server daemon (10.0.0.1:40728). Sep 9 00:30:05.050648 sshd[3957]: Accepted publickey for core from 10.0.0.1 port 40728 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:05.054040 sshd[3957]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:05.061022 systemd-logind[1441]: New session 10 of user core. Sep 9 00:30:05.068327 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 9 00:30:05.279640 sshd[3957]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:05.285138 systemd[1]: sshd@10-10.0.0.113:22-10.0.0.1:40728.service: Deactivated successfully. Sep 9 00:30:05.288453 systemd[1]: session-10.scope: Deactivated successfully. Sep 9 00:30:05.289543 systemd-logind[1441]: Session 10 logged out. Waiting for processes to exit. Sep 9 00:30:05.290798 systemd-logind[1441]: Removed session 10. Sep 9 00:30:10.265875 systemd[1]: Started sshd@11-10.0.0.113:22-10.0.0.1:58880.service - OpenSSH per-connection server daemon (10.0.0.1:58880). Sep 9 00:30:10.493311 sshd[3974]: Accepted publickey for core from 10.0.0.1 port 58880 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:10.495643 sshd[3974]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:10.500597 systemd-logind[1441]: New session 11 of user core. Sep 9 00:30:10.512250 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 9 00:30:10.726225 sshd[3974]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:10.742684 systemd[1]: sshd@11-10.0.0.113:22-10.0.0.1:58880.service: Deactivated successfully. Sep 9 00:30:10.745330 systemd[1]: session-11.scope: Deactivated successfully. Sep 9 00:30:10.747376 systemd-logind[1441]: Session 11 logged out. Waiting for processes to exit. Sep 9 00:30:10.748858 systemd[1]: Started sshd@12-10.0.0.113:22-10.0.0.1:58890.service - OpenSSH per-connection server daemon (10.0.0.1:58890). Sep 9 00:30:10.749751 systemd-logind[1441]: Removed session 11. Sep 9 00:30:10.794608 sshd[3989]: Accepted publickey for core from 10.0.0.1 port 58890 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:10.796907 sshd[3989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:10.801601 systemd-logind[1441]: New session 12 of user core. Sep 9 00:30:10.811265 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 9 00:30:11.165269 sshd[3989]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:11.177859 systemd[1]: sshd@12-10.0.0.113:22-10.0.0.1:58890.service: Deactivated successfully. Sep 9 00:30:11.183539 systemd[1]: session-12.scope: Deactivated successfully. Sep 9 00:30:11.187005 systemd-logind[1441]: Session 12 logged out. Waiting for processes to exit. Sep 9 00:30:11.194682 systemd[1]: Started sshd@13-10.0.0.113:22-10.0.0.1:58906.service - OpenSSH per-connection server daemon (10.0.0.1:58906). Sep 9 00:30:11.195803 systemd-logind[1441]: Removed session 12. Sep 9 00:30:11.237324 sshd[4002]: Accepted publickey for core from 10.0.0.1 port 58906 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:11.240138 sshd[4002]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:11.245843 systemd-logind[1441]: New session 13 of user core. Sep 9 00:30:11.253353 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 9 00:30:11.386974 sshd[4002]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:11.391741 systemd[1]: sshd@13-10.0.0.113:22-10.0.0.1:58906.service: Deactivated successfully. Sep 9 00:30:11.394329 systemd[1]: session-13.scope: Deactivated successfully. Sep 9 00:30:11.395203 systemd-logind[1441]: Session 13 logged out. Waiting for processes to exit. Sep 9 00:30:11.396500 systemd-logind[1441]: Removed session 13. Sep 9 00:30:16.400859 systemd[1]: Started sshd@14-10.0.0.113:22-10.0.0.1:58910.service - OpenSSH per-connection server daemon (10.0.0.1:58910). Sep 9 00:30:16.442125 sshd[4019]: Accepted publickey for core from 10.0.0.1 port 58910 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:16.444664 sshd[4019]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:16.450554 systemd-logind[1441]: New session 14 of user core. Sep 9 00:30:16.463477 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 9 00:30:16.607213 sshd[4019]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:16.611951 systemd[1]: sshd@14-10.0.0.113:22-10.0.0.1:58910.service: Deactivated successfully. Sep 9 00:30:16.614447 systemd[1]: session-14.scope: Deactivated successfully. Sep 9 00:30:16.615625 systemd-logind[1441]: Session 14 logged out. Waiting for processes to exit. Sep 9 00:30:16.617069 systemd-logind[1441]: Removed session 14. Sep 9 00:30:21.618656 systemd[1]: Started sshd@15-10.0.0.113:22-10.0.0.1:41130.service - OpenSSH per-connection server daemon (10.0.0.1:41130). Sep 9 00:30:21.744488 sshd[4033]: Accepted publickey for core from 10.0.0.1 port 41130 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:21.746590 sshd[4033]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:21.751360 systemd-logind[1441]: New session 15 of user core. Sep 9 00:30:21.760267 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 9 00:30:21.885875 sshd[4033]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:21.889857 systemd[1]: sshd@15-10.0.0.113:22-10.0.0.1:41130.service: Deactivated successfully. Sep 9 00:30:21.892277 systemd[1]: session-15.scope: Deactivated successfully. Sep 9 00:30:21.894288 systemd-logind[1441]: Session 15 logged out. Waiting for processes to exit. Sep 9 00:30:21.895332 systemd-logind[1441]: Removed session 15. Sep 9 00:30:26.897761 systemd[1]: Started sshd@16-10.0.0.113:22-10.0.0.1:41134.service - OpenSSH per-connection server daemon (10.0.0.1:41134). Sep 9 00:30:26.936972 sshd[4049]: Accepted publickey for core from 10.0.0.1 port 41134 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:26.939025 sshd[4049]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:26.943759 systemd-logind[1441]: New session 16 of user core. Sep 9 00:30:26.957256 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 9 00:30:27.080825 sshd[4049]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:27.090145 systemd[1]: sshd@16-10.0.0.113:22-10.0.0.1:41134.service: Deactivated successfully. Sep 9 00:30:27.092778 systemd[1]: session-16.scope: Deactivated successfully. Sep 9 00:30:27.094475 systemd-logind[1441]: Session 16 logged out. Waiting for processes to exit. Sep 9 00:30:27.106609 systemd[1]: Started sshd@17-10.0.0.113:22-10.0.0.1:41146.service - OpenSSH per-connection server daemon (10.0.0.1:41146). Sep 9 00:30:27.108056 systemd-logind[1441]: Removed session 16. Sep 9 00:30:27.140330 sshd[4063]: Accepted publickey for core from 10.0.0.1 port 41146 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:27.142305 sshd[4063]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:27.146524 systemd-logind[1441]: New session 17 of user core. Sep 9 00:30:27.153317 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 9 00:30:28.015802 sshd[4063]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:28.027824 systemd[1]: sshd@17-10.0.0.113:22-10.0.0.1:41146.service: Deactivated successfully. Sep 9 00:30:28.030036 systemd[1]: session-17.scope: Deactivated successfully. Sep 9 00:30:28.031978 systemd-logind[1441]: Session 17 logged out. Waiting for processes to exit. Sep 9 00:30:28.046456 systemd[1]: Started sshd@18-10.0.0.113:22-10.0.0.1:41158.service - OpenSSH per-connection server daemon (10.0.0.1:41158). Sep 9 00:30:28.047505 systemd-logind[1441]: Removed session 17. Sep 9 00:30:28.081434 sshd[4076]: Accepted publickey for core from 10.0.0.1 port 41158 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:28.083203 sshd[4076]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:28.089402 systemd-logind[1441]: New session 18 of user core. Sep 9 00:30:28.099339 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 9 00:30:28.778086 sshd[4076]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:28.788728 systemd[1]: sshd@18-10.0.0.113:22-10.0.0.1:41158.service: Deactivated successfully. Sep 9 00:30:28.790909 systemd[1]: session-18.scope: Deactivated successfully. Sep 9 00:30:28.793651 systemd-logind[1441]: Session 18 logged out. Waiting for processes to exit. Sep 9 00:30:28.799578 systemd[1]: Started sshd@19-10.0.0.113:22-10.0.0.1:41174.service - OpenSSH per-connection server daemon (10.0.0.1:41174). Sep 9 00:30:28.800993 systemd-logind[1441]: Removed session 18. Sep 9 00:30:28.839264 sshd[4115]: Accepted publickey for core from 10.0.0.1 port 41174 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:28.841130 sshd[4115]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:28.845363 systemd-logind[1441]: New session 19 of user core. Sep 9 00:30:28.862234 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 9 00:30:29.328908 sshd[4115]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:29.336290 systemd[1]: sshd@19-10.0.0.113:22-10.0.0.1:41174.service: Deactivated successfully. Sep 9 00:30:29.338795 systemd[1]: session-19.scope: Deactivated successfully. Sep 9 00:30:29.340841 systemd-logind[1441]: Session 19 logged out. Waiting for processes to exit. Sep 9 00:30:29.349445 systemd[1]: Started sshd@20-10.0.0.113:22-10.0.0.1:41190.service - OpenSSH per-connection server daemon (10.0.0.1:41190). Sep 9 00:30:29.350508 systemd-logind[1441]: Removed session 19. Sep 9 00:30:29.383973 sshd[4128]: Accepted publickey for core from 10.0.0.1 port 41190 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:29.386265 sshd[4128]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:29.391303 systemd-logind[1441]: New session 20 of user core. Sep 9 00:30:29.399245 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 9 00:30:29.539669 sshd[4128]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:29.544041 systemd[1]: sshd@20-10.0.0.113:22-10.0.0.1:41190.service: Deactivated successfully. Sep 9 00:30:29.546606 systemd[1]: session-20.scope: Deactivated successfully. Sep 9 00:30:29.547244 systemd-logind[1441]: Session 20 logged out. Waiting for processes to exit. Sep 9 00:30:29.548457 systemd-logind[1441]: Removed session 20. Sep 9 00:30:30.025029 kubelet[2533]: E0909 00:30:30.024968 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:34.556367 systemd[1]: Started sshd@21-10.0.0.113:22-10.0.0.1:50932.service - OpenSSH per-connection server daemon (10.0.0.1:50932). Sep 9 00:30:34.599583 sshd[4142]: Accepted publickey for core from 10.0.0.1 port 50932 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:34.601709 sshd[4142]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:34.606216 systemd-logind[1441]: New session 21 of user core. Sep 9 00:30:34.620249 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 9 00:30:34.730823 sshd[4142]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:34.735497 systemd[1]: sshd@21-10.0.0.113:22-10.0.0.1:50932.service: Deactivated successfully. Sep 9 00:30:34.738023 systemd[1]: session-21.scope: Deactivated successfully. Sep 9 00:30:34.738742 systemd-logind[1441]: Session 21 logged out. Waiting for processes to exit. Sep 9 00:30:34.739737 systemd-logind[1441]: Removed session 21. Sep 9 00:30:36.028001 kubelet[2533]: E0909 00:30:36.027949 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:36.028001 kubelet[2533]: E0909 00:30:36.027975 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:39.744145 systemd[1]: Started sshd@22-10.0.0.113:22-10.0.0.1:50940.service - OpenSSH per-connection server daemon (10.0.0.1:50940). Sep 9 00:30:39.782339 sshd[4158]: Accepted publickey for core from 10.0.0.1 port 50940 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:39.784317 sshd[4158]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:39.789719 systemd-logind[1441]: New session 22 of user core. Sep 9 00:30:39.804257 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 9 00:30:39.925061 sshd[4158]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:39.930029 systemd[1]: sshd@22-10.0.0.113:22-10.0.0.1:50940.service: Deactivated successfully. Sep 9 00:30:39.932643 systemd[1]: session-22.scope: Deactivated successfully. Sep 9 00:30:39.933503 systemd-logind[1441]: Session 22 logged out. Waiting for processes to exit. Sep 9 00:30:39.934612 systemd-logind[1441]: Removed session 22. Sep 9 00:30:40.024336 kubelet[2533]: E0909 00:30:40.024142 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:44.937465 systemd[1]: Started sshd@23-10.0.0.113:22-10.0.0.1:44278.service - OpenSSH per-connection server daemon (10.0.0.1:44278). Sep 9 00:30:44.978649 sshd[4172]: Accepted publickey for core from 10.0.0.1 port 44278 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:44.981197 sshd[4172]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:44.986090 systemd-logind[1441]: New session 23 of user core. Sep 9 00:30:44.998242 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 9 00:30:45.107067 sshd[4172]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:45.111705 systemd[1]: sshd@23-10.0.0.113:22-10.0.0.1:44278.service: Deactivated successfully. Sep 9 00:30:45.113798 systemd[1]: session-23.scope: Deactivated successfully. Sep 9 00:30:45.114520 systemd-logind[1441]: Session 23 logged out. Waiting for processes to exit. Sep 9 00:30:45.115468 systemd-logind[1441]: Removed session 23. Sep 9 00:30:50.127143 systemd[1]: Started sshd@24-10.0.0.113:22-10.0.0.1:56982.service - OpenSSH per-connection server daemon (10.0.0.1:56982). Sep 9 00:30:50.163316 sshd[4186]: Accepted publickey for core from 10.0.0.1 port 56982 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:50.165116 sshd[4186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:50.169294 systemd-logind[1441]: New session 24 of user core. Sep 9 00:30:50.177261 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 9 00:30:50.292246 sshd[4186]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:50.304141 systemd[1]: sshd@24-10.0.0.113:22-10.0.0.1:56982.service: Deactivated successfully. Sep 9 00:30:50.306130 systemd[1]: session-24.scope: Deactivated successfully. Sep 9 00:30:50.307662 systemd-logind[1441]: Session 24 logged out. Waiting for processes to exit. Sep 9 00:30:50.312363 systemd[1]: Started sshd@25-10.0.0.113:22-10.0.0.1:56984.service - OpenSSH per-connection server daemon (10.0.0.1:56984). Sep 9 00:30:50.313763 systemd-logind[1441]: Removed session 24. Sep 9 00:30:50.349939 sshd[4200]: Accepted publickey for core from 10.0.0.1 port 56984 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:50.352139 sshd[4200]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:50.356900 systemd-logind[1441]: New session 25 of user core. Sep 9 00:30:50.367342 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 9 00:30:51.775592 containerd[1457]: time="2025-09-09T00:30:51.775521837Z" level=info msg="StopContainer for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" with timeout 30 (s)" Sep 9 00:30:51.776332 containerd[1457]: time="2025-09-09T00:30:51.776280518Z" level=info msg="Stop container \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" with signal terminated" Sep 9 00:30:51.797616 systemd[1]: cri-containerd-347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2.scope: Deactivated successfully. Sep 9 00:30:51.810522 containerd[1457]: time="2025-09-09T00:30:51.810463151Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 9 00:30:51.812992 containerd[1457]: time="2025-09-09T00:30:51.812957597Z" level=info msg="StopContainer for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" with timeout 2 (s)" Sep 9 00:30:51.813379 containerd[1457]: time="2025-09-09T00:30:51.813354059Z" level=info msg="Stop container \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" with signal terminated" Sep 9 00:30:51.815262 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2-rootfs.mount: Deactivated successfully. Sep 9 00:30:51.822154 systemd-networkd[1384]: lxc_health: Link DOWN Sep 9 00:30:51.822162 systemd-networkd[1384]: lxc_health: Lost carrier Sep 9 00:30:51.826335 containerd[1457]: time="2025-09-09T00:30:51.826255385Z" level=info msg="shim disconnected" id=347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2 namespace=k8s.io Sep 9 00:30:51.826335 containerd[1457]: time="2025-09-09T00:30:51.826326189Z" level=warning msg="cleaning up after shim disconnected" id=347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2 namespace=k8s.io Sep 9 00:30:51.826335 containerd[1457]: time="2025-09-09T00:30:51.826334805Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:51.845528 containerd[1457]: time="2025-09-09T00:30:51.845470569Z" level=info msg="StopContainer for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" returns successfully" Sep 9 00:30:51.848561 systemd[1]: cri-containerd-de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87.scope: Deactivated successfully. Sep 9 00:30:51.848912 systemd[1]: cri-containerd-de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87.scope: Consumed 7.619s CPU time. Sep 9 00:30:51.850406 containerd[1457]: time="2025-09-09T00:30:51.850334684Z" level=info msg="StopPodSandbox for \"e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3\"" Sep 9 00:30:51.850600 containerd[1457]: time="2025-09-09T00:30:51.850409536Z" level=info msg="Container to stop \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 9 00:30:51.852857 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3-shm.mount: Deactivated successfully. Sep 9 00:30:51.870867 systemd[1]: cri-containerd-e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3.scope: Deactivated successfully. Sep 9 00:30:51.879058 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87-rootfs.mount: Deactivated successfully. Sep 9 00:30:51.890559 containerd[1457]: time="2025-09-09T00:30:51.890473746Z" level=info msg="shim disconnected" id=de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87 namespace=k8s.io Sep 9 00:30:51.890559 containerd[1457]: time="2025-09-09T00:30:51.890544511Z" level=warning msg="cleaning up after shim disconnected" id=de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87 namespace=k8s.io Sep 9 00:30:51.890559 containerd[1457]: time="2025-09-09T00:30:51.890557485Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:51.900137 containerd[1457]: time="2025-09-09T00:30:51.900023839Z" level=info msg="shim disconnected" id=e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3 namespace=k8s.io Sep 9 00:30:51.900137 containerd[1457]: time="2025-09-09T00:30:51.900120452Z" level=warning msg="cleaning up after shim disconnected" id=e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3 namespace=k8s.io Sep 9 00:30:51.900137 containerd[1457]: time="2025-09-09T00:30:51.900133447Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:51.913251 containerd[1457]: time="2025-09-09T00:30:51.913147106Z" level=info msg="StopContainer for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" returns successfully" Sep 9 00:30:51.913978 containerd[1457]: time="2025-09-09T00:30:51.913926596Z" level=info msg="StopPodSandbox for \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\"" Sep 9 00:30:51.914233 containerd[1457]: time="2025-09-09T00:30:51.913990568Z" level=info msg="Container to stop \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 9 00:30:51.914233 containerd[1457]: time="2025-09-09T00:30:51.914010776Z" level=info msg="Container to stop \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 9 00:30:51.914233 containerd[1457]: time="2025-09-09T00:30:51.914024513Z" level=info msg="Container to stop \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 9 00:30:51.914233 containerd[1457]: time="2025-09-09T00:30:51.914037006Z" level=info msg="Container to stop \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 9 00:30:51.914233 containerd[1457]: time="2025-09-09T00:30:51.914049289Z" level=info msg="Container to stop \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 9 00:30:51.922755 systemd[1]: cri-containerd-bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161.scope: Deactivated successfully. Sep 9 00:30:51.941374 containerd[1457]: time="2025-09-09T00:30:51.941293367Z" level=info msg="TearDown network for sandbox \"e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3\" successfully" Sep 9 00:30:51.941374 containerd[1457]: time="2025-09-09T00:30:51.941354082Z" level=info msg="StopPodSandbox for \"e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3\" returns successfully" Sep 9 00:30:51.957580 containerd[1457]: time="2025-09-09T00:30:51.957500971Z" level=info msg="shim disconnected" id=bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161 namespace=k8s.io Sep 9 00:30:51.957580 containerd[1457]: time="2025-09-09T00:30:51.957570773Z" level=warning msg="cleaning up after shim disconnected" id=bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161 namespace=k8s.io Sep 9 00:30:51.957580 containerd[1457]: time="2025-09-09T00:30:51.957582015Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:51.979149 containerd[1457]: time="2025-09-09T00:30:51.979062279Z" level=info msg="TearDown network for sandbox \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" successfully" Sep 9 00:30:51.979149 containerd[1457]: time="2025-09-09T00:30:51.979133714Z" level=info msg="StopPodSandbox for \"bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161\" returns successfully" Sep 9 00:30:51.989347 kubelet[2533]: I0909 00:30:51.989299 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-run\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.989347 kubelet[2533]: I0909 00:30:51.989340 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hostproc\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.989989 kubelet[2533]: I0909 00:30:51.989375 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-config-path\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.989989 kubelet[2533]: I0909 00:30:51.989399 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-cgroup\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.989989 kubelet[2533]: I0909 00:30:51.989417 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-kernel\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.989989 kubelet[2533]: I0909 00:30:51.989439 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-cilium-config-path\") pod \"26b20ccf-b4d8-42dc-8854-c6374a65c9b4\" (UID: \"26b20ccf-b4d8-42dc-8854-c6374a65c9b4\") " Sep 9 00:30:51.989989 kubelet[2533]: I0909 00:30:51.989432 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.989989 kubelet[2533]: I0909 00:30:51.989457 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-bpf-maps\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990260 kubelet[2533]: I0909 00:30:51.989476 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cni-path\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990260 kubelet[2533]: I0909 00:30:51.989495 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-lib-modules\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990260 kubelet[2533]: I0909 00:30:51.989497 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.990260 kubelet[2533]: I0909 00:30:51.989521 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hostproc" (OuterVolumeSpecName: "hostproc") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.990260 kubelet[2533]: I0909 00:30:51.989520 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nxvs\" (UniqueName: \"kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-kube-api-access-8nxvs\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990260 kubelet[2533]: I0909 00:30:51.989579 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-etc-cni-netd\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989602 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-xtables-lock\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989623 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-net\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989656 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-clustermesh-secrets\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989677 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hubble-tls\") pod \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\" (UID: \"8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a\") " Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989701 2533 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mzz8\" (UniqueName: \"kubernetes.io/projected/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-kube-api-access-9mzz8\") pod \"26b20ccf-b4d8-42dc-8854-c6374a65c9b4\" (UID: \"26b20ccf-b4d8-42dc-8854-c6374a65c9b4\") " Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989759 2533 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:51.990469 kubelet[2533]: I0909 00:30:51.989773 2533 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:51.990693 kubelet[2533]: I0909 00:30:51.989786 2533 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:51.990693 kubelet[2533]: I0909 00:30:51.990058 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.990693 kubelet[2533]: I0909 00:30:51.990208 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.992120 kubelet[2533]: I0909 00:30:51.990849 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cni-path" (OuterVolumeSpecName: "cni-path") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.992120 kubelet[2533]: I0909 00:30:51.990894 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.992942 kubelet[2533]: I0909 00:30:51.992729 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.993312 kubelet[2533]: I0909 00:30:51.992879 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.993525 kubelet[2533]: I0909 00:30:51.993048 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 9 00:30:51.994286 kubelet[2533]: I0909 00:30:51.994246 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-kube-api-access-8nxvs" (OuterVolumeSpecName: "kube-api-access-8nxvs") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "kube-api-access-8nxvs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 9 00:30:51.994452 kubelet[2533]: I0909 00:30:51.994281 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-kube-api-access-9mzz8" (OuterVolumeSpecName: "kube-api-access-9mzz8") pod "26b20ccf-b4d8-42dc-8854-c6374a65c9b4" (UID: "26b20ccf-b4d8-42dc-8854-c6374a65c9b4"). InnerVolumeSpecName "kube-api-access-9mzz8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 9 00:30:52.000020 kubelet[2533]: I0909 00:30:51.999962 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "26b20ccf-b4d8-42dc-8854-c6374a65c9b4" (UID: "26b20ccf-b4d8-42dc-8854-c6374a65c9b4"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 9 00:30:52.000385 kubelet[2533]: I0909 00:30:52.000325 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 9 00:30:52.000778 kubelet[2533]: I0909 00:30:52.000737 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 9 00:30:52.000925 kubelet[2533]: I0909 00:30:52.000895 2533 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" (UID: "8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 9 00:30:52.035550 systemd[1]: Removed slice kubepods-besteffort-pod26b20ccf_b4d8_42dc_8854_c6374a65c9b4.slice - libcontainer container kubepods-besteffort-pod26b20ccf_b4d8_42dc_8854_c6374a65c9b4.slice. Sep 9 00:30:52.037971 systemd[1]: Removed slice kubepods-burstable-pod8dfa58a8_5c75_49b7_bb91_bfbce0f73f4a.slice - libcontainer container kubepods-burstable-pod8dfa58a8_5c75_49b7_bb91_bfbce0f73f4a.slice. Sep 9 00:30:52.038142 systemd[1]: kubepods-burstable-pod8dfa58a8_5c75_49b7_bb91_bfbce0f73f4a.slice: Consumed 7.739s CPU time. Sep 9 00:30:52.059236 kubelet[2533]: I0909 00:30:52.059192 2533 scope.go:117] "RemoveContainer" containerID="de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87" Sep 9 00:30:52.061355 containerd[1457]: time="2025-09-09T00:30:52.061310498Z" level=info msg="RemoveContainer for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090835 2533 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090872 2533 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nxvs\" (UniqueName: \"kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-kube-api-access-8nxvs\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090887 2533 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090897 2533 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090905 2533 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090914 2533 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090923 2533 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9mzz8\" (UniqueName: \"kubernetes.io/projected/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-kube-api-access-9mzz8\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.090909 kubelet[2533]: I0909 00:30:52.090933 2533 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.091314 kubelet[2533]: I0909 00:30:52.090941 2533 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.091314 kubelet[2533]: I0909 00:30:52.090950 2533 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/26b20ccf-b4d8-42dc-8854-c6374a65c9b4-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.091314 kubelet[2533]: I0909 00:30:52.090959 2533 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.091314 kubelet[2533]: I0909 00:30:52.090967 2533 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.091314 kubelet[2533]: I0909 00:30:52.090975 2533 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 9 00:30:52.194366 containerd[1457]: time="2025-09-09T00:30:52.194287268Z" level=info msg="RemoveContainer for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" returns successfully" Sep 9 00:30:52.194727 kubelet[2533]: I0909 00:30:52.194690 2533 scope.go:117] "RemoveContainer" containerID="3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea" Sep 9 00:30:52.196143 containerd[1457]: time="2025-09-09T00:30:52.195874913Z" level=info msg="RemoveContainer for \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\"" Sep 9 00:30:52.307667 containerd[1457]: time="2025-09-09T00:30:52.307514494Z" level=info msg="RemoveContainer for \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\" returns successfully" Sep 9 00:30:52.307892 kubelet[2533]: I0909 00:30:52.307819 2533 scope.go:117] "RemoveContainer" containerID="110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364" Sep 9 00:30:52.308910 containerd[1457]: time="2025-09-09T00:30:52.308883933Z" level=info msg="RemoveContainer for \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\"" Sep 9 00:30:52.409476 containerd[1457]: time="2025-09-09T00:30:52.409396392Z" level=info msg="RemoveContainer for \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\" returns successfully" Sep 9 00:30:52.409710 kubelet[2533]: I0909 00:30:52.409681 2533 scope.go:117] "RemoveContainer" containerID="e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62" Sep 9 00:30:52.411119 containerd[1457]: time="2025-09-09T00:30:52.411067294Z" level=info msg="RemoveContainer for \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\"" Sep 9 00:30:52.486322 containerd[1457]: time="2025-09-09T00:30:52.486249543Z" level=info msg="RemoveContainer for \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\" returns successfully" Sep 9 00:30:52.486665 kubelet[2533]: I0909 00:30:52.486606 2533 scope.go:117] "RemoveContainer" containerID="e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f" Sep 9 00:30:52.488285 containerd[1457]: time="2025-09-09T00:30:52.488248929Z" level=info msg="RemoveContainer for \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\"" Sep 9 00:30:52.751177 containerd[1457]: time="2025-09-09T00:30:52.751124083Z" level=info msg="RemoveContainer for \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\" returns successfully" Sep 9 00:30:52.751587 kubelet[2533]: I0909 00:30:52.751548 2533 scope.go:117] "RemoveContainer" containerID="de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87" Sep 9 00:30:52.756292 containerd[1457]: time="2025-09-09T00:30:52.756217733Z" level=error msg="ContainerStatus for \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\": not found" Sep 9 00:30:52.769901 kubelet[2533]: E0909 00:30:52.769859 2533 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\": not found" containerID="de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87" Sep 9 00:30:52.770017 kubelet[2533]: I0909 00:30:52.769910 2533 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87"} err="failed to get container status \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\": rpc error: code = NotFound desc = an error occurred when try to find container \"de9f69c4829466159353e324ba620c00cd9ae65851e54fbcdcf55fac74675f87\": not found" Sep 9 00:30:52.770017 kubelet[2533]: I0909 00:30:52.770009 2533 scope.go:117] "RemoveContainer" containerID="3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea" Sep 9 00:30:52.770454 containerd[1457]: time="2025-09-09T00:30:52.770398124Z" level=error msg="ContainerStatus for \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\": not found" Sep 9 00:30:52.770606 kubelet[2533]: E0909 00:30:52.770568 2533 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\": not found" containerID="3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea" Sep 9 00:30:52.770668 kubelet[2533]: I0909 00:30:52.770607 2533 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea"} err="failed to get container status \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\": rpc error: code = NotFound desc = an error occurred when try to find container \"3d96d6e11fd3744023f9281b291452521754b50217d257874b383632467999ea\": not found" Sep 9 00:30:52.770668 kubelet[2533]: I0909 00:30:52.770629 2533 scope.go:117] "RemoveContainer" containerID="110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364" Sep 9 00:30:52.770947 containerd[1457]: time="2025-09-09T00:30:52.770849081Z" level=error msg="ContainerStatus for \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\": not found" Sep 9 00:30:52.771165 kubelet[2533]: E0909 00:30:52.771098 2533 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\": not found" containerID="110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364" Sep 9 00:30:52.771252 kubelet[2533]: I0909 00:30:52.771146 2533 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364"} err="failed to get container status \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\": rpc error: code = NotFound desc = an error occurred when try to find container \"110f6b36b7995c75dc4ad6d537bfb3a55131c6abe7a7f8088e79cfd3b081b364\": not found" Sep 9 00:30:52.771252 kubelet[2533]: I0909 00:30:52.771187 2533 scope.go:117] "RemoveContainer" containerID="e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62" Sep 9 00:30:52.771478 containerd[1457]: time="2025-09-09T00:30:52.771438430Z" level=error msg="ContainerStatus for \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\": not found" Sep 9 00:30:52.771607 kubelet[2533]: E0909 00:30:52.771577 2533 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\": not found" containerID="e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62" Sep 9 00:30:52.771683 kubelet[2533]: I0909 00:30:52.771606 2533 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62"} err="failed to get container status \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\": rpc error: code = NotFound desc = an error occurred when try to find container \"e739222da403de9fda47c2a1d9aa3d6a41a4baca0bfbb2774ffe372f7586cf62\": not found" Sep 9 00:30:52.771683 kubelet[2533]: I0909 00:30:52.771628 2533 scope.go:117] "RemoveContainer" containerID="e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f" Sep 9 00:30:52.771918 containerd[1457]: time="2025-09-09T00:30:52.771866252Z" level=error msg="ContainerStatus for \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\": not found" Sep 9 00:30:52.772134 kubelet[2533]: E0909 00:30:52.772058 2533 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\": not found" containerID="e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f" Sep 9 00:30:52.772134 kubelet[2533]: I0909 00:30:52.772124 2533 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f"} err="failed to get container status \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\": rpc error: code = NotFound desc = an error occurred when try to find container \"e6595fbe9bdacfe6058c0ec8a2fa936b9f435795c7cc18b48f546e849baca58f\": not found" Sep 9 00:30:52.772216 kubelet[2533]: I0909 00:30:52.772150 2533 scope.go:117] "RemoveContainer" containerID="347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2" Sep 9 00:30:52.773619 containerd[1457]: time="2025-09-09T00:30:52.773580937Z" level=info msg="RemoveContainer for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\"" Sep 9 00:30:52.790560 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161-rootfs.mount: Deactivated successfully. Sep 9 00:30:52.790683 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e62c69acc9da22b245a88e12d51cfb39d30a4a3b8e86abb75ad8b3eb238c0cc3-rootfs.mount: Deactivated successfully. Sep 9 00:30:52.790799 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-bd64b3d5f91bcb0b40457a3c64ce2b455c78362d52e815d25c57cc1c617b6161-shm.mount: Deactivated successfully. Sep 9 00:30:52.790882 systemd[1]: var-lib-kubelet-pods-8dfa58a8\x2d5c75\x2d49b7\x2dbb91\x2dbfbce0f73f4a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d8nxvs.mount: Deactivated successfully. Sep 9 00:30:52.790967 systemd[1]: var-lib-kubelet-pods-26b20ccf\x2db4d8\x2d42dc\x2d8854\x2dc6374a65c9b4-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d9mzz8.mount: Deactivated successfully. Sep 9 00:30:52.791060 systemd[1]: var-lib-kubelet-pods-8dfa58a8\x2d5c75\x2d49b7\x2dbb91\x2dbfbce0f73f4a-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 9 00:30:52.791169 systemd[1]: var-lib-kubelet-pods-8dfa58a8\x2d5c75\x2d49b7\x2dbb91\x2dbfbce0f73f4a-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 9 00:30:53.045603 containerd[1457]: time="2025-09-09T00:30:53.045429427Z" level=info msg="RemoveContainer for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" returns successfully" Sep 9 00:30:53.046066 containerd[1457]: time="2025-09-09T00:30:53.046008525Z" level=error msg="ContainerStatus for \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\": not found" Sep 9 00:30:53.046125 kubelet[2533]: I0909 00:30:53.045726 2533 scope.go:117] "RemoveContainer" containerID="347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2" Sep 9 00:30:53.046407 kubelet[2533]: E0909 00:30:53.046246 2533 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\": not found" containerID="347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2" Sep 9 00:30:53.046407 kubelet[2533]: I0909 00:30:53.046275 2533 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2"} err="failed to get container status \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\": rpc error: code = NotFound desc = an error occurred when try to find container \"347c7d2676e039dbdee8537f207f4b84b8f955f604ace848d9dab11e17497ba2\": not found" Sep 9 00:30:53.776373 sshd[4200]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:53.785181 systemd[1]: sshd@25-10.0.0.113:22-10.0.0.1:56984.service: Deactivated successfully. Sep 9 00:30:53.787291 systemd[1]: session-25.scope: Deactivated successfully. Sep 9 00:30:53.788868 systemd-logind[1441]: Session 25 logged out. Waiting for processes to exit. Sep 9 00:30:53.795475 systemd[1]: Started sshd@26-10.0.0.113:22-10.0.0.1:57000.service - OpenSSH per-connection server daemon (10.0.0.1:57000). Sep 9 00:30:53.796637 systemd-logind[1441]: Removed session 25. Sep 9 00:30:53.832358 sshd[4368]: Accepted publickey for core from 10.0.0.1 port 57000 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:53.833959 sshd[4368]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:53.838227 systemd-logind[1441]: New session 26 of user core. Sep 9 00:30:53.850262 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 9 00:30:54.027170 kubelet[2533]: I0909 00:30:54.027020 2533 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26b20ccf-b4d8-42dc-8854-c6374a65c9b4" path="/var/lib/kubelet/pods/26b20ccf-b4d8-42dc-8854-c6374a65c9b4/volumes" Sep 9 00:30:54.027866 kubelet[2533]: I0909 00:30:54.027840 2533 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" path="/var/lib/kubelet/pods/8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a/volumes" Sep 9 00:30:54.362309 sshd[4368]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:54.372684 systemd[1]: sshd@26-10.0.0.113:22-10.0.0.1:57000.service: Deactivated successfully. Sep 9 00:30:54.377667 systemd[1]: session-26.scope: Deactivated successfully. Sep 9 00:30:54.379267 systemd-logind[1441]: Session 26 logged out. Waiting for processes to exit. Sep 9 00:30:54.387922 systemd[1]: Started sshd@27-10.0.0.113:22-10.0.0.1:57012.service - OpenSSH per-connection server daemon (10.0.0.1:57012). Sep 9 00:30:54.392135 kubelet[2533]: I0909 00:30:54.389136 2533 memory_manager.go:355] "RemoveStaleState removing state" podUID="26b20ccf-b4d8-42dc-8854-c6374a65c9b4" containerName="cilium-operator" Sep 9 00:30:54.392135 kubelet[2533]: I0909 00:30:54.389203 2533 memory_manager.go:355] "RemoveStaleState removing state" podUID="8dfa58a8-5c75-49b7-bb91-bfbce0f73f4a" containerName="cilium-agent" Sep 9 00:30:54.397117 systemd-logind[1441]: Removed session 26. Sep 9 00:30:54.403576 kubelet[2533]: I0909 00:30:54.403524 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-cilium-cgroup\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403576 kubelet[2533]: I0909 00:30:54.403566 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-host-proc-sys-kernel\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403576 kubelet[2533]: I0909 00:30:54.403584 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-hostproc\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403801 kubelet[2533]: I0909 00:30:54.403598 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-bpf-maps\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403801 kubelet[2533]: I0909 00:30:54.403613 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-cilium-run\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403801 kubelet[2533]: I0909 00:30:54.403628 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/7624245c-d16d-4cc3-9a41-1d1995b43245-cilium-ipsec-secrets\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403801 kubelet[2533]: I0909 00:30:54.403642 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7624245c-d16d-4cc3-9a41-1d1995b43245-cilium-config-path\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403801 kubelet[2533]: I0909 00:30:54.403656 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7624245c-d16d-4cc3-9a41-1d1995b43245-hubble-tls\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.403801 kubelet[2533]: I0909 00:30:54.403670 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-cni-path\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.404030 kubelet[2533]: I0909 00:30:54.403690 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7624245c-d16d-4cc3-9a41-1d1995b43245-clustermesh-secrets\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.404030 kubelet[2533]: I0909 00:30:54.403705 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg5xq\" (UniqueName: \"kubernetes.io/projected/7624245c-d16d-4cc3-9a41-1d1995b43245-kube-api-access-mg5xq\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.404030 kubelet[2533]: I0909 00:30:54.403721 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-etc-cni-netd\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.404030 kubelet[2533]: I0909 00:30:54.403744 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-lib-modules\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.404030 kubelet[2533]: I0909 00:30:54.403758 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-xtables-lock\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.404030 kubelet[2533]: I0909 00:30:54.403771 2533 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7624245c-d16d-4cc3-9a41-1d1995b43245-host-proc-sys-net\") pod \"cilium-v54jp\" (UID: \"7624245c-d16d-4cc3-9a41-1d1995b43245\") " pod="kube-system/cilium-v54jp" Sep 9 00:30:54.413951 systemd[1]: Created slice kubepods-burstable-pod7624245c_d16d_4cc3_9a41_1d1995b43245.slice - libcontainer container kubepods-burstable-pod7624245c_d16d_4cc3_9a41_1d1995b43245.slice. Sep 9 00:30:54.451159 sshd[4381]: Accepted publickey for core from 10.0.0.1 port 57012 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:54.452689 sshd[4381]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:54.457947 systemd-logind[1441]: New session 27 of user core. Sep 9 00:30:54.464272 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 9 00:30:54.523595 sshd[4381]: pam_unix(sshd:session): session closed for user core Sep 9 00:30:54.535240 systemd[1]: sshd@27-10.0.0.113:22-10.0.0.1:57012.service: Deactivated successfully. Sep 9 00:30:54.538179 systemd[1]: session-27.scope: Deactivated successfully. Sep 9 00:30:54.539136 systemd-logind[1441]: Session 27 logged out. Waiting for processes to exit. Sep 9 00:30:54.555886 systemd[1]: Started sshd@28-10.0.0.113:22-10.0.0.1:57028.service - OpenSSH per-connection server daemon (10.0.0.1:57028). Sep 9 00:30:54.557128 systemd-logind[1441]: Removed session 27. Sep 9 00:30:54.608462 sshd[4393]: Accepted publickey for core from 10.0.0.1 port 57028 ssh2: RSA SHA256:LXnX38+Ld2WFs+27NUMweVJjFK11ltvOlPZ2mP3Dpgo Sep 9 00:30:54.610330 sshd[4393]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 9 00:30:54.615580 systemd-logind[1441]: New session 28 of user core. Sep 9 00:30:54.626275 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 9 00:30:54.717156 kubelet[2533]: E0909 00:30:54.717092 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:54.717897 containerd[1457]: time="2025-09-09T00:30:54.717825697Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-v54jp,Uid:7624245c-d16d-4cc3-9a41-1d1995b43245,Namespace:kube-system,Attempt:0,}" Sep 9 00:30:54.779255 containerd[1457]: time="2025-09-09T00:30:54.779098270Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 9 00:30:54.779255 containerd[1457]: time="2025-09-09T00:30:54.779175266Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 9 00:30:54.779255 containerd[1457]: time="2025-09-09T00:30:54.779190134Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:30:54.779454 containerd[1457]: time="2025-09-09T00:30:54.779300002Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 9 00:30:54.805277 systemd[1]: Started cri-containerd-92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83.scope - libcontainer container 92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83. Sep 9 00:30:54.832624 containerd[1457]: time="2025-09-09T00:30:54.832574716Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-v54jp,Uid:7624245c-d16d-4cc3-9a41-1d1995b43245,Namespace:kube-system,Attempt:0,} returns sandbox id \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\"" Sep 9 00:30:54.833783 kubelet[2533]: E0909 00:30:54.833462 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:54.835839 containerd[1457]: time="2025-09-09T00:30:54.835770412Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 9 00:30:54.850907 containerd[1457]: time="2025-09-09T00:30:54.850822891Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763\"" Sep 9 00:30:54.851573 containerd[1457]: time="2025-09-09T00:30:54.851487012Z" level=info msg="StartContainer for \"27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763\"" Sep 9 00:30:54.884307 systemd[1]: Started cri-containerd-27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763.scope - libcontainer container 27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763. Sep 9 00:30:54.918469 containerd[1457]: time="2025-09-09T00:30:54.918401242Z" level=info msg="StartContainer for \"27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763\" returns successfully" Sep 9 00:30:54.930191 systemd[1]: cri-containerd-27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763.scope: Deactivated successfully. Sep 9 00:30:54.969055 containerd[1457]: time="2025-09-09T00:30:54.968955042Z" level=info msg="shim disconnected" id=27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763 namespace=k8s.io Sep 9 00:30:54.969055 containerd[1457]: time="2025-09-09T00:30:54.969047377Z" level=warning msg="cleaning up after shim disconnected" id=27157c1cc60eceec3e7aa39fc6d4b19a7f00a403b57a5982cba45ff373128763 namespace=k8s.io Sep 9 00:30:54.969055 containerd[1457]: time="2025-09-09T00:30:54.969061844Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:55.069721 kubelet[2533]: E0909 00:30:55.069683 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:55.071581 containerd[1457]: time="2025-09-09T00:30:55.071457983Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 9 00:30:55.088886 containerd[1457]: time="2025-09-09T00:30:55.088810324Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d\"" Sep 9 00:30:55.089544 containerd[1457]: time="2025-09-09T00:30:55.089500753Z" level=info msg="StartContainer for \"976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d\"" Sep 9 00:30:55.121330 systemd[1]: Started cri-containerd-976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d.scope - libcontainer container 976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d. Sep 9 00:30:55.155181 containerd[1457]: time="2025-09-09T00:30:55.155026184Z" level=info msg="StartContainer for \"976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d\" returns successfully" Sep 9 00:30:55.165924 systemd[1]: cri-containerd-976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d.scope: Deactivated successfully. Sep 9 00:30:55.194535 containerd[1457]: time="2025-09-09T00:30:55.194448647Z" level=info msg="shim disconnected" id=976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d namespace=k8s.io Sep 9 00:30:55.194535 containerd[1457]: time="2025-09-09T00:30:55.194521586Z" level=warning msg="cleaning up after shim disconnected" id=976ca8a36cc5c9c20bd20dffe88e4006641393c46605fc29fb7347e9150d888d namespace=k8s.io Sep 9 00:30:55.194535 containerd[1457]: time="2025-09-09T00:30:55.194532105Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:56.073438 kubelet[2533]: E0909 00:30:56.073401 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:56.075258 containerd[1457]: time="2025-09-09T00:30:56.075128974Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 9 00:30:56.089560 kubelet[2533]: E0909 00:30:56.089513 2533 kubelet.go:3002] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 9 00:30:56.194829 containerd[1457]: time="2025-09-09T00:30:56.194722795Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517\"" Sep 9 00:30:56.195490 containerd[1457]: time="2025-09-09T00:30:56.195439004Z" level=info msg="StartContainer for \"f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517\"" Sep 9 00:30:56.233270 systemd[1]: Started cri-containerd-f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517.scope - libcontainer container f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517. Sep 9 00:30:56.270421 systemd[1]: cri-containerd-f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517.scope: Deactivated successfully. Sep 9 00:30:56.272593 containerd[1457]: time="2025-09-09T00:30:56.272537159Z" level=info msg="StartContainer for \"f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517\" returns successfully" Sep 9 00:30:56.302102 containerd[1457]: time="2025-09-09T00:30:56.302008672Z" level=info msg="shim disconnected" id=f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517 namespace=k8s.io Sep 9 00:30:56.302102 containerd[1457]: time="2025-09-09T00:30:56.302095847Z" level=warning msg="cleaning up after shim disconnected" id=f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517 namespace=k8s.io Sep 9 00:30:56.302102 containerd[1457]: time="2025-09-09T00:30:56.302110846Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:56.512613 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f5cb062aed221a78677a8cc52b059a16a34b0fdf57a77d96051a496e839a0517-rootfs.mount: Deactivated successfully. Sep 9 00:30:57.077809 kubelet[2533]: E0909 00:30:57.077765 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:57.079484 containerd[1457]: time="2025-09-09T00:30:57.079430598Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 9 00:30:57.094811 containerd[1457]: time="2025-09-09T00:30:57.094755092Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec\"" Sep 9 00:30:57.097756 containerd[1457]: time="2025-09-09T00:30:57.096554187Z" level=info msg="StartContainer for \"6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec\"" Sep 9 00:30:57.131957 systemd[1]: Started cri-containerd-6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec.scope - libcontainer container 6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec. Sep 9 00:30:57.161521 systemd[1]: cri-containerd-6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec.scope: Deactivated successfully. Sep 9 00:30:57.163861 containerd[1457]: time="2025-09-09T00:30:57.163779453Z" level=info msg="StartContainer for \"6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec\" returns successfully" Sep 9 00:30:57.193851 containerd[1457]: time="2025-09-09T00:30:57.193757891Z" level=info msg="shim disconnected" id=6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec namespace=k8s.io Sep 9 00:30:57.193851 containerd[1457]: time="2025-09-09T00:30:57.193840498Z" level=warning msg="cleaning up after shim disconnected" id=6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec namespace=k8s.io Sep 9 00:30:57.193851 containerd[1457]: time="2025-09-09T00:30:57.193853753Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 9 00:30:57.513180 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6521388a971cbdfa8306375c900d5a30533b51bc840917f3cf891ad6cd2c6bec-rootfs.mount: Deactivated successfully. Sep 9 00:30:58.082024 kubelet[2533]: E0909 00:30:58.081985 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:30:58.083879 containerd[1457]: time="2025-09-09T00:30:58.083835963Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 9 00:30:58.095101 kubelet[2533]: I0909 00:30:58.095031 2533 setters.go:602] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-09T00:30:58Z","lastTransitionTime":"2025-09-09T00:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 9 00:30:58.103472 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1443763844.mount: Deactivated successfully. Sep 9 00:30:58.105263 containerd[1457]: time="2025-09-09T00:30:58.105214150Z" level=info msg="CreateContainer within sandbox \"92ed6a1948d1d04d68822498972a9cb65d7688238b5b8275baed6a5761548e83\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"dbae34b304788156b7bd5f3ed3dffa7591192f04e65311a74430cbd806f6fe5c\"" Sep 9 00:30:58.105934 containerd[1457]: time="2025-09-09T00:30:58.105895412Z" level=info msg="StartContainer for \"dbae34b304788156b7bd5f3ed3dffa7591192f04e65311a74430cbd806f6fe5c\"" Sep 9 00:30:58.148248 systemd[1]: Started cri-containerd-dbae34b304788156b7bd5f3ed3dffa7591192f04e65311a74430cbd806f6fe5c.scope - libcontainer container dbae34b304788156b7bd5f3ed3dffa7591192f04e65311a74430cbd806f6fe5c. Sep 9 00:30:58.189194 containerd[1457]: time="2025-09-09T00:30:58.189136679Z" level=info msg="StartContainer for \"dbae34b304788156b7bd5f3ed3dffa7591192f04e65311a74430cbd806f6fe5c\" returns successfully" Sep 9 00:30:58.647121 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 9 00:30:59.087355 kubelet[2533]: E0909 00:30:59.087315 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:31:00.718947 kubelet[2533]: E0909 00:31:00.718365 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:31:01.921492 systemd-networkd[1384]: lxc_health: Link UP Sep 9 00:31:01.932657 systemd-networkd[1384]: lxc_health: Gained carrier Sep 9 00:31:02.719622 kubelet[2533]: E0909 00:31:02.719248 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:31:02.874837 kubelet[2533]: I0909 00:31:02.874746 2533 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-v54jp" podStartSLOduration=8.874702499 podStartE2EDuration="8.874702499s" podCreationTimestamp="2025-09-09 00:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-09 00:30:59.101255306 +0000 UTC m=+103.201009748" watchObservedRunningTime="2025-09-09 00:31:02.874702499 +0000 UTC m=+106.974456922" Sep 9 00:31:03.096595 kubelet[2533]: E0909 00:31:03.096538 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:31:03.387230 systemd-networkd[1384]: lxc_health: Gained IPv6LL Sep 9 00:31:04.097537 kubelet[2533]: E0909 00:31:04.097489 2533 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 9 00:31:07.300065 sshd[4393]: pam_unix(sshd:session): session closed for user core Sep 9 00:31:07.304793 systemd[1]: sshd@28-10.0.0.113:22-10.0.0.1:57028.service: Deactivated successfully. Sep 9 00:31:07.307441 systemd[1]: session-28.scope: Deactivated successfully. Sep 9 00:31:07.308145 systemd-logind[1441]: Session 28 logged out. Waiting for processes to exit. Sep 9 00:31:07.309522 systemd-logind[1441]: Removed session 28.