Aug 13 07:08:30.887708 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Aug 12 22:14:58 -00 2025 Aug 13 07:08:30.887730 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:08:30.887741 kernel: BIOS-provided physical RAM map: Aug 13 07:08:30.887748 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Aug 13 07:08:30.887754 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Aug 13 07:08:30.887760 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Aug 13 07:08:30.887767 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Aug 13 07:08:30.887774 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Aug 13 07:08:30.887780 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable Aug 13 07:08:30.887786 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS Aug 13 07:08:30.887795 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable Aug 13 07:08:30.887801 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009c9eefff] reserved Aug 13 07:08:30.887811 kernel: BIOS-e820: [mem 0x000000009c9ef000-0x000000009caeefff] type 20 Aug 13 07:08:30.887818 kernel: BIOS-e820: [mem 0x000000009caef000-0x000000009cb6efff] reserved Aug 13 07:08:30.887830 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data Aug 13 07:08:30.887839 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Aug 13 07:08:30.887853 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable Aug 13 07:08:30.887861 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved Aug 13 07:08:30.887868 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Aug 13 07:08:30.887875 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Aug 13 07:08:30.887881 kernel: NX (Execute Disable) protection: active Aug 13 07:08:30.887888 kernel: APIC: Static calls initialized Aug 13 07:08:30.887895 kernel: efi: EFI v2.7 by EDK II Aug 13 07:08:30.887901 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b674118 Aug 13 07:08:30.887908 kernel: SMBIOS 2.8 present. Aug 13 07:08:30.887928 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 0.0.0 02/06/2015 Aug 13 07:08:30.887935 kernel: Hypervisor detected: KVM Aug 13 07:08:30.887945 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Aug 13 07:08:30.887953 kernel: kvm-clock: using sched offset of 5135042059 cycles Aug 13 07:08:30.887960 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Aug 13 07:08:30.887967 kernel: tsc: Detected 2794.750 MHz processor Aug 13 07:08:30.887974 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Aug 13 07:08:30.887982 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Aug 13 07:08:30.887989 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x400000000 Aug 13 07:08:30.887996 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Aug 13 07:08:30.888003 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Aug 13 07:08:30.888012 kernel: Using GB pages for direct mapping Aug 13 07:08:30.888019 kernel: Secure boot disabled Aug 13 07:08:30.888026 kernel: ACPI: Early table checksum verification disabled Aug 13 07:08:30.888033 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Aug 13 07:08:30.888044 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Aug 13 07:08:30.888052 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 07:08:30.888059 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 07:08:30.888069 kernel: ACPI: FACS 0x000000009CBDD000 000040 Aug 13 07:08:30.888076 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 07:08:30.888087 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 07:08:30.888095 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 07:08:30.888102 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 07:08:30.888109 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Aug 13 07:08:30.888117 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Aug 13 07:08:30.888127 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Aug 13 07:08:30.888134 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Aug 13 07:08:30.888142 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Aug 13 07:08:30.888149 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Aug 13 07:08:30.888156 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Aug 13 07:08:30.888163 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Aug 13 07:08:30.888170 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Aug 13 07:08:30.888178 kernel: No NUMA configuration found Aug 13 07:08:30.888187 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] Aug 13 07:08:30.888198 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] Aug 13 07:08:30.888205 kernel: Zone ranges: Aug 13 07:08:30.888213 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Aug 13 07:08:30.888220 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] Aug 13 07:08:30.888227 kernel: Normal empty Aug 13 07:08:30.888234 kernel: Movable zone start for each node Aug 13 07:08:30.888242 kernel: Early memory node ranges Aug 13 07:08:30.888249 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Aug 13 07:08:30.888256 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Aug 13 07:08:30.888263 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Aug 13 07:08:30.888277 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] Aug 13 07:08:30.888291 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] Aug 13 07:08:30.888309 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] Aug 13 07:08:30.888329 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] Aug 13 07:08:30.888348 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 07:08:30.888367 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Aug 13 07:08:30.888385 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Aug 13 07:08:30.888403 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 07:08:30.888421 kernel: On node 0, zone DMA: 240 pages in unavailable ranges Aug 13 07:08:30.888442 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Aug 13 07:08:30.888450 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges Aug 13 07:08:30.888457 kernel: ACPI: PM-Timer IO Port: 0x608 Aug 13 07:08:30.888464 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Aug 13 07:08:30.888471 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Aug 13 07:08:30.888479 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Aug 13 07:08:30.888486 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Aug 13 07:08:30.888493 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Aug 13 07:08:30.888501 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Aug 13 07:08:30.888511 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Aug 13 07:08:30.888518 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Aug 13 07:08:30.888525 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Aug 13 07:08:30.888532 kernel: TSC deadline timer available Aug 13 07:08:30.888540 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Aug 13 07:08:30.888547 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Aug 13 07:08:30.888554 kernel: kvm-guest: KVM setup pv remote TLB flush Aug 13 07:08:30.888561 kernel: kvm-guest: setup PV sched yield Aug 13 07:08:30.888569 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Aug 13 07:08:30.888578 kernel: Booting paravirtualized kernel on KVM Aug 13 07:08:30.888586 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Aug 13 07:08:30.888593 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Aug 13 07:08:30.888600 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u524288 Aug 13 07:08:30.888608 kernel: pcpu-alloc: s197096 r8192 d32280 u524288 alloc=1*2097152 Aug 13 07:08:30.888615 kernel: pcpu-alloc: [0] 0 1 2 3 Aug 13 07:08:30.888622 kernel: kvm-guest: PV spinlocks enabled Aug 13 07:08:30.888629 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Aug 13 07:08:30.888638 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:08:30.888651 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 07:08:30.888659 kernel: random: crng init done Aug 13 07:08:30.888666 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 07:08:30.888674 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:08:30.888681 kernel: Fallback order for Node 0: 0 Aug 13 07:08:30.888688 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 Aug 13 07:08:30.888714 kernel: Policy zone: DMA32 Aug 13 07:08:30.888722 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 07:08:30.888743 kernel: Memory: 2395616K/2567000K available (12288K kernel code, 2295K rwdata, 22748K rodata, 42876K init, 2316K bss, 171124K reserved, 0K cma-reserved) Aug 13 07:08:30.888751 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Aug 13 07:08:30.888758 kernel: ftrace: allocating 37968 entries in 149 pages Aug 13 07:08:30.888766 kernel: ftrace: allocated 149 pages with 4 groups Aug 13 07:08:30.888773 kernel: Dynamic Preempt: voluntary Aug 13 07:08:30.888789 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 07:08:30.888800 kernel: rcu: RCU event tracing is enabled. Aug 13 07:08:30.888808 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Aug 13 07:08:30.888816 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 07:08:30.888823 kernel: Rude variant of Tasks RCU enabled. Aug 13 07:08:30.888833 kernel: Tracing variant of Tasks RCU enabled. Aug 13 07:08:30.888844 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 07:08:30.888858 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Aug 13 07:08:30.888865 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Aug 13 07:08:30.888876 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 07:08:30.888884 kernel: Console: colour dummy device 80x25 Aug 13 07:08:30.888891 kernel: printk: console [ttyS0] enabled Aug 13 07:08:30.888902 kernel: ACPI: Core revision 20230628 Aug 13 07:08:30.888910 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Aug 13 07:08:30.888931 kernel: APIC: Switch to symmetric I/O mode setup Aug 13 07:08:30.888939 kernel: x2apic enabled Aug 13 07:08:30.888947 kernel: APIC: Switched APIC routing to: physical x2apic Aug 13 07:08:30.888954 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Aug 13 07:08:30.888962 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Aug 13 07:08:30.888970 kernel: kvm-guest: setup PV IPIs Aug 13 07:08:30.888977 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Aug 13 07:08:30.888988 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Aug 13 07:08:30.888996 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Aug 13 07:08:30.889004 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Aug 13 07:08:30.889011 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Aug 13 07:08:30.889019 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Aug 13 07:08:30.889027 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Aug 13 07:08:30.889034 kernel: Spectre V2 : Mitigation: Retpolines Aug 13 07:08:30.889042 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Aug 13 07:08:30.889049 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Aug 13 07:08:30.889060 kernel: RETBleed: Mitigation: untrained return thunk Aug 13 07:08:30.889068 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Aug 13 07:08:30.889077 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Aug 13 07:08:30.889086 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Aug 13 07:08:30.889099 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Aug 13 07:08:30.889107 kernel: x86/bugs: return thunk changed Aug 13 07:08:30.889115 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Aug 13 07:08:30.889122 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Aug 13 07:08:30.889130 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Aug 13 07:08:30.889140 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Aug 13 07:08:30.889148 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Aug 13 07:08:30.889155 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Aug 13 07:08:30.889163 kernel: Freeing SMP alternatives memory: 32K Aug 13 07:08:30.889171 kernel: pid_max: default: 32768 minimum: 301 Aug 13 07:08:30.889178 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 07:08:30.889186 kernel: landlock: Up and running. Aug 13 07:08:30.889193 kernel: SELinux: Initializing. Aug 13 07:08:30.889201 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 07:08:30.889211 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 07:08:30.889219 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Aug 13 07:08:30.889227 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Aug 13 07:08:30.889234 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Aug 13 07:08:30.889242 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Aug 13 07:08:30.889250 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Aug 13 07:08:30.889258 kernel: ... version: 0 Aug 13 07:08:30.889265 kernel: ... bit width: 48 Aug 13 07:08:30.889276 kernel: ... generic registers: 6 Aug 13 07:08:30.889283 kernel: ... value mask: 0000ffffffffffff Aug 13 07:08:30.889291 kernel: ... max period: 00007fffffffffff Aug 13 07:08:30.889298 kernel: ... fixed-purpose events: 0 Aug 13 07:08:30.889306 kernel: ... event mask: 000000000000003f Aug 13 07:08:30.889314 kernel: signal: max sigframe size: 1776 Aug 13 07:08:30.889321 kernel: rcu: Hierarchical SRCU implementation. Aug 13 07:08:30.889329 kernel: rcu: Max phase no-delay instances is 400. Aug 13 07:08:30.889337 kernel: smp: Bringing up secondary CPUs ... Aug 13 07:08:30.889344 kernel: smpboot: x86: Booting SMP configuration: Aug 13 07:08:30.889354 kernel: .... node #0, CPUs: #1 #2 #3 Aug 13 07:08:30.889362 kernel: smp: Brought up 1 node, 4 CPUs Aug 13 07:08:30.889369 kernel: smpboot: Max logical packages: 1 Aug 13 07:08:30.889377 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Aug 13 07:08:30.889385 kernel: devtmpfs: initialized Aug 13 07:08:30.889392 kernel: x86/mm: Memory block size: 128MB Aug 13 07:08:30.889400 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Aug 13 07:08:30.889408 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Aug 13 07:08:30.889415 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) Aug 13 07:08:30.889426 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Aug 13 07:08:30.889433 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Aug 13 07:08:30.889441 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 07:08:30.889449 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Aug 13 07:08:30.889457 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 07:08:30.889464 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 07:08:30.889472 kernel: audit: initializing netlink subsys (disabled) Aug 13 07:08:30.889480 kernel: audit: type=2000 audit(1755068910.266:1): state=initialized audit_enabled=0 res=1 Aug 13 07:08:30.889490 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 07:08:30.889497 kernel: thermal_sys: Registered thermal governor 'user_space' Aug 13 07:08:30.889505 kernel: cpuidle: using governor menu Aug 13 07:08:30.889512 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 07:08:30.889520 kernel: dca service started, version 1.12.1 Aug 13 07:08:30.889528 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Aug 13 07:08:30.889535 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Aug 13 07:08:30.889543 kernel: PCI: Using configuration type 1 for base access Aug 13 07:08:30.889551 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Aug 13 07:08:30.889561 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 07:08:30.889569 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 07:08:30.889577 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 07:08:30.889585 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 07:08:30.889592 kernel: ACPI: Added _OSI(Module Device) Aug 13 07:08:30.889600 kernel: ACPI: Added _OSI(Processor Device) Aug 13 07:08:30.889607 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 07:08:30.889615 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Aug 13 07:08:30.889623 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Aug 13 07:08:30.889633 kernel: ACPI: Interpreter enabled Aug 13 07:08:30.889640 kernel: ACPI: PM: (supports S0 S3 S5) Aug 13 07:08:30.889648 kernel: ACPI: Using IOAPIC for interrupt routing Aug 13 07:08:30.889655 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Aug 13 07:08:30.889663 kernel: PCI: Using E820 reservations for host bridge windows Aug 13 07:08:30.889671 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Aug 13 07:08:30.889678 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Aug 13 07:08:30.889981 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 07:08:30.890153 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Aug 13 07:08:30.890284 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Aug 13 07:08:30.890294 kernel: PCI host bridge to bus 0000:00 Aug 13 07:08:30.890445 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Aug 13 07:08:30.890566 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Aug 13 07:08:30.890683 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Aug 13 07:08:30.890811 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Aug 13 07:08:30.891038 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Aug 13 07:08:30.891159 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0xfffffffff window] Aug 13 07:08:30.891275 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Aug 13 07:08:30.891443 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Aug 13 07:08:30.891612 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Aug 13 07:08:30.891753 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Aug 13 07:08:30.891892 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Aug 13 07:08:30.892047 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Aug 13 07:08:30.892180 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Aug 13 07:08:30.892305 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Aug 13 07:08:30.892457 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Aug 13 07:08:30.892587 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Aug 13 07:08:30.892725 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Aug 13 07:08:30.892862 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] Aug 13 07:08:30.893049 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Aug 13 07:08:30.893209 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Aug 13 07:08:30.893359 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Aug 13 07:08:30.893490 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] Aug 13 07:08:30.893658 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Aug 13 07:08:30.893848 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Aug 13 07:08:30.894069 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Aug 13 07:08:30.894220 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] Aug 13 07:08:30.894350 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Aug 13 07:08:30.894498 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Aug 13 07:08:30.894626 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Aug 13 07:08:30.894782 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Aug 13 07:08:30.894978 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Aug 13 07:08:30.895120 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Aug 13 07:08:30.895261 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Aug 13 07:08:30.895406 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Aug 13 07:08:30.895418 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Aug 13 07:08:30.895426 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Aug 13 07:08:30.895434 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Aug 13 07:08:30.895442 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Aug 13 07:08:30.895450 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Aug 13 07:08:30.895463 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Aug 13 07:08:30.895471 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Aug 13 07:08:30.895478 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Aug 13 07:08:30.895486 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Aug 13 07:08:30.895493 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Aug 13 07:08:30.895501 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Aug 13 07:08:30.895509 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Aug 13 07:08:30.895516 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Aug 13 07:08:30.895524 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Aug 13 07:08:30.895534 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Aug 13 07:08:30.895541 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Aug 13 07:08:30.895549 kernel: iommu: Default domain type: Translated Aug 13 07:08:30.895557 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Aug 13 07:08:30.895564 kernel: efivars: Registered efivars operations Aug 13 07:08:30.895572 kernel: PCI: Using ACPI for IRQ routing Aug 13 07:08:30.895580 kernel: PCI: pci_cache_line_size set to 64 bytes Aug 13 07:08:30.895587 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Aug 13 07:08:30.895595 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] Aug 13 07:08:30.895605 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] Aug 13 07:08:30.895613 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] Aug 13 07:08:30.895762 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Aug 13 07:08:30.895904 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Aug 13 07:08:30.896050 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Aug 13 07:08:30.896061 kernel: vgaarb: loaded Aug 13 07:08:30.896069 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Aug 13 07:08:30.896076 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Aug 13 07:08:30.896089 kernel: clocksource: Switched to clocksource kvm-clock Aug 13 07:08:30.896097 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 07:08:30.896104 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 07:08:30.896112 kernel: pnp: PnP ACPI init Aug 13 07:08:30.896286 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Aug 13 07:08:30.896298 kernel: pnp: PnP ACPI: found 6 devices Aug 13 07:08:30.896307 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Aug 13 07:08:30.896315 kernel: NET: Registered PF_INET protocol family Aug 13 07:08:30.896322 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 07:08:30.896334 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Aug 13 07:08:30.896342 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 07:08:30.896350 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Aug 13 07:08:30.896358 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Aug 13 07:08:30.896366 kernel: TCP: Hash tables configured (established 32768 bind 32768) Aug 13 07:08:30.896374 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 07:08:30.896382 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 07:08:30.896390 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 07:08:30.896400 kernel: NET: Registered PF_XDP protocol family Aug 13 07:08:30.896532 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Aug 13 07:08:30.896659 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Aug 13 07:08:30.896789 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Aug 13 07:08:30.896933 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Aug 13 07:08:30.897055 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Aug 13 07:08:30.897171 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Aug 13 07:08:30.897289 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Aug 13 07:08:30.897413 kernel: pci_bus 0000:00: resource 9 [mem 0x800000000-0xfffffffff window] Aug 13 07:08:30.897423 kernel: PCI: CLS 0 bytes, default 64 Aug 13 07:08:30.897431 kernel: Initialise system trusted keyrings Aug 13 07:08:30.897439 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Aug 13 07:08:30.897447 kernel: Key type asymmetric registered Aug 13 07:08:30.897454 kernel: Asymmetric key parser 'x509' registered Aug 13 07:08:30.897462 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Aug 13 07:08:30.897470 kernel: io scheduler mq-deadline registered Aug 13 07:08:30.897478 kernel: io scheduler kyber registered Aug 13 07:08:30.897489 kernel: io scheduler bfq registered Aug 13 07:08:30.897497 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Aug 13 07:08:30.897505 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Aug 13 07:08:30.897513 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Aug 13 07:08:30.897521 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Aug 13 07:08:30.897528 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 07:08:30.897536 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 07:08:30.897544 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Aug 13 07:08:30.897552 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Aug 13 07:08:30.897562 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Aug 13 07:08:30.897725 kernel: rtc_cmos 00:04: RTC can wake from S4 Aug 13 07:08:30.897737 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Aug 13 07:08:30.897867 kernel: rtc_cmos 00:04: registered as rtc0 Aug 13 07:08:30.898084 kernel: rtc_cmos 00:04: setting system clock to 2025-08-13T07:08:30 UTC (1755068910) Aug 13 07:08:30.898205 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Aug 13 07:08:30.898216 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Aug 13 07:08:30.898224 kernel: efifb: probing for efifb Aug 13 07:08:30.898237 kernel: efifb: framebuffer at 0xc0000000, using 1408k, total 1408k Aug 13 07:08:30.898245 kernel: efifb: mode is 800x600x24, linelength=2400, pages=1 Aug 13 07:08:30.898252 kernel: efifb: scrolling: redraw Aug 13 07:08:30.898260 kernel: efifb: Truecolor: size=0:8:8:8, shift=0:16:8:0 Aug 13 07:08:30.898268 kernel: Console: switching to colour frame buffer device 100x37 Aug 13 07:08:30.898276 kernel: fb0: EFI VGA frame buffer device Aug 13 07:08:30.898302 kernel: pstore: Using crash dump compression: deflate Aug 13 07:08:30.898313 kernel: pstore: Registered efi_pstore as persistent store backend Aug 13 07:08:30.898321 kernel: NET: Registered PF_INET6 protocol family Aug 13 07:08:30.898331 kernel: Segment Routing with IPv6 Aug 13 07:08:30.898341 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 07:08:30.898349 kernel: NET: Registered PF_PACKET protocol family Aug 13 07:08:30.898357 kernel: Key type dns_resolver registered Aug 13 07:08:30.898365 kernel: IPI shorthand broadcast: enabled Aug 13 07:08:30.898373 kernel: sched_clock: Marking stable (1043002784, 109201912)->(1334864324, -182659628) Aug 13 07:08:30.898381 kernel: registered taskstats version 1 Aug 13 07:08:30.898389 kernel: Loading compiled-in X.509 certificates Aug 13 07:08:30.898397 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: 264e720147fa8df9744bb9dc1c08171c0cb20041' Aug 13 07:08:30.898408 kernel: Key type .fscrypt registered Aug 13 07:08:30.898415 kernel: Key type fscrypt-provisioning registered Aug 13 07:08:30.898423 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 07:08:30.898431 kernel: ima: Allocated hash algorithm: sha1 Aug 13 07:08:30.898439 kernel: ima: No architecture policies found Aug 13 07:08:30.898447 kernel: clk: Disabling unused clocks Aug 13 07:08:30.898455 kernel: Freeing unused kernel image (initmem) memory: 42876K Aug 13 07:08:30.898463 kernel: Write protecting the kernel read-only data: 36864k Aug 13 07:08:30.898471 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Aug 13 07:08:30.898482 kernel: Run /init as init process Aug 13 07:08:30.898489 kernel: with arguments: Aug 13 07:08:30.898497 kernel: /init Aug 13 07:08:30.898505 kernel: with environment: Aug 13 07:08:30.898513 kernel: HOME=/ Aug 13 07:08:30.898521 kernel: TERM=linux Aug 13 07:08:30.898529 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 07:08:30.898539 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:08:30.898552 systemd[1]: Detected virtualization kvm. Aug 13 07:08:30.898560 systemd[1]: Detected architecture x86-64. Aug 13 07:08:30.898568 systemd[1]: Running in initrd. Aug 13 07:08:30.898576 systemd[1]: No hostname configured, using default hostname. Aug 13 07:08:30.898584 systemd[1]: Hostname set to . Aug 13 07:08:30.898598 systemd[1]: Initializing machine ID from VM UUID. Aug 13 07:08:30.898606 systemd[1]: Queued start job for default target initrd.target. Aug 13 07:08:30.898615 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:08:30.898623 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:08:30.898632 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 07:08:30.898641 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:08:30.898649 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 07:08:30.898660 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 07:08:30.898670 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 07:08:30.898679 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 07:08:30.898687 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:08:30.898705 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:08:30.898713 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:08:30.898721 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:08:30.898732 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:08:30.898741 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:08:30.898749 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:08:30.898757 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:08:30.898766 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 07:08:30.898774 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 07:08:30.898782 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:08:30.898791 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:08:30.898799 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:08:30.898810 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:08:30.898819 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 07:08:30.898828 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:08:30.898839 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 07:08:30.898851 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 07:08:30.898862 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:08:30.898870 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:08:30.898879 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:30.898890 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 07:08:30.898899 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:08:30.898907 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 07:08:30.898930 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 07:08:30.898959 systemd-journald[193]: Collecting audit messages is disabled. Aug 13 07:08:30.898981 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:30.898990 systemd-journald[193]: Journal started Aug 13 07:08:30.899011 systemd-journald[193]: Runtime Journal (/run/log/journal/e56dd7c6833b442fb27df894dda8b6a0) is 6.0M, max 48.3M, 42.2M free. Aug 13 07:08:30.888776 systemd-modules-load[194]: Inserted module 'overlay' Aug 13 07:08:30.903566 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:30.903593 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:08:30.904256 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 07:08:30.908343 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:08:30.910481 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:08:30.918542 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:30.920000 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:08:30.932942 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 07:08:30.934167 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 07:08:30.934637 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:08:30.938854 systemd-modules-load[194]: Inserted module 'br_netfilter' Aug 13 07:08:30.939742 kernel: Bridge firewalling registered Aug 13 07:08:30.941180 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:08:30.943160 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:08:30.950350 dracut-cmdline[221]: dracut-dracut-053 Aug 13 07:08:30.953799 dracut-cmdline[221]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:08:30.955841 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:08:30.966054 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:08:30.997257 systemd-resolved[247]: Positive Trust Anchors: Aug 13 07:08:30.997276 systemd-resolved[247]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:08:30.997307 systemd-resolved[247]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:08:30.999895 systemd-resolved[247]: Defaulting to hostname 'linux'. Aug 13 07:08:31.001235 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:08:31.006626 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:08:31.035949 kernel: SCSI subsystem initialized Aug 13 07:08:31.044943 kernel: Loading iSCSI transport class v2.0-870. Aug 13 07:08:31.055943 kernel: iscsi: registered transport (tcp) Aug 13 07:08:31.076959 kernel: iscsi: registered transport (qla4xxx) Aug 13 07:08:31.077029 kernel: QLogic iSCSI HBA Driver Aug 13 07:08:31.126937 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 07:08:31.144107 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 07:08:31.168761 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 07:08:31.168801 kernel: device-mapper: uevent: version 1.0.3 Aug 13 07:08:31.168820 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 07:08:31.210940 kernel: raid6: avx2x4 gen() 29944 MB/s Aug 13 07:08:31.227935 kernel: raid6: avx2x2 gen() 30864 MB/s Aug 13 07:08:31.244969 kernel: raid6: avx2x1 gen() 25992 MB/s Aug 13 07:08:31.244987 kernel: raid6: using algorithm avx2x2 gen() 30864 MB/s Aug 13 07:08:31.263044 kernel: raid6: .... xor() 19887 MB/s, rmw enabled Aug 13 07:08:31.263073 kernel: raid6: using avx2x2 recovery algorithm Aug 13 07:08:31.283947 kernel: xor: automatically using best checksumming function avx Aug 13 07:08:31.437953 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 07:08:31.451972 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:08:31.462062 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:08:31.475619 systemd-udevd[414]: Using default interface naming scheme 'v255'. Aug 13 07:08:31.480477 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:08:31.492062 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 07:08:31.505362 dracut-pre-trigger[420]: rd.md=0: removing MD RAID activation Aug 13 07:08:31.538485 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:08:31.553049 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:08:31.622522 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:08:31.633085 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 07:08:31.650492 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 07:08:31.653377 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:08:31.654636 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:08:31.655864 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:08:31.669944 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Aug 13 07:08:31.670398 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 07:08:31.674259 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 07:08:31.677970 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Aug 13 07:08:31.678971 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:08:31.679403 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:31.685047 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 07:08:31.685072 kernel: GPT:9289727 != 19775487 Aug 13 07:08:31.685082 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 07:08:31.685151 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:31.690047 kernel: GPT:9289727 != 19775487 Aug 13 07:08:31.690074 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 07:08:31.690085 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 07:08:31.688685 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:08:31.688865 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:31.691176 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:31.697951 kernel: libata version 3.00 loaded. Aug 13 07:08:31.700292 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:31.704459 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:08:31.711964 kernel: AVX2 version of gcm_enc/dec engaged. Aug 13 07:08:31.712011 kernel: AES CTR mode by8 optimization enabled Aug 13 07:08:31.712237 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:08:31.716398 kernel: ahci 0000:00:1f.2: version 3.0 Aug 13 07:08:31.716647 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Aug 13 07:08:31.716661 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Aug 13 07:08:31.716838 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Aug 13 07:08:31.712401 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:31.723968 kernel: scsi host0: ahci Aug 13 07:08:31.727823 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (471) Aug 13 07:08:31.730934 kernel: BTRFS: device fsid 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad devid 1 transid 37 /dev/vda3 scanned by (udev-worker) (461) Aug 13 07:08:31.732944 kernel: scsi host1: ahci Aug 13 07:08:31.733242 kernel: scsi host2: ahci Aug 13 07:08:31.733399 kernel: scsi host3: ahci Aug 13 07:08:31.734228 kernel: scsi host4: ahci Aug 13 07:08:31.739689 kernel: scsi host5: ahci Aug 13 07:08:31.739882 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Aug 13 07:08:31.739901 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Aug 13 07:08:31.739977 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Aug 13 07:08:31.739989 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Aug 13 07:08:31.740000 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Aug 13 07:08:31.740010 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Aug 13 07:08:31.740139 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Aug 13 07:08:31.751513 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Aug 13 07:08:31.758399 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Aug 13 07:08:31.764431 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Aug 13 07:08:31.767589 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Aug 13 07:08:31.785063 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 07:08:31.788007 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:31.793150 disk-uuid[555]: Primary Header is updated. Aug 13 07:08:31.793150 disk-uuid[555]: Secondary Entries is updated. Aug 13 07:08:31.793150 disk-uuid[555]: Secondary Header is updated. Aug 13 07:08:31.797445 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 07:08:31.800943 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 07:08:31.806233 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:31.817078 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:31.841046 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:32.052952 kernel: ata2: SATA link down (SStatus 0 SControl 300) Aug 13 07:08:32.053008 kernel: ata4: SATA link down (SStatus 0 SControl 300) Aug 13 07:08:32.053942 kernel: ata6: SATA link down (SStatus 0 SControl 300) Aug 13 07:08:32.053959 kernel: ata5: SATA link down (SStatus 0 SControl 300) Aug 13 07:08:32.054946 kernel: ata1: SATA link down (SStatus 0 SControl 300) Aug 13 07:08:32.055955 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Aug 13 07:08:32.057103 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Aug 13 07:08:32.057117 kernel: ata3.00: applying bridge limits Aug 13 07:08:32.057949 kernel: ata3.00: configured for UDMA/100 Aug 13 07:08:32.057962 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Aug 13 07:08:32.118441 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Aug 13 07:08:32.118681 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Aug 13 07:08:32.131943 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Aug 13 07:08:32.802528 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 07:08:32.802591 disk-uuid[559]: The operation has completed successfully. Aug 13 07:08:32.832787 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 07:08:32.832929 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 07:08:32.855057 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 07:08:32.858515 sh[598]: Success Aug 13 07:08:32.870949 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Aug 13 07:08:32.903762 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 07:08:32.914508 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 07:08:32.917105 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 07:08:32.932723 kernel: BTRFS info (device dm-0): first mount of filesystem 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad Aug 13 07:08:32.932760 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:08:32.932771 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 07:08:32.934415 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 07:08:32.934431 kernel: BTRFS info (device dm-0): using free space tree Aug 13 07:08:32.938899 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 07:08:32.941228 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 07:08:32.954062 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 07:08:32.956561 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 07:08:32.964704 kernel: BTRFS info (device vda6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:08:32.964731 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:08:32.964742 kernel: BTRFS info (device vda6): using free space tree Aug 13 07:08:32.967938 kernel: BTRFS info (device vda6): auto enabling async discard Aug 13 07:08:32.977318 systemd[1]: mnt-oem.mount: Deactivated successfully. Aug 13 07:08:32.979022 kernel: BTRFS info (device vda6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:08:32.988021 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 07:08:32.993098 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 07:08:33.055121 ignition[686]: Ignition 2.19.0 Aug 13 07:08:33.055134 ignition[686]: Stage: fetch-offline Aug 13 07:08:33.055173 ignition[686]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:33.055184 ignition[686]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 07:08:33.055290 ignition[686]: parsed url from cmdline: "" Aug 13 07:08:33.055294 ignition[686]: no config URL provided Aug 13 07:08:33.055299 ignition[686]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 07:08:33.055311 ignition[686]: no config at "/usr/lib/ignition/user.ign" Aug 13 07:08:33.055339 ignition[686]: op(1): [started] loading QEMU firmware config module Aug 13 07:08:33.055357 ignition[686]: op(1): executing: "modprobe" "qemu_fw_cfg" Aug 13 07:08:33.068110 ignition[686]: op(1): [finished] loading QEMU firmware config module Aug 13 07:08:33.080020 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:08:33.091056 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:08:33.109589 ignition[686]: parsing config with SHA512: 64808a788790b0383c8c0be75248c3ec6b3b9b5003dc30dc13fbe8218f3fc1d7edf9c374f5e547a08618da7c1edb8a3a4a4d6f284b0e734e7ca9659fe502a293 Aug 13 07:08:33.113460 systemd-networkd[787]: lo: Link UP Aug 13 07:08:33.113470 systemd-networkd[787]: lo: Gained carrier Aug 13 07:08:33.114068 unknown[686]: fetched base config from "system" Aug 13 07:08:33.115103 ignition[686]: fetch-offline: fetch-offline passed Aug 13 07:08:33.114086 unknown[686]: fetched user config from "qemu" Aug 13 07:08:33.115199 ignition[686]: Ignition finished successfully Aug 13 07:08:33.115346 systemd-networkd[787]: Enumeration completed Aug 13 07:08:33.115493 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:08:33.115883 systemd-networkd[787]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:33.115887 systemd-networkd[787]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:08:33.117702 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:08:33.117820 systemd-networkd[787]: eth0: Link UP Aug 13 07:08:33.117824 systemd-networkd[787]: eth0: Gained carrier Aug 13 07:08:33.117831 systemd-networkd[787]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:33.120237 systemd[1]: Reached target network.target - Network. Aug 13 07:08:33.121290 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Aug 13 07:08:33.128078 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 07:08:33.131970 systemd-networkd[787]: eth0: DHCPv4 address 10.0.0.69/16, gateway 10.0.0.1 acquired from 10.0.0.1 Aug 13 07:08:33.144129 ignition[790]: Ignition 2.19.0 Aug 13 07:08:33.144142 ignition[790]: Stage: kargs Aug 13 07:08:33.144305 ignition[790]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:33.144317 ignition[790]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 07:08:33.145110 ignition[790]: kargs: kargs passed Aug 13 07:08:33.145158 ignition[790]: Ignition finished successfully Aug 13 07:08:33.148228 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 07:08:33.156160 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 07:08:33.171251 ignition[799]: Ignition 2.19.0 Aug 13 07:08:33.171262 ignition[799]: Stage: disks Aug 13 07:08:33.171455 ignition[799]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:33.171468 ignition[799]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 07:08:33.172403 ignition[799]: disks: disks passed Aug 13 07:08:33.175092 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 07:08:33.172446 ignition[799]: Ignition finished successfully Aug 13 07:08:33.176443 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 07:08:33.178242 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 07:08:33.180151 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:08:33.180213 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:08:33.180535 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:08:33.192049 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 07:08:33.202837 systemd-resolved[247]: Detected conflict on linux IN A 10.0.0.69 Aug 13 07:08:33.202850 systemd-resolved[247]: Hostname conflict, changing published hostname from 'linux' to 'linux3'. Aug 13 07:08:33.205466 systemd-fsck[810]: ROOT: clean, 14/553520 files, 52654/553472 blocks Aug 13 07:08:33.211153 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 07:08:33.222993 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 07:08:33.310951 kernel: EXT4-fs (vda9): mounted filesystem 98cc0201-e9ec-4d2c-8a62-5b521bf9317d r/w with ordered data mode. Quota mode: none. Aug 13 07:08:33.311090 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 07:08:33.313412 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 07:08:33.326003 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:08:33.328459 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 07:08:33.330820 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Aug 13 07:08:33.330866 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 07:08:33.339998 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (819) Aug 13 07:08:33.340025 kernel: BTRFS info (device vda6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:08:33.340037 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:08:33.340048 kernel: BTRFS info (device vda6): using free space tree Aug 13 07:08:33.340060 kernel: BTRFS info (device vda6): auto enabling async discard Aug 13 07:08:33.332554 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:08:33.341477 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 07:08:33.344300 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:08:33.347910 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 07:08:33.383391 initrd-setup-root[843]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 07:08:33.389020 initrd-setup-root[850]: cut: /sysroot/etc/group: No such file or directory Aug 13 07:08:33.394066 initrd-setup-root[857]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 07:08:33.399262 initrd-setup-root[864]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 07:08:33.486169 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 07:08:33.496040 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 07:08:33.498576 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 07:08:33.506951 kernel: BTRFS info (device vda6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:08:33.522658 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 07:08:33.526969 ignition[933]: INFO : Ignition 2.19.0 Aug 13 07:08:33.526969 ignition[933]: INFO : Stage: mount Aug 13 07:08:33.528526 ignition[933]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:33.528526 ignition[933]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 07:08:33.528526 ignition[933]: INFO : mount: mount passed Aug 13 07:08:33.528526 ignition[933]: INFO : Ignition finished successfully Aug 13 07:08:33.534147 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 07:08:33.546068 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 07:08:33.932251 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 07:08:33.941176 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:08:33.948943 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (946) Aug 13 07:08:33.951040 kernel: BTRFS info (device vda6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:08:33.951062 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:08:33.951073 kernel: BTRFS info (device vda6): using free space tree Aug 13 07:08:33.953945 kernel: BTRFS info (device vda6): auto enabling async discard Aug 13 07:08:33.955347 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:08:33.984294 ignition[963]: INFO : Ignition 2.19.0 Aug 13 07:08:33.984294 ignition[963]: INFO : Stage: files Aug 13 07:08:33.985997 ignition[963]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:33.985997 ignition[963]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 07:08:33.985997 ignition[963]: DEBUG : files: compiled without relabeling support, skipping Aug 13 07:08:33.989380 ignition[963]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 07:08:33.989380 ignition[963]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 07:08:33.991911 ignition[963]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 07:08:33.993312 ignition[963]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 07:08:33.994961 unknown[963]: wrote ssh authorized keys file for user: core Aug 13 07:08:33.996022 ignition[963]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 07:08:33.998262 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Aug 13 07:08:34.000060 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Aug 13 07:08:34.042063 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 07:08:34.141125 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Aug 13 07:08:34.141125 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Aug 13 07:08:34.144807 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Aug 13 07:08:34.224173 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Aug 13 07:08:34.312495 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Aug 13 07:08:34.312495 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:08:34.316347 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Aug 13 07:08:34.566115 systemd-networkd[787]: eth0: Gained IPv6LL Aug 13 07:08:34.670035 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Aug 13 07:08:35.016480 ignition[963]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:08:35.016480 ignition[963]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Aug 13 07:08:35.020276 ignition[963]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Aug 13 07:08:35.045253 ignition[963]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Aug 13 07:08:35.050374 ignition[963]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Aug 13 07:08:35.051914 ignition[963]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Aug 13 07:08:35.051914 ignition[963]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Aug 13 07:08:35.051914 ignition[963]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 07:08:35.051914 ignition[963]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:08:35.051914 ignition[963]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:08:35.051914 ignition[963]: INFO : files: files passed Aug 13 07:08:35.051914 ignition[963]: INFO : Ignition finished successfully Aug 13 07:08:35.062956 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 07:08:35.075075 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 07:08:35.075904 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 07:08:35.083276 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 07:08:35.084333 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 07:08:35.086661 initrd-setup-root-after-ignition[990]: grep: /sysroot/oem/oem-release: No such file or directory Aug 13 07:08:35.089642 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:08:35.089642 initrd-setup-root-after-ignition[992]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:08:35.092646 initrd-setup-root-after-ignition[997]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:08:35.096795 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:08:35.097097 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 07:08:35.116114 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 07:08:35.147911 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 07:08:35.148099 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 07:08:35.151994 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 07:08:35.153894 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 07:08:35.154078 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 07:08:35.157349 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 07:08:35.180050 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:08:35.195093 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 07:08:35.205362 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:08:35.207787 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:08:35.210180 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 07:08:35.212005 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 07:08:35.213022 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:08:35.215535 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 07:08:35.216785 systemd[1]: Stopped target basic.target - Basic System. Aug 13 07:08:35.217703 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 07:08:35.220384 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:08:35.221476 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 07:08:35.221823 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 07:08:35.222314 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:08:35.222651 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 07:08:35.222982 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 07:08:35.223454 systemd[1]: Stopped target swap.target - Swaps. Aug 13 07:08:35.223759 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 07:08:35.223892 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:08:35.237186 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:08:35.237329 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:08:35.239317 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 07:08:35.239426 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:08:35.241473 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 07:08:35.241598 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 07:08:35.243828 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 07:08:35.243958 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:08:35.244381 systemd[1]: Stopped target paths.target - Path Units. Aug 13 07:08:35.244620 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 07:08:35.253994 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:08:35.256667 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 07:08:35.256803 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 07:08:35.258474 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 07:08:35.258585 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:08:35.260111 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 07:08:35.260205 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:08:35.261769 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 07:08:35.261884 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:08:35.263570 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 07:08:35.263677 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 07:08:35.275054 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 07:08:35.275142 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 07:08:35.275254 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:08:35.277664 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 07:08:35.278896 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 07:08:35.279105 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:08:35.279516 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 07:08:35.279628 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:08:35.285609 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 07:08:35.285728 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 07:08:35.304949 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 07:08:35.318895 ignition[1018]: INFO : Ignition 2.19.0 Aug 13 07:08:35.318895 ignition[1018]: INFO : Stage: umount Aug 13 07:08:35.320594 ignition[1018]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:35.320594 ignition[1018]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 07:08:35.320594 ignition[1018]: INFO : umount: umount passed Aug 13 07:08:35.320594 ignition[1018]: INFO : Ignition finished successfully Aug 13 07:08:35.324350 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 07:08:35.324520 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 07:08:35.325758 systemd[1]: Stopped target network.target - Network. Aug 13 07:08:35.327336 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 07:08:35.327394 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 07:08:35.329229 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 07:08:35.329282 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 07:08:35.331003 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 07:08:35.331055 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 07:08:35.331940 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 07:08:35.331992 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 07:08:35.333885 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 07:08:35.336725 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 07:08:35.339978 systemd-networkd[787]: eth0: DHCPv6 lease lost Aug 13 07:08:35.342442 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 07:08:35.342616 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 07:08:35.345140 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 07:08:35.345189 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:08:35.352026 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 07:08:35.352095 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 07:08:35.352151 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:08:35.354044 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:08:35.361556 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 07:08:35.361716 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 07:08:35.363652 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:08:35.363713 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:08:35.365672 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 07:08:35.365725 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 07:08:35.365831 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 07:08:35.365876 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:08:35.366519 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 07:08:35.366647 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 07:08:35.385668 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 07:08:35.385855 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:08:35.388042 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 07:08:35.388093 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 07:08:35.390044 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 07:08:35.390088 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:08:35.391997 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 07:08:35.392049 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:08:35.394083 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 07:08:35.394134 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 07:08:35.396011 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:08:35.396061 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:35.406068 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 07:08:35.407129 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 07:08:35.407187 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:08:35.409384 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:08:35.409436 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:35.414206 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 07:08:35.414320 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 07:08:35.540336 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 07:08:35.540486 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 07:08:35.542598 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 07:08:35.544380 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 07:08:35.544437 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 07:08:35.554126 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 07:08:35.561373 systemd[1]: Switching root. Aug 13 07:08:35.590718 systemd-journald[193]: Journal stopped Aug 13 07:08:37.228249 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). Aug 13 07:08:37.228356 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 07:08:37.228373 kernel: SELinux: policy capability open_perms=1 Aug 13 07:08:37.228385 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 07:08:37.228403 kernel: SELinux: policy capability always_check_network=0 Aug 13 07:08:37.228429 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 07:08:37.228445 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 07:08:37.228459 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 07:08:37.228472 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 07:08:37.228491 kernel: audit: type=1403 audit(1755068916.179:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 07:08:37.228510 systemd[1]: Successfully loaded SELinux policy in 39.892ms. Aug 13 07:08:37.228539 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.893ms. Aug 13 07:08:37.228558 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:08:37.228578 systemd[1]: Detected virtualization kvm. Aug 13 07:08:37.228590 systemd[1]: Detected architecture x86-64. Aug 13 07:08:37.228606 systemd[1]: Detected first boot. Aug 13 07:08:37.228618 systemd[1]: Initializing machine ID from VM UUID. Aug 13 07:08:37.228630 zram_generator::config[1065]: No configuration found. Aug 13 07:08:37.228653 systemd[1]: Populated /etc with preset unit settings. Aug 13 07:08:37.228668 systemd[1]: initrd-switch-root.service: Deactivated successfully. Aug 13 07:08:37.228680 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Aug 13 07:08:37.228697 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Aug 13 07:08:37.228711 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Aug 13 07:08:37.228726 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Aug 13 07:08:37.228738 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Aug 13 07:08:37.228750 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Aug 13 07:08:37.228763 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Aug 13 07:08:37.228778 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Aug 13 07:08:37.228790 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Aug 13 07:08:37.228806 systemd[1]: Created slice user.slice - User and Session Slice. Aug 13 07:08:37.228819 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:08:37.228831 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:08:37.228844 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Aug 13 07:08:37.228856 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Aug 13 07:08:37.228868 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Aug 13 07:08:37.228881 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:08:37.228905 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Aug 13 07:08:37.242856 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:08:37.242898 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Aug 13 07:08:37.242937 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Aug 13 07:08:37.242960 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Aug 13 07:08:37.242985 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Aug 13 07:08:37.243013 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:08:37.243056 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:08:37.243082 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:08:37.243108 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:08:37.243145 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Aug 13 07:08:37.243168 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Aug 13 07:08:37.243191 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:08:37.243214 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:08:37.243237 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:08:37.243257 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Aug 13 07:08:37.243273 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Aug 13 07:08:37.243297 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Aug 13 07:08:37.243321 systemd[1]: Mounting media.mount - External Media Directory... Aug 13 07:08:37.243356 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:37.243378 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Aug 13 07:08:37.243406 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Aug 13 07:08:37.243419 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Aug 13 07:08:37.243432 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 07:08:37.243445 systemd[1]: Reached target machines.target - Containers. Aug 13 07:08:37.243460 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Aug 13 07:08:37.243472 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:08:37.243498 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:08:37.243519 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Aug 13 07:08:37.243532 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:08:37.243546 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:08:37.243559 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:08:37.243571 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Aug 13 07:08:37.243583 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:08:37.243596 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 07:08:37.243608 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Aug 13 07:08:37.243629 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Aug 13 07:08:37.243641 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Aug 13 07:08:37.243653 systemd[1]: Stopped systemd-fsck-usr.service. Aug 13 07:08:37.243665 kernel: fuse: init (API version 7.39) Aug 13 07:08:37.243678 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:08:37.243691 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:08:37.243705 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 07:08:37.243718 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Aug 13 07:08:37.243729 kernel: loop: module loaded Aug 13 07:08:37.243747 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:08:37.243762 systemd[1]: verity-setup.service: Deactivated successfully. Aug 13 07:08:37.243776 systemd[1]: Stopped verity-setup.service. Aug 13 07:08:37.243792 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:37.243807 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Aug 13 07:08:37.243822 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Aug 13 07:08:37.243838 systemd[1]: Mounted media.mount - External Media Directory. Aug 13 07:08:37.243907 systemd-journald[1132]: Collecting audit messages is disabled. Aug 13 07:08:37.244481 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Aug 13 07:08:37.244496 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Aug 13 07:08:37.244516 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Aug 13 07:08:37.244529 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:08:37.244550 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 07:08:37.244563 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Aug 13 07:08:37.244576 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:08:37.244588 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:08:37.244601 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:08:37.244613 kernel: ACPI: bus type drm_connector registered Aug 13 07:08:37.244626 systemd-journald[1132]: Journal started Aug 13 07:08:37.244656 systemd-journald[1132]: Runtime Journal (/run/log/journal/e56dd7c6833b442fb27df894dda8b6a0) is 6.0M, max 48.3M, 42.2M free. Aug 13 07:08:37.244710 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:08:36.999766 systemd[1]: Queued start job for default target multi-user.target. Aug 13 07:08:37.020522 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Aug 13 07:08:37.021057 systemd[1]: systemd-journald.service: Deactivated successfully. Aug 13 07:08:37.249991 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:08:37.250818 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:08:37.251022 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:08:37.252517 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 07:08:37.252721 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Aug 13 07:08:37.254169 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:08:37.254350 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:08:37.255772 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:08:37.257377 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 07:08:37.259366 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Aug 13 07:08:37.265237 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Aug 13 07:08:37.280409 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 07:08:37.292099 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Aug 13 07:08:37.294684 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Aug 13 07:08:37.295809 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 07:08:37.295848 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:08:37.297856 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Aug 13 07:08:37.300901 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Aug 13 07:08:37.305153 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Aug 13 07:08:37.306792 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:08:37.309863 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Aug 13 07:08:37.315014 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Aug 13 07:08:37.316195 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:08:37.321035 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Aug 13 07:08:37.322395 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:08:37.323710 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:08:37.330215 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Aug 13 07:08:37.335807 systemd[1]: Starting systemd-sysusers.service - Create System Users... Aug 13 07:08:37.338739 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Aug 13 07:08:37.340589 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Aug 13 07:08:37.343453 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Aug 13 07:08:37.401742 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:08:37.414043 systemd-journald[1132]: Time spent on flushing to /var/log/journal/e56dd7c6833b442fb27df894dda8b6a0 is 41.499ms for 1002 entries. Aug 13 07:08:37.414043 systemd-journald[1132]: System Journal (/var/log/journal/e56dd7c6833b442fb27df894dda8b6a0) is 8.0M, max 195.6M, 187.6M free. Aug 13 07:08:37.469965 systemd-journald[1132]: Received client request to flush runtime journal. Aug 13 07:08:37.470018 kernel: loop0: detected capacity change from 0 to 140768 Aug 13 07:08:37.470033 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 07:08:37.419720 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Aug 13 07:08:37.421530 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Aug 13 07:08:37.423174 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:08:37.426882 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Aug 13 07:08:37.460441 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Aug 13 07:08:37.470111 udevadm[1187]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Aug 13 07:08:37.476397 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Aug 13 07:08:37.489897 systemd[1]: Finished systemd-sysusers.service - Create System Users. Aug 13 07:08:37.497089 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:08:37.499096 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 07:08:37.499705 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Aug 13 07:08:37.501118 kernel: loop1: detected capacity change from 0 to 142488 Aug 13 07:08:37.601113 kernel: loop2: detected capacity change from 0 to 229808 Aug 13 07:08:37.612998 systemd-tmpfiles[1197]: ACLs are not supported, ignoring. Aug 13 07:08:37.613017 systemd-tmpfiles[1197]: ACLs are not supported, ignoring. Aug 13 07:08:37.623559 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:08:37.673959 kernel: loop3: detected capacity change from 0 to 140768 Aug 13 07:08:37.689956 kernel: loop4: detected capacity change from 0 to 142488 Aug 13 07:08:37.702941 kernel: loop5: detected capacity change from 0 to 229808 Aug 13 07:08:37.712418 (sd-merge)[1203]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Aug 13 07:08:37.713113 (sd-merge)[1203]: Merged extensions into '/usr'. Aug 13 07:08:37.717594 systemd[1]: Reloading requested from client PID 1179 ('systemd-sysext') (unit systemd-sysext.service)... Aug 13 07:08:37.717611 systemd[1]: Reloading... Aug 13 07:08:37.824637 zram_generator::config[1238]: No configuration found. Aug 13 07:08:37.908165 ldconfig[1174]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 07:08:37.977936 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:08:38.039654 systemd[1]: Reloading finished in 321 ms. Aug 13 07:08:38.076156 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Aug 13 07:08:38.077716 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Aug 13 07:08:38.092194 systemd[1]: Starting ensure-sysext.service... Aug 13 07:08:38.094860 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:08:38.099671 systemd[1]: Reloading requested from client PID 1266 ('systemctl') (unit ensure-sysext.service)... Aug 13 07:08:38.099686 systemd[1]: Reloading... Aug 13 07:08:38.133203 systemd-tmpfiles[1267]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 07:08:38.133600 systemd-tmpfiles[1267]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Aug 13 07:08:38.134657 systemd-tmpfiles[1267]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 07:08:38.134997 systemd-tmpfiles[1267]: ACLs are not supported, ignoring. Aug 13 07:08:38.135078 systemd-tmpfiles[1267]: ACLs are not supported, ignoring. Aug 13 07:08:38.138825 systemd-tmpfiles[1267]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:08:38.139016 systemd-tmpfiles[1267]: Skipping /boot Aug 13 07:08:38.155488 systemd-tmpfiles[1267]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:08:38.155557 systemd-tmpfiles[1267]: Skipping /boot Aug 13 07:08:38.176938 zram_generator::config[1296]: No configuration found. Aug 13 07:08:38.317414 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:08:38.368428 systemd[1]: Reloading finished in 268 ms. Aug 13 07:08:38.390982 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Aug 13 07:08:38.405564 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:08:38.413607 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 07:08:38.416680 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Aug 13 07:08:38.419306 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Aug 13 07:08:38.424338 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:08:38.430137 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:08:38.433096 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Aug 13 07:08:38.440226 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Aug 13 07:08:38.445117 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:38.445418 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:08:38.447204 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:08:38.450085 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:08:38.452509 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:08:38.453756 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:08:38.453856 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:38.457843 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:38.459117 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:08:38.459338 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:08:38.459491 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:38.464386 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Aug 13 07:08:38.466536 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:08:38.466748 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:08:38.470664 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:38.471113 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:08:38.475424 augenrules[1359]: No rules Aug 13 07:08:38.477873 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:08:38.479697 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:08:38.479831 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:08:38.481885 systemd-udevd[1337]: Using default interface naming scheme 'v255'. Aug 13 07:08:38.482982 systemd[1]: Starting systemd-update-done.service - Update is Completed... Aug 13 07:08:38.484748 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:08:38.487224 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 07:08:38.489265 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Aug 13 07:08:38.491187 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:08:38.491376 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:08:38.493045 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:08:38.493228 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:08:38.494819 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:08:38.495017 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:08:38.500546 systemd[1]: Finished systemd-update-done.service - Update is Completed. Aug 13 07:08:38.505372 systemd[1]: Finished ensure-sysext.service. Aug 13 07:08:38.509686 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:08:38.521171 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Aug 13 07:08:38.522463 systemd[1]: Started systemd-userdbd.service - User Database Manager. Aug 13 07:08:38.523708 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:08:38.530299 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:08:38.535356 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Aug 13 07:08:38.540774 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 07:08:38.575960 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1391) Aug 13 07:08:38.692611 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Aug 13 07:08:38.694291 systemd[1]: Reached target time-set.target - System Time Set. Aug 13 07:08:38.695392 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Aug 13 07:08:38.712102 systemd-networkd[1382]: lo: Link UP Aug 13 07:08:38.712116 systemd-networkd[1382]: lo: Gained carrier Aug 13 07:08:38.714859 systemd-networkd[1382]: Enumeration completed Aug 13 07:08:38.715071 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:08:38.719362 systemd-networkd[1382]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:38.719374 systemd-networkd[1382]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:08:38.721906 systemd-networkd[1382]: eth0: Link UP Aug 13 07:08:38.721916 systemd-networkd[1382]: eth0: Gained carrier Aug 13 07:08:38.722793 systemd-resolved[1336]: Positive Trust Anchors: Aug 13 07:08:38.722810 systemd-resolved[1336]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:08:38.722841 systemd-resolved[1336]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:08:38.723217 systemd-networkd[1382]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:38.726173 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Aug 13 07:08:38.729421 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Aug 13 07:08:38.730287 systemd-resolved[1336]: Defaulting to hostname 'linux'. Aug 13 07:08:38.731654 systemd-networkd[1382]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:38.732783 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Aug 13 07:08:38.734061 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:08:38.736086 systemd[1]: Reached target network.target - Network. Aug 13 07:08:38.737025 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:08:38.741052 systemd-networkd[1382]: eth0: DHCPv4 address 10.0.0.69/16, gateway 10.0.0.1 acquired from 10.0.0.1 Aug 13 07:08:38.742674 systemd-timesyncd[1377]: Network configuration changed, trying to establish connection. Aug 13 07:08:39.675471 systemd-resolved[1336]: Clock change detected. Flushing caches. Aug 13 07:08:39.675553 systemd-timesyncd[1377]: Contacted time server 10.0.0.1:123 (10.0.0.1). Aug 13 07:08:39.675621 systemd-timesyncd[1377]: Initial clock synchronization to Wed 2025-08-13 07:08:39.675407 UTC. Aug 13 07:08:39.688220 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Aug 13 07:08:39.692875 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Aug 13 07:08:39.702314 kernel: ACPI: button: Power Button [PWRF] Aug 13 07:08:39.702358 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Aug 13 07:08:39.708434 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Aug 13 07:08:39.712311 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Aug 13 07:08:39.712517 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Aug 13 07:08:39.712701 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Aug 13 07:08:39.814082 kernel: mousedev: PS/2 mouse device common for all mice Aug 13 07:08:39.860195 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:39.912142 kernel: kvm_amd: TSC scaling supported Aug 13 07:08:39.912213 kernel: kvm_amd: Nested Virtualization enabled Aug 13 07:08:39.912228 kernel: kvm_amd: Nested Paging enabled Aug 13 07:08:39.912270 kernel: kvm_amd: LBR virtualization supported Aug 13 07:08:39.913273 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Aug 13 07:08:39.913296 kernel: kvm_amd: Virtual GIF supported Aug 13 07:08:39.938879 kernel: EDAC MC: Ver: 3.0.0 Aug 13 07:08:39.946162 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:39.971218 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Aug 13 07:08:39.987493 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Aug 13 07:08:39.995161 lvm[1427]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:08:40.028620 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Aug 13 07:08:40.030169 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:08:40.031261 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:08:40.032417 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Aug 13 07:08:40.033637 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Aug 13 07:08:40.035058 systemd[1]: Started logrotate.timer - Daily rotation of log files. Aug 13 07:08:40.036251 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Aug 13 07:08:40.037451 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Aug 13 07:08:40.038640 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 07:08:40.038672 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:08:40.039541 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:08:40.041349 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Aug 13 07:08:40.044354 systemd[1]: Starting docker.socket - Docker Socket for the API... Aug 13 07:08:40.058636 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Aug 13 07:08:40.061050 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Aug 13 07:08:40.062626 systemd[1]: Listening on docker.socket - Docker Socket for the API. Aug 13 07:08:40.063751 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:08:40.064687 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:08:40.065630 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:08:40.065656 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:08:40.066683 systemd[1]: Starting containerd.service - containerd container runtime... Aug 13 07:08:40.068786 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Aug 13 07:08:40.072993 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Aug 13 07:08:40.079473 lvm[1431]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:08:40.079356 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Aug 13 07:08:40.080543 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Aug 13 07:08:40.081787 jq[1434]: false Aug 13 07:08:40.083566 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Aug 13 07:08:40.089971 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Aug 13 07:08:40.093072 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Aug 13 07:08:40.097620 extend-filesystems[1435]: Found loop3 Aug 13 07:08:40.097620 extend-filesystems[1435]: Found loop4 Aug 13 07:08:40.097620 extend-filesystems[1435]: Found loop5 Aug 13 07:08:40.097620 extend-filesystems[1435]: Found sr0 Aug 13 07:08:40.097620 extend-filesystems[1435]: Found vda Aug 13 07:08:40.097620 extend-filesystems[1435]: Found vda1 Aug 13 07:08:40.097620 extend-filesystems[1435]: Found vda2 Aug 13 07:08:40.141890 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1379) Aug 13 07:08:40.141939 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Aug 13 07:08:40.141959 extend-filesystems[1435]: Found vda3 Aug 13 07:08:40.141959 extend-filesystems[1435]: Found usr Aug 13 07:08:40.141959 extend-filesystems[1435]: Found vda4 Aug 13 07:08:40.141959 extend-filesystems[1435]: Found vda6 Aug 13 07:08:40.141959 extend-filesystems[1435]: Found vda7 Aug 13 07:08:40.141959 extend-filesystems[1435]: Found vda9 Aug 13 07:08:40.141959 extend-filesystems[1435]: Checking size of /dev/vda9 Aug 13 07:08:40.141959 extend-filesystems[1435]: Resized partition /dev/vda9 Aug 13 07:08:40.188933 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Aug 13 07:08:40.104361 dbus-daemon[1433]: [system] SELinux support is enabled Aug 13 07:08:40.101007 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Aug 13 07:08:40.197226 extend-filesystems[1451]: resize2fs 1.47.1 (20-May-2024) Aug 13 07:08:40.197226 extend-filesystems[1451]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Aug 13 07:08:40.197226 extend-filesystems[1451]: old_desc_blocks = 1, new_desc_blocks = 1 Aug 13 07:08:40.197226 extend-filesystems[1451]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Aug 13 07:08:40.112012 systemd[1]: Starting systemd-logind.service - User Login Management... Aug 13 07:08:40.202239 extend-filesystems[1435]: Resized filesystem in /dev/vda9 Aug 13 07:08:40.113740 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 07:08:40.114805 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 07:08:40.203893 update_engine[1452]: I20250813 07:08:40.162562 1452 main.cc:92] Flatcar Update Engine starting Aug 13 07:08:40.203893 update_engine[1452]: I20250813 07:08:40.164068 1452 update_check_scheduler.cc:74] Next update check in 11m52s Aug 13 07:08:40.116552 systemd[1]: Starting update-engine.service - Update Engine... Aug 13 07:08:40.204282 jq[1453]: true Aug 13 07:08:40.128940 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Aug 13 07:08:40.132146 systemd[1]: Started dbus.service - D-Bus System Message Bus. Aug 13 07:08:40.146856 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Aug 13 07:08:40.204880 jq[1460]: true Aug 13 07:08:40.149049 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 07:08:40.149906 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Aug 13 07:08:40.150254 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 07:08:40.150473 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Aug 13 07:08:40.163278 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 07:08:40.163500 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Aug 13 07:08:40.174921 (ntainerd)[1461]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Aug 13 07:08:40.180363 systemd-logind[1450]: Watching system buttons on /dev/input/event1 (Power Button) Aug 13 07:08:40.180396 systemd-logind[1450]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Aug 13 07:08:40.181108 systemd-logind[1450]: New seat seat0. Aug 13 07:08:40.185252 systemd[1]: Started systemd-logind.service - User Login Management. Aug 13 07:08:40.186812 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 07:08:40.187169 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Aug 13 07:08:40.210908 systemd[1]: Started update-engine.service - Update Engine. Aug 13 07:08:40.219815 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 07:08:40.220130 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Aug 13 07:08:40.221943 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 07:08:40.222240 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Aug 13 07:08:40.247458 sshd_keygen[1458]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 07:08:40.268951 bash[1489]: Updated "/home/core/.ssh/authorized_keys" Aug 13 07:08:40.268388 systemd[1]: Started locksmithd.service - Cluster reboot manager. Aug 13 07:08:40.269272 tar[1459]: linux-amd64/LICENSE Aug 13 07:08:40.269272 tar[1459]: linux-amd64/helm Aug 13 07:08:40.274295 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Aug 13 07:08:40.282210 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Aug 13 07:08:40.302208 systemd[1]: Starting issuegen.service - Generate /run/issue... Aug 13 07:08:40.303465 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Aug 13 07:08:40.309441 locksmithd[1488]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 07:08:40.310813 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 07:08:40.311106 systemd[1]: Finished issuegen.service - Generate /run/issue. Aug 13 07:08:40.320321 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Aug 13 07:08:40.393111 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Aug 13 07:08:40.401137 systemd[1]: Started getty@tty1.service - Getty on tty1. Aug 13 07:08:40.403916 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Aug 13 07:08:40.405305 systemd[1]: Reached target getty.target - Login Prompts. Aug 13 07:08:40.720080 containerd[1461]: time="2025-08-13T07:08:40.719877179Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Aug 13 07:08:40.752181 containerd[1461]: time="2025-08-13T07:08:40.752121546Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.754786 containerd[1461]: time="2025-08-13T07:08:40.754744022Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.100-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:08:40.754786 containerd[1461]: time="2025-08-13T07:08:40.754774439Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 07:08:40.754878 containerd[1461]: time="2025-08-13T07:08:40.754794757Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 07:08:40.755150 containerd[1461]: time="2025-08-13T07:08:40.755118464Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Aug 13 07:08:40.755150 containerd[1461]: time="2025-08-13T07:08:40.755144833Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755251 containerd[1461]: time="2025-08-13T07:08:40.755227638Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755251 containerd[1461]: time="2025-08-13T07:08:40.755247335Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755541 containerd[1461]: time="2025-08-13T07:08:40.755519235Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755570 containerd[1461]: time="2025-08-13T07:08:40.755550794Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755590 containerd[1461]: time="2025-08-13T07:08:40.755568407Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755590 containerd[1461]: time="2025-08-13T07:08:40.755582704Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.755714 containerd[1461]: time="2025-08-13T07:08:40.755694834Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.756020 containerd[1461]: time="2025-08-13T07:08:40.755996279Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:08:40.756162 containerd[1461]: time="2025-08-13T07:08:40.756140770Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:08:40.756162 containerd[1461]: time="2025-08-13T07:08:40.756158543Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 07:08:40.756337 containerd[1461]: time="2025-08-13T07:08:40.756273138Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 07:08:40.756402 containerd[1461]: time="2025-08-13T07:08:40.756379738Z" level=info msg="metadata content store policy set" policy=shared Aug 13 07:08:40.762974 containerd[1461]: time="2025-08-13T07:08:40.762919688Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 07:08:40.762974 containerd[1461]: time="2025-08-13T07:08:40.762984971Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 07:08:40.763127 containerd[1461]: time="2025-08-13T07:08:40.763053990Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Aug 13 07:08:40.763127 containerd[1461]: time="2025-08-13T07:08:40.763098433Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Aug 13 07:08:40.763166 containerd[1461]: time="2025-08-13T07:08:40.763145932Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 07:08:40.763323 containerd[1461]: time="2025-08-13T07:08:40.763299711Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 07:08:40.763650 containerd[1461]: time="2025-08-13T07:08:40.763628818Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 07:08:40.763796 containerd[1461]: time="2025-08-13T07:08:40.763772647Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Aug 13 07:08:40.763796 containerd[1461]: time="2025-08-13T07:08:40.763792875Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Aug 13 07:08:40.763884 containerd[1461]: time="2025-08-13T07:08:40.763808174Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Aug 13 07:08:40.763884 containerd[1461]: time="2025-08-13T07:08:40.763821579Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763884 containerd[1461]: time="2025-08-13T07:08:40.763834323Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763884 containerd[1461]: time="2025-08-13T07:08:40.763865521Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763884 containerd[1461]: time="2025-08-13T07:08:40.763882343Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763985 containerd[1461]: time="2025-08-13T07:08:40.763896319Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763985 containerd[1461]: time="2025-08-13T07:08:40.763909384Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763985 containerd[1461]: time="2025-08-13T07:08:40.763925724Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763985 containerd[1461]: time="2025-08-13T07:08:40.763938538Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 07:08:40.763985 containerd[1461]: time="2025-08-13T07:08:40.763959988Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.763985 containerd[1461]: time="2025-08-13T07:08:40.763986578Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.763999613Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764012577Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764026974Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764039898Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764053504Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764066378Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764078390Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764093258Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764101 containerd[1461]: time="2025-08-13T07:08:40.764104419Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764116662Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764129697Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764146258Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764180442Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764193526Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764204837Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764280289Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764301098Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764312469Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764323730Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764348276Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764382701Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764397659Z" level=info msg="NRI interface is disabled by configuration." Aug 13 07:08:40.764458 containerd[1461]: time="2025-08-13T07:08:40.764408179Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 07:08:40.764960 containerd[1461]: time="2025-08-13T07:08:40.764731074Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 07:08:40.764960 containerd[1461]: time="2025-08-13T07:08:40.764825511Z" level=info msg="Connect containerd service" Aug 13 07:08:40.764960 containerd[1461]: time="2025-08-13T07:08:40.764915960Z" level=info msg="using legacy CRI server" Aug 13 07:08:40.764960 containerd[1461]: time="2025-08-13T07:08:40.764928394Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Aug 13 07:08:40.765248 containerd[1461]: time="2025-08-13T07:08:40.765063738Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 07:08:40.765962 containerd[1461]: time="2025-08-13T07:08:40.765935161Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:08:40.766254 containerd[1461]: time="2025-08-13T07:08:40.766169260Z" level=info msg="Start subscribing containerd event" Aug 13 07:08:40.766632 containerd[1461]: time="2025-08-13T07:08:40.766365808Z" level=info msg="Start recovering state" Aug 13 07:08:40.766632 containerd[1461]: time="2025-08-13T07:08:40.766485032Z" level=info msg="Start event monitor" Aug 13 07:08:40.766632 containerd[1461]: time="2025-08-13T07:08:40.766495071Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 07:08:40.766632 containerd[1461]: time="2025-08-13T07:08:40.766514177Z" level=info msg="Start snapshots syncer" Aug 13 07:08:40.766632 containerd[1461]: time="2025-08-13T07:08:40.766530327Z" level=info msg="Start cni network conf syncer for default" Aug 13 07:08:40.766632 containerd[1461]: time="2025-08-13T07:08:40.766542630Z" level=info msg="Start streaming server" Aug 13 07:08:40.767935 containerd[1461]: time="2025-08-13T07:08:40.767908541Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 07:08:40.768665 systemd[1]: Started containerd.service - containerd container runtime. Aug 13 07:08:40.770152 containerd[1461]: time="2025-08-13T07:08:40.768706817Z" level=info msg="containerd successfully booted in 0.050526s" Aug 13 07:08:40.955651 tar[1459]: linux-amd64/README.md Aug 13 07:08:40.976090 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Aug 13 07:08:41.450487 systemd-networkd[1382]: eth0: Gained IPv6LL Aug 13 07:08:41.454326 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Aug 13 07:08:41.456200 systemd[1]: Reached target network-online.target - Network is Online. Aug 13 07:08:41.469098 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Aug 13 07:08:41.471427 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:08:41.473649 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Aug 13 07:08:41.494713 systemd[1]: coreos-metadata.service: Deactivated successfully. Aug 13 07:08:41.494994 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Aug 13 07:08:41.496829 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Aug 13 07:08:41.499109 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Aug 13 07:08:42.741190 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:08:42.742957 systemd[1]: Reached target multi-user.target - Multi-User System. Aug 13 07:08:42.746957 systemd[1]: Startup finished in 1.176s (kernel) + 5.473s (initrd) + 5.673s (userspace) = 12.324s. Aug 13 07:08:42.751470 (kubelet)[1547]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:08:43.426075 kubelet[1547]: E0813 07:08:43.425992 1547 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:08:43.430842 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:08:43.431088 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:08:43.431474 systemd[1]: kubelet.service: Consumed 1.818s CPU time. Aug 13 07:08:44.170012 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Aug 13 07:08:44.171521 systemd[1]: Started sshd@0-10.0.0.69:22-10.0.0.1:39464.service - OpenSSH per-connection server daemon (10.0.0.1:39464). Aug 13 07:08:44.222088 sshd[1560]: Accepted publickey for core from 10.0.0.1 port 39464 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:44.224427 sshd[1560]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:44.233642 systemd-logind[1450]: New session 1 of user core. Aug 13 07:08:44.235019 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Aug 13 07:08:44.243084 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Aug 13 07:08:44.255406 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Aug 13 07:08:44.258240 systemd[1]: Starting user@500.service - User Manager for UID 500... Aug 13 07:08:44.267376 (systemd)[1564]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 07:08:44.380769 systemd[1564]: Queued start job for default target default.target. Aug 13 07:08:44.393210 systemd[1564]: Created slice app.slice - User Application Slice. Aug 13 07:08:44.393238 systemd[1564]: Reached target paths.target - Paths. Aug 13 07:08:44.393262 systemd[1564]: Reached target timers.target - Timers. Aug 13 07:08:44.395074 systemd[1564]: Starting dbus.socket - D-Bus User Message Bus Socket... Aug 13 07:08:44.408111 systemd[1564]: Listening on dbus.socket - D-Bus User Message Bus Socket. Aug 13 07:08:44.408279 systemd[1564]: Reached target sockets.target - Sockets. Aug 13 07:08:44.408298 systemd[1564]: Reached target basic.target - Basic System. Aug 13 07:08:44.408341 systemd[1564]: Reached target default.target - Main User Target. Aug 13 07:08:44.408380 systemd[1564]: Startup finished in 133ms. Aug 13 07:08:44.408784 systemd[1]: Started user@500.service - User Manager for UID 500. Aug 13 07:08:44.410558 systemd[1]: Started session-1.scope - Session 1 of User core. Aug 13 07:08:44.469899 systemd[1]: Started sshd@1-10.0.0.69:22-10.0.0.1:39478.service - OpenSSH per-connection server daemon (10.0.0.1:39478). Aug 13 07:08:44.507867 sshd[1575]: Accepted publickey for core from 10.0.0.1 port 39478 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:44.509480 sshd[1575]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:44.513693 systemd-logind[1450]: New session 2 of user core. Aug 13 07:08:44.523980 systemd[1]: Started session-2.scope - Session 2 of User core. Aug 13 07:08:44.579217 sshd[1575]: pam_unix(sshd:session): session closed for user core Aug 13 07:08:44.591740 systemd[1]: sshd@1-10.0.0.69:22-10.0.0.1:39478.service: Deactivated successfully. Aug 13 07:08:44.593514 systemd[1]: session-2.scope: Deactivated successfully. Aug 13 07:08:44.595054 systemd-logind[1450]: Session 2 logged out. Waiting for processes to exit. Aug 13 07:08:44.606118 systemd[1]: Started sshd@2-10.0.0.69:22-10.0.0.1:39488.service - OpenSSH per-connection server daemon (10.0.0.1:39488). Aug 13 07:08:44.607195 systemd-logind[1450]: Removed session 2. Aug 13 07:08:44.637084 sshd[1582]: Accepted publickey for core from 10.0.0.1 port 39488 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:44.638644 sshd[1582]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:44.643065 systemd-logind[1450]: New session 3 of user core. Aug 13 07:08:44.662065 systemd[1]: Started session-3.scope - Session 3 of User core. Aug 13 07:08:44.713699 sshd[1582]: pam_unix(sshd:session): session closed for user core Aug 13 07:08:44.725494 systemd[1]: sshd@2-10.0.0.69:22-10.0.0.1:39488.service: Deactivated successfully. Aug 13 07:08:44.727169 systemd[1]: session-3.scope: Deactivated successfully. Aug 13 07:08:44.728801 systemd-logind[1450]: Session 3 logged out. Waiting for processes to exit. Aug 13 07:08:44.746084 systemd[1]: Started sshd@3-10.0.0.69:22-10.0.0.1:39492.service - OpenSSH per-connection server daemon (10.0.0.1:39492). Aug 13 07:08:44.747072 systemd-logind[1450]: Removed session 3. Aug 13 07:08:44.777703 sshd[1589]: Accepted publickey for core from 10.0.0.1 port 39492 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:44.779331 sshd[1589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:44.783437 systemd-logind[1450]: New session 4 of user core. Aug 13 07:08:44.793982 systemd[1]: Started session-4.scope - Session 4 of User core. Aug 13 07:08:44.851164 sshd[1589]: pam_unix(sshd:session): session closed for user core Aug 13 07:08:44.871891 systemd[1]: sshd@3-10.0.0.69:22-10.0.0.1:39492.service: Deactivated successfully. Aug 13 07:08:44.873768 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 07:08:44.875450 systemd-logind[1450]: Session 4 logged out. Waiting for processes to exit. Aug 13 07:08:44.876818 systemd[1]: Started sshd@4-10.0.0.69:22-10.0.0.1:39494.service - OpenSSH per-connection server daemon (10.0.0.1:39494). Aug 13 07:08:44.877650 systemd-logind[1450]: Removed session 4. Aug 13 07:08:44.912792 sshd[1596]: Accepted publickey for core from 10.0.0.1 port 39494 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:44.914518 sshd[1596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:44.918936 systemd-logind[1450]: New session 5 of user core. Aug 13 07:08:44.927996 systemd[1]: Started session-5.scope - Session 5 of User core. Aug 13 07:08:44.989422 sudo[1599]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 07:08:44.989771 sudo[1599]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:08:45.009312 sudo[1599]: pam_unix(sudo:session): session closed for user root Aug 13 07:08:45.011731 sshd[1596]: pam_unix(sshd:session): session closed for user core Aug 13 07:08:45.025897 systemd[1]: sshd@4-10.0.0.69:22-10.0.0.1:39494.service: Deactivated successfully. Aug 13 07:08:45.027676 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 07:08:45.029311 systemd-logind[1450]: Session 5 logged out. Waiting for processes to exit. Aug 13 07:08:45.031423 systemd[1]: Started sshd@5-10.0.0.69:22-10.0.0.1:39506.service - OpenSSH per-connection server daemon (10.0.0.1:39506). Aug 13 07:08:45.032484 systemd-logind[1450]: Removed session 5. Aug 13 07:08:45.069815 sshd[1604]: Accepted publickey for core from 10.0.0.1 port 39506 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:45.071461 sshd[1604]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:45.075365 systemd-logind[1450]: New session 6 of user core. Aug 13 07:08:45.086980 systemd[1]: Started session-6.scope - Session 6 of User core. Aug 13 07:08:45.142162 sudo[1608]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 07:08:45.142520 sudo[1608]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:08:45.147188 sudo[1608]: pam_unix(sudo:session): session closed for user root Aug 13 07:08:45.154109 sudo[1607]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Aug 13 07:08:45.154464 sudo[1607]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:08:45.176115 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Aug 13 07:08:45.178372 auditctl[1611]: No rules Aug 13 07:08:45.179838 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 07:08:45.180114 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Aug 13 07:08:45.182101 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 07:08:45.217199 augenrules[1629]: No rules Aug 13 07:08:45.219309 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 07:08:45.220649 sudo[1607]: pam_unix(sudo:session): session closed for user root Aug 13 07:08:45.222772 sshd[1604]: pam_unix(sshd:session): session closed for user core Aug 13 07:08:45.235779 systemd[1]: sshd@5-10.0.0.69:22-10.0.0.1:39506.service: Deactivated successfully. Aug 13 07:08:45.237720 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 07:08:45.239462 systemd-logind[1450]: Session 6 logged out. Waiting for processes to exit. Aug 13 07:08:45.245276 systemd[1]: Started sshd@6-10.0.0.69:22-10.0.0.1:39514.service - OpenSSH per-connection server daemon (10.0.0.1:39514). Aug 13 07:08:45.246825 systemd-logind[1450]: Removed session 6. Aug 13 07:08:45.277210 sshd[1637]: Accepted publickey for core from 10.0.0.1 port 39514 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:08:45.278997 sshd[1637]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:08:45.283620 systemd-logind[1450]: New session 7 of user core. Aug 13 07:08:45.296984 systemd[1]: Started session-7.scope - Session 7 of User core. Aug 13 07:08:45.351470 sudo[1640]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 07:08:45.351960 sudo[1640]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:08:46.310166 systemd[1]: Starting docker.service - Docker Application Container Engine... Aug 13 07:08:46.310339 (dockerd)[1658]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Aug 13 07:08:46.954114 dockerd[1658]: time="2025-08-13T07:08:46.954026623Z" level=info msg="Starting up" Aug 13 07:08:47.514447 dockerd[1658]: time="2025-08-13T07:08:47.514367852Z" level=info msg="Loading containers: start." Aug 13 07:08:47.641878 kernel: Initializing XFRM netlink socket Aug 13 07:08:47.723602 systemd-networkd[1382]: docker0: Link UP Aug 13 07:08:47.745515 dockerd[1658]: time="2025-08-13T07:08:47.745461160Z" level=info msg="Loading containers: done." Aug 13 07:08:47.760094 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3750259743-merged.mount: Deactivated successfully. Aug 13 07:08:47.762939 dockerd[1658]: time="2025-08-13T07:08:47.762885254Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 07:08:47.763028 dockerd[1658]: time="2025-08-13T07:08:47.763000279Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Aug 13 07:08:47.763196 dockerd[1658]: time="2025-08-13T07:08:47.763154108Z" level=info msg="Daemon has completed initialization" Aug 13 07:08:47.802628 dockerd[1658]: time="2025-08-13T07:08:47.802431298Z" level=info msg="API listen on /run/docker.sock" Aug 13 07:08:47.802923 systemd[1]: Started docker.service - Docker Application Container Engine. Aug 13 07:08:48.820452 containerd[1461]: time="2025-08-13T07:08:48.820373005Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\"" Aug 13 07:08:49.347902 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1317501502.mount: Deactivated successfully. Aug 13 07:08:50.885617 containerd[1461]: time="2025-08-13T07:08:50.885504499Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:50.886136 containerd[1461]: time="2025-08-13T07:08:50.885950435Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.3: active requests=0, bytes read=30078237" Aug 13 07:08:50.887155 containerd[1461]: time="2025-08-13T07:08:50.887121831Z" level=info msg="ImageCreate event name:\"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:50.889999 containerd[1461]: time="2025-08-13T07:08:50.889925036Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:50.891248 containerd[1461]: time="2025-08-13T07:08:50.891201529Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.3\" with image id \"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\", size \"30075037\" in 2.070756258s" Aug 13 07:08:50.891291 containerd[1461]: time="2025-08-13T07:08:50.891269717Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\" returns image reference \"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\"" Aug 13 07:08:50.892274 containerd[1461]: time="2025-08-13T07:08:50.892242691Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\"" Aug 13 07:08:52.161334 containerd[1461]: time="2025-08-13T07:08:52.161264333Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:52.162169 containerd[1461]: time="2025-08-13T07:08:52.162105711Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.3: active requests=0, bytes read=26019361" Aug 13 07:08:52.163777 containerd[1461]: time="2025-08-13T07:08:52.163724445Z" level=info msg="ImageCreate event name:\"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:52.166894 containerd[1461]: time="2025-08-13T07:08:52.166808527Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:52.168636 containerd[1461]: time="2025-08-13T07:08:52.168563627Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.3\" with image id \"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\", size \"27646922\" in 1.276282404s" Aug 13 07:08:52.168636 containerd[1461]: time="2025-08-13T07:08:52.168627657Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\" returns image reference \"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\"" Aug 13 07:08:52.169342 containerd[1461]: time="2025-08-13T07:08:52.169303825Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\"" Aug 13 07:08:53.681561 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 07:08:53.744944 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:08:53.751867 containerd[1461]: time="2025-08-13T07:08:53.751805751Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:53.753000 containerd[1461]: time="2025-08-13T07:08:53.752880185Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.3: active requests=0, bytes read=20155013" Aug 13 07:08:53.754578 containerd[1461]: time="2025-08-13T07:08:53.754539846Z" level=info msg="ImageCreate event name:\"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:53.758036 containerd[1461]: time="2025-08-13T07:08:53.757995464Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:53.759493 containerd[1461]: time="2025-08-13T07:08:53.759427809Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.3\" with image id \"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\", size \"21782592\" in 1.590079391s" Aug 13 07:08:53.759493 containerd[1461]: time="2025-08-13T07:08:53.759478695Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\" returns image reference \"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\"" Aug 13 07:08:53.760045 containerd[1461]: time="2025-08-13T07:08:53.760004781Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\"" Aug 13 07:08:54.043943 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:08:54.049169 (kubelet)[1875]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:08:54.178717 kubelet[1875]: E0813 07:08:54.178638 1875 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:08:54.186934 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:08:54.187213 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:08:55.477739 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3929221632.mount: Deactivated successfully. Aug 13 07:08:56.345344 containerd[1461]: time="2025-08-13T07:08:56.345276452Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:56.346177 containerd[1461]: time="2025-08-13T07:08:56.346135582Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.3: active requests=0, bytes read=31892666" Aug 13 07:08:56.347412 containerd[1461]: time="2025-08-13T07:08:56.347378823Z" level=info msg="ImageCreate event name:\"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:56.349705 containerd[1461]: time="2025-08-13T07:08:56.349640423Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:56.350356 containerd[1461]: time="2025-08-13T07:08:56.350301532Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.3\" with image id \"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\", repo tag \"registry.k8s.io/kube-proxy:v1.33.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\", size \"31891685\" in 2.590251546s" Aug 13 07:08:56.350388 containerd[1461]: time="2025-08-13T07:08:56.350356395Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\" returns image reference \"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\"" Aug 13 07:08:56.351088 containerd[1461]: time="2025-08-13T07:08:56.351057038Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Aug 13 07:08:56.879410 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2081334370.mount: Deactivated successfully. Aug 13 07:08:58.225260 containerd[1461]: time="2025-08-13T07:08:58.225177314Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:58.226217 containerd[1461]: time="2025-08-13T07:08:58.226102850Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Aug 13 07:08:58.227544 containerd[1461]: time="2025-08-13T07:08:58.227507122Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:58.231371 containerd[1461]: time="2025-08-13T07:08:58.231330590Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:58.232817 containerd[1461]: time="2025-08-13T07:08:58.232760430Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.881666893s" Aug 13 07:08:58.232817 containerd[1461]: time="2025-08-13T07:08:58.232793692Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Aug 13 07:08:58.233386 containerd[1461]: time="2025-08-13T07:08:58.233315120Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 07:08:58.780649 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2842819766.mount: Deactivated successfully. Aug 13 07:08:58.786305 containerd[1461]: time="2025-08-13T07:08:58.786257189Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:58.787044 containerd[1461]: time="2025-08-13T07:08:58.786973341Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Aug 13 07:08:58.788108 containerd[1461]: time="2025-08-13T07:08:58.788061371Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:58.790338 containerd[1461]: time="2025-08-13T07:08:58.790302482Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:08:58.791056 containerd[1461]: time="2025-08-13T07:08:58.791016882Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 557.647781ms" Aug 13 07:08:58.791096 containerd[1461]: time="2025-08-13T07:08:58.791054963Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Aug 13 07:08:58.791598 containerd[1461]: time="2025-08-13T07:08:58.791572674Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Aug 13 07:08:59.367820 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1737509783.mount: Deactivated successfully. Aug 13 07:09:01.602869 containerd[1461]: time="2025-08-13T07:09:01.602798758Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:01.603582 containerd[1461]: time="2025-08-13T07:09:01.603495925Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58247175" Aug 13 07:09:01.604894 containerd[1461]: time="2025-08-13T07:09:01.604864711Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:01.608709 containerd[1461]: time="2025-08-13T07:09:01.608669554Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:01.609799 containerd[1461]: time="2025-08-13T07:09:01.609747054Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.818135788s" Aug 13 07:09:01.609799 containerd[1461]: time="2025-08-13T07:09:01.609792118Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Aug 13 07:09:04.437710 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 07:09:04.631472 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:04.822104 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:04.828533 (kubelet)[2036]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:04.878068 kubelet[2036]: E0813 07:09:04.877986 2036 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:04.882458 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:04.882687 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:05.100214 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:05.117258 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:05.144023 systemd[1]: Reloading requested from client PID 2052 ('systemctl') (unit session-7.scope)... Aug 13 07:09:05.144039 systemd[1]: Reloading... Aug 13 07:09:05.236876 zram_generator::config[2091]: No configuration found. Aug 13 07:09:06.017576 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:09:06.103704 systemd[1]: Reloading finished in 959 ms. Aug 13 07:09:06.156486 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:06.160700 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 07:09:06.160980 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:06.169264 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:06.335342 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:06.341809 (kubelet)[2141]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:09:06.427440 kubelet[2141]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:09:06.427440 kubelet[2141]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 07:09:06.427440 kubelet[2141]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:09:06.427939 kubelet[2141]: I0813 07:09:06.427499 2141 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:09:07.857806 kubelet[2141]: I0813 07:09:07.857751 2141 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 07:09:07.857806 kubelet[2141]: I0813 07:09:07.857786 2141 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:09:07.858301 kubelet[2141]: I0813 07:09:07.858077 2141 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 07:09:07.880044 kubelet[2141]: E0813 07:09:07.879998 2141 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.69:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Aug 13 07:09:07.880614 kubelet[2141]: I0813 07:09:07.880583 2141 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:09:07.886258 kubelet[2141]: E0813 07:09:07.886210 2141 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:09:07.886258 kubelet[2141]: I0813 07:09:07.886248 2141 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:09:07.892682 kubelet[2141]: I0813 07:09:07.892647 2141 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:09:07.893087 kubelet[2141]: I0813 07:09:07.893050 2141 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:09:07.893405 kubelet[2141]: I0813 07:09:07.893077 2141 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 07:09:07.893543 kubelet[2141]: I0813 07:09:07.893423 2141 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:09:07.893543 kubelet[2141]: I0813 07:09:07.893440 2141 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 07:09:07.893709 kubelet[2141]: I0813 07:09:07.893690 2141 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:09:07.895657 kubelet[2141]: I0813 07:09:07.895439 2141 kubelet.go:480] "Attempting to sync node with API server" Aug 13 07:09:07.895657 kubelet[2141]: I0813 07:09:07.895466 2141 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:09:07.895657 kubelet[2141]: I0813 07:09:07.895511 2141 kubelet.go:386] "Adding apiserver pod source" Aug 13 07:09:07.895657 kubelet[2141]: I0813 07:09:07.895538 2141 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:09:07.901472 kubelet[2141]: I0813 07:09:07.901362 2141 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 07:09:07.902438 kubelet[2141]: I0813 07:09:07.901952 2141 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 07:09:07.903198 kubelet[2141]: W0813 07:09:07.903165 2141 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 07:09:07.904036 kubelet[2141]: E0813 07:09:07.903918 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.69:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 07:09:07.904036 kubelet[2141]: E0813 07:09:07.903921 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.69:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Aug 13 07:09:07.906478 kubelet[2141]: I0813 07:09:07.906450 2141 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 07:09:07.906540 kubelet[2141]: I0813 07:09:07.906523 2141 server.go:1289] "Started kubelet" Aug 13 07:09:07.906782 kubelet[2141]: I0813 07:09:07.906729 2141 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:09:07.907496 kubelet[2141]: I0813 07:09:07.906841 2141 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:09:07.907496 kubelet[2141]: I0813 07:09:07.907303 2141 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:09:07.910041 kubelet[2141]: I0813 07:09:07.909198 2141 server.go:317] "Adding debug handlers to kubelet server" Aug 13 07:09:07.910041 kubelet[2141]: I0813 07:09:07.909861 2141 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:09:07.910430 kubelet[2141]: I0813 07:09:07.910407 2141 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:09:07.912802 kubelet[2141]: E0813 07:09:07.911791 2141 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.69:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.69:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.185b41e8cc48c006 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-08-13 07:09:07.906478086 +0000 UTC m=+1.559359644,LastTimestamp:2025-08-13 07:09:07.906478086 +0000 UTC m=+1.559359644,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Aug 13 07:09:07.913559 kubelet[2141]: E0813 07:09:07.912999 2141 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 07:09:07.913559 kubelet[2141]: I0813 07:09:07.913045 2141 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 07:09:07.913559 kubelet[2141]: I0813 07:09:07.913140 2141 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 07:09:07.913559 kubelet[2141]: I0813 07:09:07.913204 2141 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:09:07.913871 kubelet[2141]: E0813 07:09:07.913813 2141 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="200ms" Aug 13 07:09:07.914054 kubelet[2141]: E0813 07:09:07.914027 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.69:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Aug 13 07:09:07.914244 kubelet[2141]: I0813 07:09:07.914069 2141 factory.go:223] Registration of the systemd container factory successfully Aug 13 07:09:07.914441 kubelet[2141]: I0813 07:09:07.914406 2141 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:09:07.914526 kubelet[2141]: E0813 07:09:07.914451 2141 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:09:07.915414 kubelet[2141]: I0813 07:09:07.915385 2141 factory.go:223] Registration of the containerd container factory successfully Aug 13 07:09:07.930151 kubelet[2141]: I0813 07:09:07.929881 2141 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 07:09:07.930151 kubelet[2141]: I0813 07:09:07.929907 2141 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 07:09:07.930151 kubelet[2141]: I0813 07:09:07.929931 2141 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:09:07.932927 kubelet[2141]: I0813 07:09:07.932893 2141 policy_none.go:49] "None policy: Start" Aug 13 07:09:07.932971 kubelet[2141]: I0813 07:09:07.932939 2141 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 07:09:07.932971 kubelet[2141]: I0813 07:09:07.932960 2141 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:09:07.936723 kubelet[2141]: I0813 07:09:07.936659 2141 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 07:09:07.938985 kubelet[2141]: I0813 07:09:07.938235 2141 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 07:09:07.938985 kubelet[2141]: I0813 07:09:07.938287 2141 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 07:09:07.938985 kubelet[2141]: I0813 07:09:07.938320 2141 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 07:09:07.938985 kubelet[2141]: I0813 07:09:07.938337 2141 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 07:09:07.938985 kubelet[2141]: E0813 07:09:07.938400 2141 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:09:07.939267 kubelet[2141]: E0813 07:09:07.939235 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.69:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Aug 13 07:09:07.944443 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Aug 13 07:09:07.970527 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Aug 13 07:09:07.973862 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Aug 13 07:09:07.983785 kubelet[2141]: E0813 07:09:07.983745 2141 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 07:09:07.984028 kubelet[2141]: I0813 07:09:07.984002 2141 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:09:07.984125 kubelet[2141]: I0813 07:09:07.984023 2141 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:09:07.984594 kubelet[2141]: I0813 07:09:07.984286 2141 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:09:07.985144 kubelet[2141]: E0813 07:09:07.985120 2141 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 07:09:07.985216 kubelet[2141]: E0813 07:09:07.985177 2141 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Aug 13 07:09:08.049102 systemd[1]: Created slice kubepods-burstable-pod718653644aa13bbbaba997fad3f175cb.slice - libcontainer container kubepods-burstable-pod718653644aa13bbbaba997fad3f175cb.slice. Aug 13 07:09:08.060767 kubelet[2141]: E0813 07:09:08.060725 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:08.063105 systemd[1]: Created slice kubepods-burstable-podee495458985854145bfdfbfdfe0cc6b2.slice - libcontainer container kubepods-burstable-podee495458985854145bfdfbfdfe0cc6b2.slice. Aug 13 07:09:08.072255 kubelet[2141]: E0813 07:09:08.072215 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:08.075056 systemd[1]: Created slice kubepods-burstable-pod9f30683e4d57ebf2ca7dbf4704079d65.slice - libcontainer container kubepods-burstable-pod9f30683e4d57ebf2ca7dbf4704079d65.slice. Aug 13 07:09:08.076628 kubelet[2141]: E0813 07:09:08.076596 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:08.086526 kubelet[2141]: I0813 07:09:08.086503 2141 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Aug 13 07:09:08.086938 kubelet[2141]: E0813 07:09:08.086898 2141 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Aug 13 07:09:08.114562 kubelet[2141]: E0813 07:09:08.114422 2141 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="400ms" Aug 13 07:09:08.214987 kubelet[2141]: I0813 07:09:08.214916 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/718653644aa13bbbaba997fad3f175cb-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"718653644aa13bbbaba997fad3f175cb\") " pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:08.214987 kubelet[2141]: I0813 07:09:08.214967 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/718653644aa13bbbaba997fad3f175cb-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"718653644aa13bbbaba997fad3f175cb\") " pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:08.214987 kubelet[2141]: I0813 07:09:08.214986 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/718653644aa13bbbaba997fad3f175cb-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"718653644aa13bbbaba997fad3f175cb\") " pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:08.215143 kubelet[2141]: I0813 07:09:08.215011 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:08.215143 kubelet[2141]: I0813 07:09:08.215055 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:08.215143 kubelet[2141]: I0813 07:09:08.215073 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:08.215143 kubelet[2141]: I0813 07:09:08.215093 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:08.215143 kubelet[2141]: I0813 07:09:08.215113 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:08.215323 kubelet[2141]: I0813 07:09:08.215134 2141 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9f30683e4d57ebf2ca7dbf4704079d65-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"9f30683e4d57ebf2ca7dbf4704079d65\") " pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:08.289331 kubelet[2141]: I0813 07:09:08.289302 2141 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Aug 13 07:09:08.289632 kubelet[2141]: E0813 07:09:08.289591 2141 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Aug 13 07:09:08.362251 kubelet[2141]: E0813 07:09:08.362209 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:08.362974 containerd[1461]: time="2025-08-13T07:09:08.362933637Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:718653644aa13bbbaba997fad3f175cb,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:08.373311 kubelet[2141]: E0813 07:09:08.373192 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:08.373942 containerd[1461]: time="2025-08-13T07:09:08.373880829Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:ee495458985854145bfdfbfdfe0cc6b2,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:08.377254 kubelet[2141]: E0813 07:09:08.377231 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:08.377691 containerd[1461]: time="2025-08-13T07:09:08.377641298Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:9f30683e4d57ebf2ca7dbf4704079d65,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:08.515273 kubelet[2141]: E0813 07:09:08.515225 2141 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="800ms" Aug 13 07:09:08.691296 kubelet[2141]: I0813 07:09:08.691162 2141 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Aug 13 07:09:08.691616 kubelet[2141]: E0813 07:09:08.691562 2141 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Aug 13 07:09:08.864684 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2228710663.mount: Deactivated successfully. Aug 13 07:09:08.875164 containerd[1461]: time="2025-08-13T07:09:08.875118991Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:09:08.877144 containerd[1461]: time="2025-08-13T07:09:08.877068777Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:09:08.878263 containerd[1461]: time="2025-08-13T07:09:08.878222479Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:09:08.879270 containerd[1461]: time="2025-08-13T07:09:08.879216433Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:09:08.880344 containerd[1461]: time="2025-08-13T07:09:08.880295335Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:09:08.881204 containerd[1461]: time="2025-08-13T07:09:08.881173131Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:09:08.882257 containerd[1461]: time="2025-08-13T07:09:08.882215044Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Aug 13 07:09:08.884690 containerd[1461]: time="2025-08-13T07:09:08.884654798Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:09:08.886255 containerd[1461]: time="2025-08-13T07:09:08.886220252Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 508.510195ms" Aug 13 07:09:08.886978 containerd[1461]: time="2025-08-13T07:09:08.886950591Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 512.959125ms" Aug 13 07:09:08.887685 containerd[1461]: time="2025-08-13T07:09:08.887648991Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 524.624464ms" Aug 13 07:09:08.931616 kubelet[2141]: E0813 07:09:08.931568 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.69:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Aug 13 07:09:09.274818 kubelet[2141]: E0813 07:09:09.274651 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.69:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Aug 13 07:09:09.284051 kubelet[2141]: E0813 07:09:09.283973 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.69:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Aug 13 07:09:09.316249 kubelet[2141]: E0813 07:09:09.316155 2141 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.69:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.69:6443: connect: connection refused" interval="1.6s" Aug 13 07:09:09.403723 kubelet[2141]: E0813 07:09:09.403657 2141 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.69:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.69:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 07:09:09.493584 kubelet[2141]: I0813 07:09:09.493540 2141 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Aug 13 07:09:09.494034 kubelet[2141]: E0813 07:09:09.493974 2141 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.69:6443/api/v1/nodes\": dial tcp 10.0.0.69:6443: connect: connection refused" node="localhost" Aug 13 07:09:09.510058 containerd[1461]: time="2025-08-13T07:09:09.508928979Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:09.510058 containerd[1461]: time="2025-08-13T07:09:09.508997427Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:09.510058 containerd[1461]: time="2025-08-13T07:09:09.509012816Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:09.510058 containerd[1461]: time="2025-08-13T07:09:09.509120017Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:09.512191 containerd[1461]: time="2025-08-13T07:09:09.512083933Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:09.512191 containerd[1461]: time="2025-08-13T07:09:09.512137644Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:09.512191 containerd[1461]: time="2025-08-13T07:09:09.512148214Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:09.512293 containerd[1461]: time="2025-08-13T07:09:09.512223645Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:09.513021 containerd[1461]: time="2025-08-13T07:09:09.512948404Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:09.513091 containerd[1461]: time="2025-08-13T07:09:09.513002165Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:09.513091 containerd[1461]: time="2025-08-13T07:09:09.513015880Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:09.513151 containerd[1461]: time="2025-08-13T07:09:09.513100218Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:09.550038 systemd[1]: Started cri-containerd-18e670ce3c09ef85aaa075b8c3ca2e579fea479d75a957844de6c4d659cf218b.scope - libcontainer container 18e670ce3c09ef85aaa075b8c3ca2e579fea479d75a957844de6c4d659cf218b. Aug 13 07:09:09.552286 systemd[1]: Started cri-containerd-254a1fc9a6222594b345b1f4d28f14e1955b15a530099abbf131bbe723deeaf0.scope - libcontainer container 254a1fc9a6222594b345b1f4d28f14e1955b15a530099abbf131bbe723deeaf0. Aug 13 07:09:09.556952 systemd[1]: Started cri-containerd-75f8e5d2c6ccc5cb57378eeb1aaf810e3f57b44f54e032b2ecca33c54137f4b9.scope - libcontainer container 75f8e5d2c6ccc5cb57378eeb1aaf810e3f57b44f54e032b2ecca33c54137f4b9. Aug 13 07:09:09.604765 containerd[1461]: time="2025-08-13T07:09:09.604723690Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:9f30683e4d57ebf2ca7dbf4704079d65,Namespace:kube-system,Attempt:0,} returns sandbox id \"18e670ce3c09ef85aaa075b8c3ca2e579fea479d75a957844de6c4d659cf218b\"" Aug 13 07:09:09.606505 kubelet[2141]: E0813 07:09:09.606471 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:09.608016 containerd[1461]: time="2025-08-13T07:09:09.607977640Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:718653644aa13bbbaba997fad3f175cb,Namespace:kube-system,Attempt:0,} returns sandbox id \"75f8e5d2c6ccc5cb57378eeb1aaf810e3f57b44f54e032b2ecca33c54137f4b9\"" Aug 13 07:09:09.609071 kubelet[2141]: E0813 07:09:09.608954 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:09.612265 containerd[1461]: time="2025-08-13T07:09:09.612217077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:ee495458985854145bfdfbfdfe0cc6b2,Namespace:kube-system,Attempt:0,} returns sandbox id \"254a1fc9a6222594b345b1f4d28f14e1955b15a530099abbf131bbe723deeaf0\"" Aug 13 07:09:09.614164 kubelet[2141]: E0813 07:09:09.613526 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:09.618085 containerd[1461]: time="2025-08-13T07:09:09.617933804Z" level=info msg="CreateContainer within sandbox \"18e670ce3c09ef85aaa075b8c3ca2e579fea479d75a957844de6c4d659cf218b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 07:09:09.620572 containerd[1461]: time="2025-08-13T07:09:09.620439461Z" level=info msg="CreateContainer within sandbox \"75f8e5d2c6ccc5cb57378eeb1aaf810e3f57b44f54e032b2ecca33c54137f4b9\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 07:09:09.622780 containerd[1461]: time="2025-08-13T07:09:09.622651959Z" level=info msg="CreateContainer within sandbox \"254a1fc9a6222594b345b1f4d28f14e1955b15a530099abbf131bbe723deeaf0\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 07:09:09.639667 containerd[1461]: time="2025-08-13T07:09:09.639615179Z" level=info msg="CreateContainer within sandbox \"18e670ce3c09ef85aaa075b8c3ca2e579fea479d75a957844de6c4d659cf218b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"43cc09fd8f262f98cdb1aa2c41471b4160f947936e6080b67fee82bc2280a0fe\"" Aug 13 07:09:09.640226 containerd[1461]: time="2025-08-13T07:09:09.640190708Z" level=info msg="StartContainer for \"43cc09fd8f262f98cdb1aa2c41471b4160f947936e6080b67fee82bc2280a0fe\"" Aug 13 07:09:09.643505 containerd[1461]: time="2025-08-13T07:09:09.643457331Z" level=info msg="CreateContainer within sandbox \"75f8e5d2c6ccc5cb57378eeb1aaf810e3f57b44f54e032b2ecca33c54137f4b9\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"6a1e5edb3d1eef82e35676a7b23491777727dc05b1fb8112e4056dc337bfc7d5\"" Aug 13 07:09:09.644245 containerd[1461]: time="2025-08-13T07:09:09.644203811Z" level=info msg="StartContainer for \"6a1e5edb3d1eef82e35676a7b23491777727dc05b1fb8112e4056dc337bfc7d5\"" Aug 13 07:09:09.646941 containerd[1461]: time="2025-08-13T07:09:09.646910405Z" level=info msg="CreateContainer within sandbox \"254a1fc9a6222594b345b1f4d28f14e1955b15a530099abbf131bbe723deeaf0\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"a5034dd00acc527e0840f210e690dde5bfc444bac27cfba9f03383859a5a552b\"" Aug 13 07:09:09.647234 containerd[1461]: time="2025-08-13T07:09:09.647203013Z" level=info msg="StartContainer for \"a5034dd00acc527e0840f210e690dde5bfc444bac27cfba9f03383859a5a552b\"" Aug 13 07:09:09.667029 systemd[1]: Started cri-containerd-43cc09fd8f262f98cdb1aa2c41471b4160f947936e6080b67fee82bc2280a0fe.scope - libcontainer container 43cc09fd8f262f98cdb1aa2c41471b4160f947936e6080b67fee82bc2280a0fe. Aug 13 07:09:09.670305 systemd[1]: Started cri-containerd-6a1e5edb3d1eef82e35676a7b23491777727dc05b1fb8112e4056dc337bfc7d5.scope - libcontainer container 6a1e5edb3d1eef82e35676a7b23491777727dc05b1fb8112e4056dc337bfc7d5. Aug 13 07:09:09.682201 systemd[1]: Started cri-containerd-a5034dd00acc527e0840f210e690dde5bfc444bac27cfba9f03383859a5a552b.scope - libcontainer container a5034dd00acc527e0840f210e690dde5bfc444bac27cfba9f03383859a5a552b. Aug 13 07:09:09.726255 containerd[1461]: time="2025-08-13T07:09:09.726090138Z" level=info msg="StartContainer for \"6a1e5edb3d1eef82e35676a7b23491777727dc05b1fb8112e4056dc337bfc7d5\" returns successfully" Aug 13 07:09:09.726255 containerd[1461]: time="2025-08-13T07:09:09.726182822Z" level=info msg="StartContainer for \"43cc09fd8f262f98cdb1aa2c41471b4160f947936e6080b67fee82bc2280a0fe\" returns successfully" Aug 13 07:09:09.748116 containerd[1461]: time="2025-08-13T07:09:09.748064022Z" level=info msg="StartContainer for \"a5034dd00acc527e0840f210e690dde5bfc444bac27cfba9f03383859a5a552b\" returns successfully" Aug 13 07:09:09.949162 kubelet[2141]: E0813 07:09:09.948307 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:09.949162 kubelet[2141]: E0813 07:09:09.948447 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:09.951746 kubelet[2141]: E0813 07:09:09.951015 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:09.951746 kubelet[2141]: E0813 07:09:09.951098 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:09.953326 kubelet[2141]: E0813 07:09:09.953190 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:09.953326 kubelet[2141]: E0813 07:09:09.953272 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:10.955828 kubelet[2141]: E0813 07:09:10.955632 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:10.955828 kubelet[2141]: E0813 07:09:10.955668 2141 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Aug 13 07:09:10.955828 kubelet[2141]: E0813 07:09:10.955781 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:10.955828 kubelet[2141]: E0813 07:09:10.955781 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:11.176027 kubelet[2141]: I0813 07:09:11.175980 2141 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Aug 13 07:09:11.316981 kubelet[2141]: E0813 07:09:11.315732 2141 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Aug 13 07:09:11.511542 kubelet[2141]: I0813 07:09:11.510987 2141 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Aug 13 07:09:11.511542 kubelet[2141]: E0813 07:09:11.511038 2141 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Aug 13 07:09:11.521076 kubelet[2141]: E0813 07:09:11.521031 2141 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 07:09:11.621588 kubelet[2141]: E0813 07:09:11.621412 2141 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 07:09:11.722143 kubelet[2141]: E0813 07:09:11.722093 2141 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 07:09:11.822426 kubelet[2141]: E0813 07:09:11.822369 2141 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 07:09:11.898462 kubelet[2141]: I0813 07:09:11.898254 2141 apiserver.go:52] "Watching apiserver" Aug 13 07:09:11.913227 kubelet[2141]: I0813 07:09:11.913185 2141 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 07:09:11.913907 kubelet[2141]: I0813 07:09:11.913841 2141 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:11.919327 kubelet[2141]: E0813 07:09:11.919279 2141 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:11.919327 kubelet[2141]: I0813 07:09:11.919304 2141 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:11.920780 kubelet[2141]: E0813 07:09:11.920756 2141 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:11.920780 kubelet[2141]: I0813 07:09:11.920774 2141 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:11.921935 kubelet[2141]: E0813 07:09:11.921901 2141 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:11.955985 kubelet[2141]: I0813 07:09:11.955953 2141 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:11.957535 kubelet[2141]: E0813 07:09:11.957505 2141 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:11.957660 kubelet[2141]: E0813 07:09:11.957642 2141 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:13.293665 systemd[1]: Reloading requested from client PID 2435 ('systemctl') (unit session-7.scope)... Aug 13 07:09:13.293690 systemd[1]: Reloading... Aug 13 07:09:13.372884 zram_generator::config[2480]: No configuration found. Aug 13 07:09:13.477694 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:09:13.570678 systemd[1]: Reloading finished in 276 ms. Aug 13 07:09:13.623142 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:13.653822 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 07:09:13.654271 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:13.654358 systemd[1]: kubelet.service: Consumed 1.000s CPU time, 132.2M memory peak, 0B memory swap peak. Aug 13 07:09:13.666142 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:13.876337 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:13.889241 (kubelet)[2519]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:09:13.931955 kubelet[2519]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:09:13.931955 kubelet[2519]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 07:09:13.931955 kubelet[2519]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:09:13.932400 kubelet[2519]: I0813 07:09:13.932002 2519 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:09:13.939383 kubelet[2519]: I0813 07:09:13.939341 2519 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 07:09:13.939383 kubelet[2519]: I0813 07:09:13.939365 2519 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:09:13.939566 kubelet[2519]: I0813 07:09:13.939541 2519 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 07:09:13.940649 kubelet[2519]: I0813 07:09:13.940622 2519 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Aug 13 07:09:13.942611 kubelet[2519]: I0813 07:09:13.942583 2519 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:09:13.947199 kubelet[2519]: E0813 07:09:13.947159 2519 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:09:13.947199 kubelet[2519]: I0813 07:09:13.947188 2519 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:09:13.954621 kubelet[2519]: I0813 07:09:13.954566 2519 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:09:13.954858 kubelet[2519]: I0813 07:09:13.954809 2519 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:09:13.955008 kubelet[2519]: I0813 07:09:13.954834 2519 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 07:09:13.955203 kubelet[2519]: I0813 07:09:13.955015 2519 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:09:13.955203 kubelet[2519]: I0813 07:09:13.955027 2519 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 07:09:13.955203 kubelet[2519]: I0813 07:09:13.955093 2519 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:09:13.955313 kubelet[2519]: I0813 07:09:13.955261 2519 kubelet.go:480] "Attempting to sync node with API server" Aug 13 07:09:13.955313 kubelet[2519]: I0813 07:09:13.955277 2519 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:09:13.955381 kubelet[2519]: I0813 07:09:13.955331 2519 kubelet.go:386] "Adding apiserver pod source" Aug 13 07:09:13.955381 kubelet[2519]: I0813 07:09:13.955368 2519 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:09:13.956563 kubelet[2519]: I0813 07:09:13.956536 2519 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 07:09:13.957537 kubelet[2519]: I0813 07:09:13.957338 2519 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 07:09:13.962354 kubelet[2519]: I0813 07:09:13.962327 2519 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 07:09:13.962487 kubelet[2519]: I0813 07:09:13.962391 2519 server.go:1289] "Started kubelet" Aug 13 07:09:13.966631 kubelet[2519]: I0813 07:09:13.965656 2519 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:09:13.966631 kubelet[2519]: I0813 07:09:13.965808 2519 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:09:13.966631 kubelet[2519]: I0813 07:09:13.966108 2519 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:09:13.967240 kubelet[2519]: I0813 07:09:13.967204 2519 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:09:13.968814 kubelet[2519]: I0813 07:09:13.967207 2519 server.go:317] "Adding debug handlers to kubelet server" Aug 13 07:09:13.969588 kubelet[2519]: I0813 07:09:13.969565 2519 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:09:13.970871 kubelet[2519]: I0813 07:09:13.970830 2519 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 07:09:13.972568 kubelet[2519]: I0813 07:09:13.972539 2519 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 07:09:13.972958 kubelet[2519]: I0813 07:09:13.972704 2519 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:09:13.976836 kubelet[2519]: E0813 07:09:13.976333 2519 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:09:13.976836 kubelet[2519]: I0813 07:09:13.976654 2519 factory.go:223] Registration of the containerd container factory successfully Aug 13 07:09:13.976836 kubelet[2519]: I0813 07:09:13.976671 2519 factory.go:223] Registration of the systemd container factory successfully Aug 13 07:09:13.976836 kubelet[2519]: I0813 07:09:13.976791 2519 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:09:13.991984 kubelet[2519]: I0813 07:09:13.991931 2519 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 07:09:13.994132 kubelet[2519]: I0813 07:09:13.994112 2519 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 07:09:13.994278 kubelet[2519]: I0813 07:09:13.994262 2519 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 07:09:13.994381 kubelet[2519]: I0813 07:09:13.994363 2519 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 07:09:13.994508 kubelet[2519]: I0813 07:09:13.994492 2519 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 07:09:13.994658 kubelet[2519]: E0813 07:09:13.994635 2519 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:09:14.017596 kubelet[2519]: I0813 07:09:14.017559 2519 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 07:09:14.017596 kubelet[2519]: I0813 07:09:14.017578 2519 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 07:09:14.017596 kubelet[2519]: I0813 07:09:14.017602 2519 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:09:14.017841 kubelet[2519]: I0813 07:09:14.017757 2519 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 07:09:14.017841 kubelet[2519]: I0813 07:09:14.017772 2519 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 07:09:14.017841 kubelet[2519]: I0813 07:09:14.017794 2519 policy_none.go:49] "None policy: Start" Aug 13 07:09:14.017841 kubelet[2519]: I0813 07:09:14.017812 2519 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 07:09:14.017841 kubelet[2519]: I0813 07:09:14.017825 2519 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:09:14.018038 kubelet[2519]: I0813 07:09:14.017937 2519 state_mem.go:75] "Updated machine memory state" Aug 13 07:09:14.022421 kubelet[2519]: E0813 07:09:14.022392 2519 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 07:09:14.022724 kubelet[2519]: I0813 07:09:14.022596 2519 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:09:14.022724 kubelet[2519]: I0813 07:09:14.022617 2519 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:09:14.022874 kubelet[2519]: I0813 07:09:14.022837 2519 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:09:14.024287 kubelet[2519]: E0813 07:09:14.024253 2519 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 07:09:14.096912 kubelet[2519]: I0813 07:09:14.096802 2519 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:14.096912 kubelet[2519]: I0813 07:09:14.096927 2519 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:14.099878 kubelet[2519]: I0813 07:09:14.097896 2519 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:14.130612 kubelet[2519]: I0813 07:09:14.130451 2519 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Aug 13 07:09:14.136573 kubelet[2519]: I0813 07:09:14.136527 2519 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Aug 13 07:09:14.136685 kubelet[2519]: I0813 07:09:14.136623 2519 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Aug 13 07:09:14.274281 kubelet[2519]: I0813 07:09:14.274215 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/718653644aa13bbbaba997fad3f175cb-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"718653644aa13bbbaba997fad3f175cb\") " pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:14.274281 kubelet[2519]: I0813 07:09:14.274260 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/718653644aa13bbbaba997fad3f175cb-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"718653644aa13bbbaba997fad3f175cb\") " pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:14.274281 kubelet[2519]: I0813 07:09:14.274298 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:14.274531 kubelet[2519]: I0813 07:09:14.274320 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:14.274531 kubelet[2519]: I0813 07:09:14.274338 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:14.274531 kubelet[2519]: I0813 07:09:14.274354 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:14.274531 kubelet[2519]: I0813 07:09:14.274420 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9f30683e4d57ebf2ca7dbf4704079d65-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"9f30683e4d57ebf2ca7dbf4704079d65\") " pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:14.274531 kubelet[2519]: I0813 07:09:14.274474 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/718653644aa13bbbaba997fad3f175cb-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"718653644aa13bbbaba997fad3f175cb\") " pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:14.274728 kubelet[2519]: I0813 07:09:14.274513 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ee495458985854145bfdfbfdfe0cc6b2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ee495458985854145bfdfbfdfe0cc6b2\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:14.294318 sudo[2560]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Aug 13 07:09:14.294802 sudo[2560]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Aug 13 07:09:14.406183 kubelet[2519]: E0813 07:09:14.406012 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:14.406314 kubelet[2519]: E0813 07:09:14.406274 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:14.406450 kubelet[2519]: E0813 07:09:14.406393 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:14.774196 sudo[2560]: pam_unix(sudo:session): session closed for user root Aug 13 07:09:14.956427 kubelet[2519]: I0813 07:09:14.956353 2519 apiserver.go:52] "Watching apiserver" Aug 13 07:09:14.973696 kubelet[2519]: I0813 07:09:14.973667 2519 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 07:09:15.005338 kubelet[2519]: I0813 07:09:15.005299 2519 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:15.005428 kubelet[2519]: I0813 07:09:15.005398 2519 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:15.005579 kubelet[2519]: I0813 07:09:15.005541 2519 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:15.131668 kubelet[2519]: E0813 07:09:15.131490 2519 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Aug 13 07:09:15.131890 kubelet[2519]: E0813 07:09:15.131691 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:15.131890 kubelet[2519]: E0813 07:09:15.131813 2519 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Aug 13 07:09:15.131978 kubelet[2519]: E0813 07:09:15.131909 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:15.132340 kubelet[2519]: E0813 07:09:15.132007 2519 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Aug 13 07:09:15.132340 kubelet[2519]: E0813 07:09:15.132138 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:15.140834 kubelet[2519]: I0813 07:09:15.140760 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.14071883 podStartE2EDuration="1.14071883s" podCreationTimestamp="2025-08-13 07:09:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:09:15.132223629 +0000 UTC m=+1.237913161" watchObservedRunningTime="2025-08-13 07:09:15.14071883 +0000 UTC m=+1.246408372" Aug 13 07:09:15.147984 kubelet[2519]: I0813 07:09:15.147898 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.147879367 podStartE2EDuration="1.147879367s" podCreationTimestamp="2025-08-13 07:09:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:09:15.140978781 +0000 UTC m=+1.246668313" watchObservedRunningTime="2025-08-13 07:09:15.147879367 +0000 UTC m=+1.253568899" Aug 13 07:09:15.160007 kubelet[2519]: I0813 07:09:15.159918 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.1598970419999999 podStartE2EDuration="1.159897042s" podCreationTimestamp="2025-08-13 07:09:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:09:15.147869638 +0000 UTC m=+1.253559190" watchObservedRunningTime="2025-08-13 07:09:15.159897042 +0000 UTC m=+1.265586574" Aug 13 07:09:16.006983 kubelet[2519]: E0813 07:09:16.006947 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:16.007891 kubelet[2519]: E0813 07:09:16.007639 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:16.008396 kubelet[2519]: E0813 07:09:16.008370 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:16.531125 sudo[1640]: pam_unix(sudo:session): session closed for user root Aug 13 07:09:16.533499 sshd[1637]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:16.537983 systemd[1]: sshd@6-10.0.0.69:22-10.0.0.1:39514.service: Deactivated successfully. Aug 13 07:09:16.540081 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 07:09:16.540308 systemd[1]: session-7.scope: Consumed 7.383s CPU time, 157.3M memory peak, 0B memory swap peak. Aug 13 07:09:16.540773 systemd-logind[1450]: Session 7 logged out. Waiting for processes to exit. Aug 13 07:09:16.542162 systemd-logind[1450]: Removed session 7. Aug 13 07:09:17.008387 kubelet[2519]: E0813 07:09:17.008236 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:17.008387 kubelet[2519]: E0813 07:09:17.008326 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:20.397530 kubelet[2519]: I0813 07:09:20.397487 2519 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 07:09:20.398168 containerd[1461]: time="2025-08-13T07:09:20.397972789Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 07:09:20.398588 kubelet[2519]: I0813 07:09:20.398198 2519 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 07:09:20.554675 kubelet[2519]: E0813 07:09:20.554548 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.013304 kubelet[2519]: E0813 07:09:21.013236 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.456462 systemd[1]: Created slice kubepods-besteffort-pode150b15b_49fe_4b17_971f_93fdfdc83b45.slice - libcontainer container kubepods-besteffort-pode150b15b_49fe_4b17_971f_93fdfdc83b45.slice. Aug 13 07:09:21.476193 systemd[1]: Created slice kubepods-burstable-pod0f16dfbc_754e_4f6b_963a_7766b242d9d2.slice - libcontainer container kubepods-burstable-pod0f16dfbc_754e_4f6b_963a_7766b242d9d2.slice. Aug 13 07:09:21.511433 kubelet[2519]: I0813 07:09:21.511377 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-xtables-lock\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.511433 kubelet[2519]: I0813 07:09:21.511419 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0f16dfbc-754e-4f6b-963a-7766b242d9d2-clustermesh-secrets\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.511433 kubelet[2519]: I0813 07:09:21.511438 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-config-path\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512002 kubelet[2519]: I0813 07:09:21.511452 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-net\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512002 kubelet[2519]: I0813 07:09:21.511466 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hubble-tls\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512002 kubelet[2519]: I0813 07:09:21.511483 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zskbm\" (UniqueName: \"kubernetes.io/projected/e150b15b-49fe-4b17-971f-93fdfdc83b45-kube-api-access-zskbm\") pod \"kube-proxy-pcjz8\" (UID: \"e150b15b-49fe-4b17-971f-93fdfdc83b45\") " pod="kube-system/kube-proxy-pcjz8" Aug 13 07:09:21.512002 kubelet[2519]: I0813 07:09:21.511516 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-bpf-maps\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512002 kubelet[2519]: I0813 07:09:21.511532 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cni-path\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512002 kubelet[2519]: I0813 07:09:21.511551 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/e150b15b-49fe-4b17-971f-93fdfdc83b45-kube-proxy\") pod \"kube-proxy-pcjz8\" (UID: \"e150b15b-49fe-4b17-971f-93fdfdc83b45\") " pod="kube-system/kube-proxy-pcjz8" Aug 13 07:09:21.512151 kubelet[2519]: I0813 07:09:21.511576 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-cgroup\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512151 kubelet[2519]: I0813 07:09:21.511607 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-lib-modules\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512151 kubelet[2519]: I0813 07:09:21.511659 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-kernel\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512151 kubelet[2519]: I0813 07:09:21.511702 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mksb5\" (UniqueName: \"kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-kube-api-access-mksb5\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512151 kubelet[2519]: I0813 07:09:21.511742 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e150b15b-49fe-4b17-971f-93fdfdc83b45-xtables-lock\") pod \"kube-proxy-pcjz8\" (UID: \"e150b15b-49fe-4b17-971f-93fdfdc83b45\") " pod="kube-system/kube-proxy-pcjz8" Aug 13 07:09:21.512151 kubelet[2519]: I0813 07:09:21.511757 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hostproc\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512281 kubelet[2519]: I0813 07:09:21.511775 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-etc-cni-netd\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.512281 kubelet[2519]: I0813 07:09:21.511797 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e150b15b-49fe-4b17-971f-93fdfdc83b45-lib-modules\") pod \"kube-proxy-pcjz8\" (UID: \"e150b15b-49fe-4b17-971f-93fdfdc83b45\") " pod="kube-system/kube-proxy-pcjz8" Aug 13 07:09:21.512281 kubelet[2519]: I0813 07:09:21.511813 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-run\") pod \"cilium-lwr7l\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " pod="kube-system/cilium-lwr7l" Aug 13 07:09:21.603479 systemd[1]: Created slice kubepods-besteffort-podece56966_6890_43be_8d79_762183dde354.slice - libcontainer container kubepods-besteffort-podece56966_6890_43be_8d79_762183dde354.slice. Aug 13 07:09:21.612872 kubelet[2519]: I0813 07:09:21.612574 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqq2g\" (UniqueName: \"kubernetes.io/projected/ece56966-6890-43be-8d79-762183dde354-kube-api-access-rqq2g\") pod \"cilium-operator-6c4d7847fc-dmmq4\" (UID: \"ece56966-6890-43be-8d79-762183dde354\") " pod="kube-system/cilium-operator-6c4d7847fc-dmmq4" Aug 13 07:09:21.616757 kubelet[2519]: I0813 07:09:21.616718 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ece56966-6890-43be-8d79-762183dde354-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-dmmq4\" (UID: \"ece56966-6890-43be-8d79-762183dde354\") " pod="kube-system/cilium-operator-6c4d7847fc-dmmq4" Aug 13 07:09:21.769901 kubelet[2519]: E0813 07:09:21.769725 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.770714 containerd[1461]: time="2025-08-13T07:09:21.770667653Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-pcjz8,Uid:e150b15b-49fe-4b17-971f-93fdfdc83b45,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:21.779996 kubelet[2519]: E0813 07:09:21.779971 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.780517 containerd[1461]: time="2025-08-13T07:09:21.780452444Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lwr7l,Uid:0f16dfbc-754e-4f6b-963a-7766b242d9d2,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:21.800485 containerd[1461]: time="2025-08-13T07:09:21.800365752Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:21.800679 containerd[1461]: time="2025-08-13T07:09:21.800515749Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:21.800679 containerd[1461]: time="2025-08-13T07:09:21.800530607Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:21.801203 containerd[1461]: time="2025-08-13T07:09:21.801158765Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:21.806027 containerd[1461]: time="2025-08-13T07:09:21.805759980Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:21.806027 containerd[1461]: time="2025-08-13T07:09:21.805832258Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:21.806027 containerd[1461]: time="2025-08-13T07:09:21.805864449Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:21.806985 containerd[1461]: time="2025-08-13T07:09:21.806186424Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:21.826036 systemd[1]: Started cri-containerd-b35cad9892df728c857bce0d81737fd0a4885200ba5551d603d517f36569f9bd.scope - libcontainer container b35cad9892df728c857bce0d81737fd0a4885200ba5551d603d517f36569f9bd. Aug 13 07:09:21.829997 systemd[1]: Started cri-containerd-e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33.scope - libcontainer container e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33. Aug 13 07:09:21.853680 containerd[1461]: time="2025-08-13T07:09:21.853597715Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-pcjz8,Uid:e150b15b-49fe-4b17-971f-93fdfdc83b45,Namespace:kube-system,Attempt:0,} returns sandbox id \"b35cad9892df728c857bce0d81737fd0a4885200ba5551d603d517f36569f9bd\"" Aug 13 07:09:21.854631 kubelet[2519]: E0813 07:09:21.854559 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.858041 containerd[1461]: time="2025-08-13T07:09:21.858005181Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lwr7l,Uid:0f16dfbc-754e-4f6b-963a-7766b242d9d2,Namespace:kube-system,Attempt:0,} returns sandbox id \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\"" Aug 13 07:09:21.860062 kubelet[2519]: E0813 07:09:21.860039 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.861361 containerd[1461]: time="2025-08-13T07:09:21.861058374Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Aug 13 07:09:21.862683 containerd[1461]: time="2025-08-13T07:09:21.862652584Z" level=info msg="CreateContainer within sandbox \"b35cad9892df728c857bce0d81737fd0a4885200ba5551d603d517f36569f9bd\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 07:09:21.882694 containerd[1461]: time="2025-08-13T07:09:21.882629584Z" level=info msg="CreateContainer within sandbox \"b35cad9892df728c857bce0d81737fd0a4885200ba5551d603d517f36569f9bd\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"bedc89b68898eca5c6c7898e621288b27fe2333ed9f8db8e4a395c6fbbe6b3c7\"" Aug 13 07:09:21.883368 containerd[1461]: time="2025-08-13T07:09:21.883318418Z" level=info msg="StartContainer for \"bedc89b68898eca5c6c7898e621288b27fe2333ed9f8db8e4a395c6fbbe6b3c7\"" Aug 13 07:09:21.907615 kubelet[2519]: E0813 07:09:21.907579 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:21.908680 containerd[1461]: time="2025-08-13T07:09:21.908605043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-dmmq4,Uid:ece56966-6890-43be-8d79-762183dde354,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:21.919053 systemd[1]: Started cri-containerd-bedc89b68898eca5c6c7898e621288b27fe2333ed9f8db8e4a395c6fbbe6b3c7.scope - libcontainer container bedc89b68898eca5c6c7898e621288b27fe2333ed9f8db8e4a395c6fbbe6b3c7. Aug 13 07:09:21.938373 containerd[1461]: time="2025-08-13T07:09:21.938252285Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:21.938615 containerd[1461]: time="2025-08-13T07:09:21.938319784Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:21.938615 containerd[1461]: time="2025-08-13T07:09:21.938366794Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:21.938615 containerd[1461]: time="2025-08-13T07:09:21.938476503Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:21.956441 containerd[1461]: time="2025-08-13T07:09:21.956385468Z" level=info msg="StartContainer for \"bedc89b68898eca5c6c7898e621288b27fe2333ed9f8db8e4a395c6fbbe6b3c7\" returns successfully" Aug 13 07:09:21.964171 systemd[1]: Started cri-containerd-6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026.scope - libcontainer container 6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026. Aug 13 07:09:22.012970 containerd[1461]: time="2025-08-13T07:09:22.012802693Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-dmmq4,Uid:ece56966-6890-43be-8d79-762183dde354,Namespace:kube-system,Attempt:0,} returns sandbox id \"6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026\"" Aug 13 07:09:22.013463 kubelet[2519]: E0813 07:09:22.013439 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:22.016983 kubelet[2519]: E0813 07:09:22.016951 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:22.019480 kubelet[2519]: E0813 07:09:22.019450 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:22.026715 kubelet[2519]: I0813 07:09:22.026550 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-pcjz8" podStartSLOduration=1.02653236 podStartE2EDuration="1.02653236s" podCreationTimestamp="2025-08-13 07:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:09:22.026230745 +0000 UTC m=+8.131920277" watchObservedRunningTime="2025-08-13 07:09:22.02653236 +0000 UTC m=+8.132221892" Aug 13 07:09:25.427231 update_engine[1452]: I20250813 07:09:25.427062 1452 update_attempter.cc:509] Updating boot flags... Aug 13 07:09:25.455873 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2906) Aug 13 07:09:25.489894 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2908) Aug 13 07:09:25.514598 kubelet[2519]: E0813 07:09:25.514400 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:25.537466 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2908) Aug 13 07:09:26.000863 kubelet[2519]: E0813 07:09:26.000809 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:26.023872 kubelet[2519]: E0813 07:09:26.023827 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:34.966253 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount610822494.mount: Deactivated successfully. Aug 13 07:09:38.373326 containerd[1461]: time="2025-08-13T07:09:38.373242680Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:38.374064 containerd[1461]: time="2025-08-13T07:09:38.373988105Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Aug 13 07:09:38.375324 containerd[1461]: time="2025-08-13T07:09:38.375282706Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:38.377136 containerd[1461]: time="2025-08-13T07:09:38.377088962Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 16.515972818s" Aug 13 07:09:38.377136 containerd[1461]: time="2025-08-13T07:09:38.377131402Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Aug 13 07:09:38.381933 containerd[1461]: time="2025-08-13T07:09:38.381872201Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Aug 13 07:09:38.396513 containerd[1461]: time="2025-08-13T07:09:38.396469604Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 07:09:38.411867 containerd[1461]: time="2025-08-13T07:09:38.411798557Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\"" Aug 13 07:09:38.412605 containerd[1461]: time="2025-08-13T07:09:38.412567146Z" level=info msg="StartContainer for \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\"" Aug 13 07:09:38.447049 systemd[1]: Started cri-containerd-bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563.scope - libcontainer container bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563. Aug 13 07:09:38.496159 systemd[1]: cri-containerd-bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563.scope: Deactivated successfully. Aug 13 07:09:38.796207 containerd[1461]: time="2025-08-13T07:09:38.796134212Z" level=info msg="StartContainer for \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\" returns successfully" Aug 13 07:09:38.820408 containerd[1461]: time="2025-08-13T07:09:38.820309414Z" level=info msg="shim disconnected" id=bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563 namespace=k8s.io Aug 13 07:09:38.820408 containerd[1461]: time="2025-08-13T07:09:38.820399243Z" level=warning msg="cleaning up after shim disconnected" id=bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563 namespace=k8s.io Aug 13 07:09:38.820408 containerd[1461]: time="2025-08-13T07:09:38.820412659Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:09:39.166017 kubelet[2519]: E0813 07:09:39.165979 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:39.170967 containerd[1461]: time="2025-08-13T07:09:39.170931116Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Aug 13 07:09:39.184536 containerd[1461]: time="2025-08-13T07:09:39.184462270Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\"" Aug 13 07:09:39.185227 containerd[1461]: time="2025-08-13T07:09:39.185189200Z" level=info msg="StartContainer for \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\"" Aug 13 07:09:39.212975 systemd[1]: Started cri-containerd-8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77.scope - libcontainer container 8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77. Aug 13 07:09:39.238286 containerd[1461]: time="2025-08-13T07:09:39.238241324Z" level=info msg="StartContainer for \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\" returns successfully" Aug 13 07:09:39.250932 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:09:39.251177 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:09:39.251248 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:09:39.257162 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:09:39.257518 systemd[1]: cri-containerd-8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77.scope: Deactivated successfully. Aug 13 07:09:39.282620 containerd[1461]: time="2025-08-13T07:09:39.282562181Z" level=info msg="shim disconnected" id=8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77 namespace=k8s.io Aug 13 07:09:39.282873 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:09:39.283224 containerd[1461]: time="2025-08-13T07:09:39.282908344Z" level=warning msg="cleaning up after shim disconnected" id=8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77 namespace=k8s.io Aug 13 07:09:39.283224 containerd[1461]: time="2025-08-13T07:09:39.282921548Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:09:39.407210 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563-rootfs.mount: Deactivated successfully. Aug 13 07:09:39.732016 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount591224871.mount: Deactivated successfully. Aug 13 07:09:40.119446 containerd[1461]: time="2025-08-13T07:09:40.119301991Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:40.120134 containerd[1461]: time="2025-08-13T07:09:40.120096889Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Aug 13 07:09:40.121375 containerd[1461]: time="2025-08-13T07:09:40.121323570Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:40.122755 containerd[1461]: time="2025-08-13T07:09:40.122718118Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 1.74079985s" Aug 13 07:09:40.122801 containerd[1461]: time="2025-08-13T07:09:40.122752332Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Aug 13 07:09:40.129028 containerd[1461]: time="2025-08-13T07:09:40.128981700Z" level=info msg="CreateContainer within sandbox \"6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Aug 13 07:09:40.142305 containerd[1461]: time="2025-08-13T07:09:40.142264066Z" level=info msg="CreateContainer within sandbox \"6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\"" Aug 13 07:09:40.142718 containerd[1461]: time="2025-08-13T07:09:40.142694077Z" level=info msg="StartContainer for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\"" Aug 13 07:09:40.170994 kubelet[2519]: E0813 07:09:40.170103 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:40.172205 systemd[1]: Started cri-containerd-04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7.scope - libcontainer container 04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7. Aug 13 07:09:40.178000 containerd[1461]: time="2025-08-13T07:09:40.177621156Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Aug 13 07:09:40.321082 containerd[1461]: time="2025-08-13T07:09:40.321018393Z" level=info msg="StartContainer for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" returns successfully" Aug 13 07:09:40.339487 containerd[1461]: time="2025-08-13T07:09:40.339416688Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\"" Aug 13 07:09:40.343878 containerd[1461]: time="2025-08-13T07:09:40.341618426Z" level=info msg="StartContainer for \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\"" Aug 13 07:09:40.417324 systemd[1]: Started cri-containerd-8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba.scope - libcontainer container 8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba. Aug 13 07:09:40.449488 systemd[1]: cri-containerd-8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba.scope: Deactivated successfully. Aug 13 07:09:40.869827 containerd[1461]: time="2025-08-13T07:09:40.869714708Z" level=info msg="StartContainer for \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\" returns successfully" Aug 13 07:09:40.891274 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba-rootfs.mount: Deactivated successfully. Aug 13 07:09:40.961554 containerd[1461]: time="2025-08-13T07:09:40.961477996Z" level=info msg="shim disconnected" id=8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba namespace=k8s.io Aug 13 07:09:40.961554 containerd[1461]: time="2025-08-13T07:09:40.961542628Z" level=warning msg="cleaning up after shim disconnected" id=8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba namespace=k8s.io Aug 13 07:09:40.961554 containerd[1461]: time="2025-08-13T07:09:40.961551786Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:09:41.175413 kubelet[2519]: E0813 07:09:41.175368 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:41.178255 kubelet[2519]: E0813 07:09:41.178209 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:41.182662 containerd[1461]: time="2025-08-13T07:09:41.182625040Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Aug 13 07:09:41.210458 kubelet[2519]: I0813 07:09:41.210109 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-dmmq4" podStartSLOduration=2.100599253 podStartE2EDuration="20.210089472s" podCreationTimestamp="2025-08-13 07:09:21 +0000 UTC" firstStartedPulling="2025-08-13 07:09:22.013983362 +0000 UTC m=+8.119672894" lastFinishedPulling="2025-08-13 07:09:40.123473581 +0000 UTC m=+26.229163113" observedRunningTime="2025-08-13 07:09:41.190810558 +0000 UTC m=+27.296500090" watchObservedRunningTime="2025-08-13 07:09:41.210089472 +0000 UTC m=+27.315779004" Aug 13 07:09:41.215425 containerd[1461]: time="2025-08-13T07:09:41.215391920Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\"" Aug 13 07:09:41.216108 containerd[1461]: time="2025-08-13T07:09:41.216063004Z" level=info msg="StartContainer for \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\"" Aug 13 07:09:41.257981 systemd[1]: Started cri-containerd-6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be.scope - libcontainer container 6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be. Aug 13 07:09:41.284044 systemd[1]: cri-containerd-6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be.scope: Deactivated successfully. Aug 13 07:09:41.286502 containerd[1461]: time="2025-08-13T07:09:41.286470433Z" level=info msg="StartContainer for \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\" returns successfully" Aug 13 07:09:41.311621 containerd[1461]: time="2025-08-13T07:09:41.311538951Z" level=info msg="shim disconnected" id=6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be namespace=k8s.io Aug 13 07:09:41.311621 containerd[1461]: time="2025-08-13T07:09:41.311609243Z" level=warning msg="cleaning up after shim disconnected" id=6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be namespace=k8s.io Aug 13 07:09:41.311621 containerd[1461]: time="2025-08-13T07:09:41.311619523Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:09:41.325552 containerd[1461]: time="2025-08-13T07:09:41.325500920Z" level=warning msg="cleanup warnings time=\"2025-08-13T07:09:41Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Aug 13 07:09:41.419320 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be-rootfs.mount: Deactivated successfully. Aug 13 07:09:41.523167 systemd[1]: Started sshd@7-10.0.0.69:22-10.0.0.1:43482.service - OpenSSH per-connection server daemon (10.0.0.1:43482). Aug 13 07:09:41.563402 sshd[3238]: Accepted publickey for core from 10.0.0.1 port 43482 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:09:41.564985 sshd[3238]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:41.569097 systemd-logind[1450]: New session 8 of user core. Aug 13 07:09:41.577984 systemd[1]: Started session-8.scope - Session 8 of User core. Aug 13 07:09:41.711787 sshd[3238]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:41.716203 systemd[1]: sshd@7-10.0.0.69:22-10.0.0.1:43482.service: Deactivated successfully. Aug 13 07:09:41.718139 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 07:09:41.718734 systemd-logind[1450]: Session 8 logged out. Waiting for processes to exit. Aug 13 07:09:41.719666 systemd-logind[1450]: Removed session 8. Aug 13 07:09:42.183659 kubelet[2519]: E0813 07:09:42.183618 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:42.184213 kubelet[2519]: E0813 07:09:42.183748 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:42.190201 containerd[1461]: time="2025-08-13T07:09:42.190146010Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Aug 13 07:09:42.207661 containerd[1461]: time="2025-08-13T07:09:42.207607621Z" level=info msg="CreateContainer within sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\"" Aug 13 07:09:42.208229 containerd[1461]: time="2025-08-13T07:09:42.208190819Z" level=info msg="StartContainer for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\"" Aug 13 07:09:42.239035 systemd[1]: Started cri-containerd-cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8.scope - libcontainer container cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8. Aug 13 07:09:42.271796 containerd[1461]: time="2025-08-13T07:09:42.271747267Z" level=info msg="StartContainer for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" returns successfully" Aug 13 07:09:42.410115 kubelet[2519]: I0813 07:09:42.408335 2519 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Aug 13 07:09:42.479760 systemd[1]: Created slice kubepods-burstable-pod427b82b6_050c_49e4_9027_7cde06b29c1a.slice - libcontainer container kubepods-burstable-pod427b82b6_050c_49e4_9027_7cde06b29c1a.slice. Aug 13 07:09:42.483381 kubelet[2519]: I0813 07:09:42.483339 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fd30912-9720-4614-90a1-9316f1451ee2-config-volume\") pod \"coredns-674b8bbfcf-vghhs\" (UID: \"8fd30912-9720-4614-90a1-9316f1451ee2\") " pod="kube-system/coredns-674b8bbfcf-vghhs" Aug 13 07:09:42.483381 kubelet[2519]: I0813 07:09:42.483379 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/427b82b6-050c-49e4-9027-7cde06b29c1a-config-volume\") pod \"coredns-674b8bbfcf-gsc7l\" (UID: \"427b82b6-050c-49e4-9027-7cde06b29c1a\") " pod="kube-system/coredns-674b8bbfcf-gsc7l" Aug 13 07:09:42.483497 kubelet[2519]: I0813 07:09:42.483395 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h766t\" (UniqueName: \"kubernetes.io/projected/8fd30912-9720-4614-90a1-9316f1451ee2-kube-api-access-h766t\") pod \"coredns-674b8bbfcf-vghhs\" (UID: \"8fd30912-9720-4614-90a1-9316f1451ee2\") " pod="kube-system/coredns-674b8bbfcf-vghhs" Aug 13 07:09:42.483497 kubelet[2519]: I0813 07:09:42.483415 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcfbg\" (UniqueName: \"kubernetes.io/projected/427b82b6-050c-49e4-9027-7cde06b29c1a-kube-api-access-wcfbg\") pod \"coredns-674b8bbfcf-gsc7l\" (UID: \"427b82b6-050c-49e4-9027-7cde06b29c1a\") " pod="kube-system/coredns-674b8bbfcf-gsc7l" Aug 13 07:09:42.487826 systemd[1]: Created slice kubepods-burstable-pod8fd30912_9720_4614_90a1_9316f1451ee2.slice - libcontainer container kubepods-burstable-pod8fd30912_9720_4614_90a1_9316f1451ee2.slice. Aug 13 07:09:42.784943 kubelet[2519]: E0813 07:09:42.784792 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:42.785831 containerd[1461]: time="2025-08-13T07:09:42.785738599Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gsc7l,Uid:427b82b6-050c-49e4-9027-7cde06b29c1a,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:42.790609 kubelet[2519]: E0813 07:09:42.790573 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:42.791209 containerd[1461]: time="2025-08-13T07:09:42.791153285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-vghhs,Uid:8fd30912-9720-4614-90a1-9316f1451ee2,Namespace:kube-system,Attempt:0,}" Aug 13 07:09:43.188041 kubelet[2519]: E0813 07:09:43.188013 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:43.202149 kubelet[2519]: I0813 07:09:43.202076 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-lwr7l" podStartSLOduration=5.6811004579999995 podStartE2EDuration="22.202050761s" podCreationTimestamp="2025-08-13 07:09:21 +0000 UTC" firstStartedPulling="2025-08-13 07:09:21.86065816 +0000 UTC m=+7.966347692" lastFinishedPulling="2025-08-13 07:09:38.381608453 +0000 UTC m=+24.487297995" observedRunningTime="2025-08-13 07:09:43.201135027 +0000 UTC m=+29.306824559" watchObservedRunningTime="2025-08-13 07:09:43.202050761 +0000 UTC m=+29.307740293" Aug 13 07:09:44.190139 kubelet[2519]: E0813 07:09:44.190090 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:44.503647 systemd-networkd[1382]: cilium_host: Link UP Aug 13 07:09:44.503819 systemd-networkd[1382]: cilium_net: Link UP Aug 13 07:09:44.504040 systemd-networkd[1382]: cilium_net: Gained carrier Aug 13 07:09:44.504221 systemd-networkd[1382]: cilium_host: Gained carrier Aug 13 07:09:44.624887 systemd-networkd[1382]: cilium_vxlan: Link UP Aug 13 07:09:44.624899 systemd-networkd[1382]: cilium_vxlan: Gained carrier Aug 13 07:09:44.841879 kernel: NET: Registered PF_ALG protocol family Aug 13 07:09:45.066086 systemd-networkd[1382]: cilium_host: Gained IPv6LL Aug 13 07:09:45.192498 kubelet[2519]: E0813 07:09:45.192461 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:45.386047 systemd-networkd[1382]: cilium_net: Gained IPv6LL Aug 13 07:09:45.537942 systemd-networkd[1382]: lxc_health: Link UP Aug 13 07:09:45.551991 systemd-networkd[1382]: lxc_health: Gained carrier Aug 13 07:09:45.866020 systemd-networkd[1382]: lxcaa086143c34c: Link UP Aug 13 07:09:45.872871 kernel: eth0: renamed from tmpd9ade Aug 13 07:09:45.878352 systemd-networkd[1382]: lxc94d89968085c: Link UP Aug 13 07:09:45.885140 systemd-networkd[1382]: lxcaa086143c34c: Gained carrier Aug 13 07:09:45.891108 kernel: eth0: renamed from tmpa4726 Aug 13 07:09:45.897440 systemd-networkd[1382]: lxc94d89968085c: Gained carrier Aug 13 07:09:46.194479 kubelet[2519]: E0813 07:09:46.194254 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:46.346094 systemd-networkd[1382]: cilium_vxlan: Gained IPv6LL Aug 13 07:09:46.726210 systemd[1]: Started sshd@8-10.0.0.69:22-10.0.0.1:43484.service - OpenSSH per-connection server daemon (10.0.0.1:43484). Aug 13 07:09:46.783844 sshd[3778]: Accepted publickey for core from 10.0.0.1 port 43484 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:09:46.786063 sshd[3778]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:46.791053 systemd-logind[1450]: New session 9 of user core. Aug 13 07:09:46.798025 systemd[1]: Started session-9.scope - Session 9 of User core. Aug 13 07:09:46.986685 sshd[3778]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:46.991896 systemd[1]: sshd@8-10.0.0.69:22-10.0.0.1:43484.service: Deactivated successfully. Aug 13 07:09:46.994046 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 07:09:46.994883 systemd-logind[1450]: Session 9 logged out. Waiting for processes to exit. Aug 13 07:09:46.995867 systemd-logind[1450]: Removed session 9. Aug 13 07:09:47.195309 kubelet[2519]: E0813 07:09:47.195256 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:47.434050 systemd-networkd[1382]: lxc_health: Gained IPv6LL Aug 13 07:09:47.882080 systemd-networkd[1382]: lxc94d89968085c: Gained IPv6LL Aug 13 07:09:47.945984 systemd-networkd[1382]: lxcaa086143c34c: Gained IPv6LL Aug 13 07:09:48.197420 kubelet[2519]: E0813 07:09:48.197344 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:49.250373 containerd[1461]: time="2025-08-13T07:09:49.250238775Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:49.250373 containerd[1461]: time="2025-08-13T07:09:49.250301382Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:49.250373 containerd[1461]: time="2025-08-13T07:09:49.250316020Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:49.250994 containerd[1461]: time="2025-08-13T07:09:49.250410558Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:49.261317 containerd[1461]: time="2025-08-13T07:09:49.261150118Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:09:49.261317 containerd[1461]: time="2025-08-13T07:09:49.261228044Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:09:49.261317 containerd[1461]: time="2025-08-13T07:09:49.261266396Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:49.261623 containerd[1461]: time="2025-08-13T07:09:49.261426366Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:09:49.283005 systemd[1]: Started cri-containerd-d9ade207d1553a6107dc3be829c157f9afe9f4d54da3a83f46b3380759995bf0.scope - libcontainer container d9ade207d1553a6107dc3be829c157f9afe9f4d54da3a83f46b3380759995bf0. Aug 13 07:09:49.287943 systemd[1]: Started cri-containerd-a47265d034b8fdcbd4361e8b98635fe4aa9c03815286885d43b6c7668d7a0ed7.scope - libcontainer container a47265d034b8fdcbd4361e8b98635fe4aa9c03815286885d43b6c7668d7a0ed7. Aug 13 07:09:49.295225 systemd-resolved[1336]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Aug 13 07:09:49.301426 systemd-resolved[1336]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Aug 13 07:09:49.322717 containerd[1461]: time="2025-08-13T07:09:49.322667639Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-vghhs,Uid:8fd30912-9720-4614-90a1-9316f1451ee2,Namespace:kube-system,Attempt:0,} returns sandbox id \"d9ade207d1553a6107dc3be829c157f9afe9f4d54da3a83f46b3380759995bf0\"" Aug 13 07:09:49.323432 kubelet[2519]: E0813 07:09:49.323407 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:49.329744 containerd[1461]: time="2025-08-13T07:09:49.329680794Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gsc7l,Uid:427b82b6-050c-49e4-9027-7cde06b29c1a,Namespace:kube-system,Attempt:0,} returns sandbox id \"a47265d034b8fdcbd4361e8b98635fe4aa9c03815286885d43b6c7668d7a0ed7\"" Aug 13 07:09:49.330914 containerd[1461]: time="2025-08-13T07:09:49.330879487Z" level=info msg="CreateContainer within sandbox \"d9ade207d1553a6107dc3be829c157f9afe9f4d54da3a83f46b3380759995bf0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:09:49.331422 kubelet[2519]: E0813 07:09:49.331375 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:49.336374 containerd[1461]: time="2025-08-13T07:09:49.336329082Z" level=info msg="CreateContainer within sandbox \"a47265d034b8fdcbd4361e8b98635fe4aa9c03815286885d43b6c7668d7a0ed7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:09:49.355437 containerd[1461]: time="2025-08-13T07:09:49.355377062Z" level=info msg="CreateContainer within sandbox \"a47265d034b8fdcbd4361e8b98635fe4aa9c03815286885d43b6c7668d7a0ed7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"537283ab41840c09e850b91c89c6dcb8e3c6a3966faa9719f9f7c161020e500c\"" Aug 13 07:09:49.356184 containerd[1461]: time="2025-08-13T07:09:49.356084773Z" level=info msg="StartContainer for \"537283ab41840c09e850b91c89c6dcb8e3c6a3966faa9719f9f7c161020e500c\"" Aug 13 07:09:49.358417 containerd[1461]: time="2025-08-13T07:09:49.358369398Z" level=info msg="CreateContainer within sandbox \"d9ade207d1553a6107dc3be829c157f9afe9f4d54da3a83f46b3380759995bf0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"bb0279338d056d23459a0f5ef497a9a6caeb65396276d6a661050763c0483815\"" Aug 13 07:09:49.358909 containerd[1461]: time="2025-08-13T07:09:49.358882282Z" level=info msg="StartContainer for \"bb0279338d056d23459a0f5ef497a9a6caeb65396276d6a661050763c0483815\"" Aug 13 07:09:49.392059 systemd[1]: Started cri-containerd-537283ab41840c09e850b91c89c6dcb8e3c6a3966faa9719f9f7c161020e500c.scope - libcontainer container 537283ab41840c09e850b91c89c6dcb8e3c6a3966faa9719f9f7c161020e500c. Aug 13 07:09:49.395524 systemd[1]: Started cri-containerd-bb0279338d056d23459a0f5ef497a9a6caeb65396276d6a661050763c0483815.scope - libcontainer container bb0279338d056d23459a0f5ef497a9a6caeb65396276d6a661050763c0483815. Aug 13 07:09:49.437899 containerd[1461]: time="2025-08-13T07:09:49.437779708Z" level=info msg="StartContainer for \"bb0279338d056d23459a0f5ef497a9a6caeb65396276d6a661050763c0483815\" returns successfully" Aug 13 07:09:49.443591 containerd[1461]: time="2025-08-13T07:09:49.443541601Z" level=info msg="StartContainer for \"537283ab41840c09e850b91c89c6dcb8e3c6a3966faa9719f9f7c161020e500c\" returns successfully" Aug 13 07:09:50.204359 kubelet[2519]: E0813 07:09:50.203923 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:50.207271 kubelet[2519]: E0813 07:09:50.207220 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:50.258352 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1194415718.mount: Deactivated successfully. Aug 13 07:09:50.706047 kubelet[2519]: I0813 07:09:50.705982 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-vghhs" podStartSLOduration=29.705959838 podStartE2EDuration="29.705959838s" podCreationTimestamp="2025-08-13 07:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:09:50.426002397 +0000 UTC m=+36.531691929" watchObservedRunningTime="2025-08-13 07:09:50.705959838 +0000 UTC m=+36.811649370" Aug 13 07:09:51.008545 kubelet[2519]: I0813 07:09:51.008328 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-gsc7l" podStartSLOduration=30.008307772 podStartE2EDuration="30.008307772s" podCreationTimestamp="2025-08-13 07:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:09:50.706870639 +0000 UTC m=+36.812560171" watchObservedRunningTime="2025-08-13 07:09:51.008307772 +0000 UTC m=+37.113997294" Aug 13 07:09:51.209568 kubelet[2519]: E0813 07:09:51.209523 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:51.209875 kubelet[2519]: E0813 07:09:51.209603 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:52.001139 systemd[1]: Started sshd@9-10.0.0.69:22-10.0.0.1:48024.service - OpenSSH per-connection server daemon (10.0.0.1:48024). Aug 13 07:09:52.043395 sshd[3967]: Accepted publickey for core from 10.0.0.1 port 48024 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:09:52.045345 sshd[3967]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:52.049683 systemd-logind[1450]: New session 10 of user core. Aug 13 07:09:52.061983 systemd[1]: Started session-10.scope - Session 10 of User core. Aug 13 07:09:52.211249 kubelet[2519]: E0813 07:09:52.211210 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:52.212364 kubelet[2519]: E0813 07:09:52.212037 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:09:52.277261 sshd[3967]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:52.282706 systemd[1]: sshd@9-10.0.0.69:22-10.0.0.1:48024.service: Deactivated successfully. Aug 13 07:09:52.284980 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 07:09:52.285733 systemd-logind[1450]: Session 10 logged out. Waiting for processes to exit. Aug 13 07:09:52.286895 systemd-logind[1450]: Removed session 10. Aug 13 07:09:57.288964 systemd[1]: Started sshd@10-10.0.0.69:22-10.0.0.1:48040.service - OpenSSH per-connection server daemon (10.0.0.1:48040). Aug 13 07:09:57.324997 sshd[3987]: Accepted publickey for core from 10.0.0.1 port 48040 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:09:57.326739 sshd[3987]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:57.331007 systemd-logind[1450]: New session 11 of user core. Aug 13 07:09:57.338970 systemd[1]: Started session-11.scope - Session 11 of User core. Aug 13 07:09:57.487939 sshd[3987]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:57.492029 systemd[1]: sshd@10-10.0.0.69:22-10.0.0.1:48040.service: Deactivated successfully. Aug 13 07:09:57.494085 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 07:09:57.494652 systemd-logind[1450]: Session 11 logged out. Waiting for processes to exit. Aug 13 07:09:57.495545 systemd-logind[1450]: Removed session 11. Aug 13 07:10:02.512109 systemd[1]: Started sshd@11-10.0.0.69:22-10.0.0.1:37352.service - OpenSSH per-connection server daemon (10.0.0.1:37352). Aug 13 07:10:02.544664 sshd[4003]: Accepted publickey for core from 10.0.0.1 port 37352 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:02.546322 sshd[4003]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:02.550331 systemd-logind[1450]: New session 12 of user core. Aug 13 07:10:02.563982 systemd[1]: Started session-12.scope - Session 12 of User core. Aug 13 07:10:02.670169 sshd[4003]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:02.677751 systemd[1]: sshd@11-10.0.0.69:22-10.0.0.1:37352.service: Deactivated successfully. Aug 13 07:10:02.679727 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 07:10:02.681258 systemd-logind[1450]: Session 12 logged out. Waiting for processes to exit. Aug 13 07:10:02.682708 systemd[1]: Started sshd@12-10.0.0.69:22-10.0.0.1:37368.service - OpenSSH per-connection server daemon (10.0.0.1:37368). Aug 13 07:10:02.683435 systemd-logind[1450]: Removed session 12. Aug 13 07:10:02.718167 sshd[4018]: Accepted publickey for core from 10.0.0.1 port 37368 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:02.719908 sshd[4018]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:02.724158 systemd-logind[1450]: New session 13 of user core. Aug 13 07:10:02.734181 systemd[1]: Started session-13.scope - Session 13 of User core. Aug 13 07:10:03.037149 sshd[4018]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:03.044986 systemd[1]: sshd@12-10.0.0.69:22-10.0.0.1:37368.service: Deactivated successfully. Aug 13 07:10:03.046890 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 07:10:03.048627 systemd-logind[1450]: Session 13 logged out. Waiting for processes to exit. Aug 13 07:10:03.060100 systemd[1]: Started sshd@13-10.0.0.69:22-10.0.0.1:37380.service - OpenSSH per-connection server daemon (10.0.0.1:37380). Aug 13 07:10:03.061126 systemd-logind[1450]: Removed session 13. Aug 13 07:10:03.096786 sshd[4030]: Accepted publickey for core from 10.0.0.1 port 37380 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:03.098669 sshd[4030]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:03.103135 systemd-logind[1450]: New session 14 of user core. Aug 13 07:10:03.112996 systemd[1]: Started session-14.scope - Session 14 of User core. Aug 13 07:10:03.401282 sshd[4030]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:03.406039 systemd[1]: sshd@13-10.0.0.69:22-10.0.0.1:37380.service: Deactivated successfully. Aug 13 07:10:03.408331 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 07:10:03.409082 systemd-logind[1450]: Session 14 logged out. Waiting for processes to exit. Aug 13 07:10:03.409989 systemd-logind[1450]: Removed session 14. Aug 13 07:10:08.386315 systemd[1]: Started sshd@14-10.0.0.69:22-10.0.0.1:53564.service - OpenSSH per-connection server daemon (10.0.0.1:53564). Aug 13 07:10:08.424997 sshd[4044]: Accepted publickey for core from 10.0.0.1 port 53564 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:08.427365 sshd[4044]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:08.431936 systemd-logind[1450]: New session 15 of user core. Aug 13 07:10:08.441126 systemd[1]: Started session-15.scope - Session 15 of User core. Aug 13 07:10:08.554793 sshd[4044]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:08.559453 systemd[1]: sshd@14-10.0.0.69:22-10.0.0.1:53564.service: Deactivated successfully. Aug 13 07:10:08.561713 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 07:10:08.562515 systemd-logind[1450]: Session 15 logged out. Waiting for processes to exit. Aug 13 07:10:08.563580 systemd-logind[1450]: Removed session 15. Aug 13 07:10:13.565957 systemd[1]: Started sshd@15-10.0.0.69:22-10.0.0.1:53580.service - OpenSSH per-connection server daemon (10.0.0.1:53580). Aug 13 07:10:13.601445 sshd[4059]: Accepted publickey for core from 10.0.0.1 port 53580 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:13.602922 sshd[4059]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:13.607276 systemd-logind[1450]: New session 16 of user core. Aug 13 07:10:13.616985 systemd[1]: Started session-16.scope - Session 16 of User core. Aug 13 07:10:13.720530 sshd[4059]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:13.730786 systemd[1]: sshd@15-10.0.0.69:22-10.0.0.1:53580.service: Deactivated successfully. Aug 13 07:10:13.732689 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 07:10:13.734243 systemd-logind[1450]: Session 16 logged out. Waiting for processes to exit. Aug 13 07:10:13.743089 systemd[1]: Started sshd@16-10.0.0.69:22-10.0.0.1:53594.service - OpenSSH per-connection server daemon (10.0.0.1:53594). Aug 13 07:10:13.743897 systemd-logind[1450]: Removed session 16. Aug 13 07:10:13.773724 sshd[4074]: Accepted publickey for core from 10.0.0.1 port 53594 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:13.775891 sshd[4074]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:13.780827 systemd-logind[1450]: New session 17 of user core. Aug 13 07:10:13.792040 systemd[1]: Started session-17.scope - Session 17 of User core. Aug 13 07:10:14.066417 sshd[4074]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:14.081946 systemd[1]: sshd@16-10.0.0.69:22-10.0.0.1:53594.service: Deactivated successfully. Aug 13 07:10:14.084002 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 07:10:14.085627 systemd-logind[1450]: Session 17 logged out. Waiting for processes to exit. Aug 13 07:10:14.095108 systemd[1]: Started sshd@17-10.0.0.69:22-10.0.0.1:53598.service - OpenSSH per-connection server daemon (10.0.0.1:53598). Aug 13 07:10:14.095929 systemd-logind[1450]: Removed session 17. Aug 13 07:10:14.130755 sshd[4089]: Accepted publickey for core from 10.0.0.1 port 53598 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:14.133106 sshd[4089]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:14.137680 systemd-logind[1450]: New session 18 of user core. Aug 13 07:10:14.150014 systemd[1]: Started session-18.scope - Session 18 of User core. Aug 13 07:10:14.614313 sshd[4089]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:14.622974 systemd[1]: sshd@17-10.0.0.69:22-10.0.0.1:53598.service: Deactivated successfully. Aug 13 07:10:14.625035 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 07:10:14.627040 systemd-logind[1450]: Session 18 logged out. Waiting for processes to exit. Aug 13 07:10:14.636333 systemd[1]: Started sshd@18-10.0.0.69:22-10.0.0.1:53600.service - OpenSSH per-connection server daemon (10.0.0.1:53600). Aug 13 07:10:14.638764 systemd-logind[1450]: Removed session 18. Aug 13 07:10:14.671731 sshd[4108]: Accepted publickey for core from 10.0.0.1 port 53600 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:14.673583 sshd[4108]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:14.678096 systemd-logind[1450]: New session 19 of user core. Aug 13 07:10:14.684987 systemd[1]: Started session-19.scope - Session 19 of User core. Aug 13 07:10:14.934630 sshd[4108]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:14.944251 systemd[1]: sshd@18-10.0.0.69:22-10.0.0.1:53600.service: Deactivated successfully. Aug 13 07:10:14.947243 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 07:10:14.952441 systemd-logind[1450]: Session 19 logged out. Waiting for processes to exit. Aug 13 07:10:14.958266 systemd[1]: Started sshd@19-10.0.0.69:22-10.0.0.1:53616.service - OpenSSH per-connection server daemon (10.0.0.1:53616). Aug 13 07:10:14.959804 systemd-logind[1450]: Removed session 19. Aug 13 07:10:14.994416 sshd[4121]: Accepted publickey for core from 10.0.0.1 port 53616 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:14.996368 sshd[4121]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:15.001600 systemd-logind[1450]: New session 20 of user core. Aug 13 07:10:15.014044 systemd[1]: Started session-20.scope - Session 20 of User core. Aug 13 07:10:15.136154 sshd[4121]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:15.141186 systemd[1]: sshd@19-10.0.0.69:22-10.0.0.1:53616.service: Deactivated successfully. Aug 13 07:10:15.143673 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 07:10:15.144529 systemd-logind[1450]: Session 20 logged out. Waiting for processes to exit. Aug 13 07:10:15.145606 systemd-logind[1450]: Removed session 20. Aug 13 07:10:20.147901 systemd[1]: Started sshd@20-10.0.0.69:22-10.0.0.1:39300.service - OpenSSH per-connection server daemon (10.0.0.1:39300). Aug 13 07:10:20.186206 sshd[4135]: Accepted publickey for core from 10.0.0.1 port 39300 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:20.187730 sshd[4135]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:20.191659 systemd-logind[1450]: New session 21 of user core. Aug 13 07:10:20.200227 systemd[1]: Started session-21.scope - Session 21 of User core. Aug 13 07:10:20.395063 sshd[4135]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:20.399125 systemd[1]: sshd@20-10.0.0.69:22-10.0.0.1:39300.service: Deactivated successfully. Aug 13 07:10:20.401335 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 07:10:20.402149 systemd-logind[1450]: Session 21 logged out. Waiting for processes to exit. Aug 13 07:10:20.403086 systemd-logind[1450]: Removed session 21. Aug 13 07:10:23.995815 kubelet[2519]: E0813 07:10:23.995746 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:25.407269 systemd[1]: Started sshd@21-10.0.0.69:22-10.0.0.1:39316.service - OpenSSH per-connection server daemon (10.0.0.1:39316). Aug 13 07:10:25.445373 sshd[4153]: Accepted publickey for core from 10.0.0.1 port 39316 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:25.447102 sshd[4153]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:25.451093 systemd-logind[1450]: New session 22 of user core. Aug 13 07:10:25.462967 systemd[1]: Started session-22.scope - Session 22 of User core. Aug 13 07:10:25.569093 sshd[4153]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:25.573525 systemd[1]: sshd@21-10.0.0.69:22-10.0.0.1:39316.service: Deactivated successfully. Aug 13 07:10:25.575422 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 07:10:25.576174 systemd-logind[1450]: Session 22 logged out. Waiting for processes to exit. Aug 13 07:10:25.577156 systemd-logind[1450]: Removed session 22. Aug 13 07:10:30.581033 systemd[1]: Started sshd@22-10.0.0.69:22-10.0.0.1:52034.service - OpenSSH per-connection server daemon (10.0.0.1:52034). Aug 13 07:10:30.617549 sshd[4167]: Accepted publickey for core from 10.0.0.1 port 52034 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:30.619141 sshd[4167]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:30.623378 systemd-logind[1450]: New session 23 of user core. Aug 13 07:10:30.632993 systemd[1]: Started session-23.scope - Session 23 of User core. Aug 13 07:10:30.738604 sshd[4167]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:30.743019 systemd[1]: sshd@22-10.0.0.69:22-10.0.0.1:52034.service: Deactivated successfully. Aug 13 07:10:30.744927 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 07:10:30.745677 systemd-logind[1450]: Session 23 logged out. Waiting for processes to exit. Aug 13 07:10:30.746574 systemd-logind[1450]: Removed session 23. Aug 13 07:10:31.995954 kubelet[2519]: E0813 07:10:31.995908 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:35.757510 systemd[1]: Started sshd@23-10.0.0.69:22-10.0.0.1:52036.service - OpenSSH per-connection server daemon (10.0.0.1:52036). Aug 13 07:10:35.793361 sshd[4181]: Accepted publickey for core from 10.0.0.1 port 52036 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:35.795013 sshd[4181]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:35.799214 systemd-logind[1450]: New session 24 of user core. Aug 13 07:10:35.809016 systemd[1]: Started session-24.scope - Session 24 of User core. Aug 13 07:10:35.947647 sshd[4181]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:35.961894 systemd[1]: sshd@23-10.0.0.69:22-10.0.0.1:52036.service: Deactivated successfully. Aug 13 07:10:35.964061 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 07:10:35.965689 systemd-logind[1450]: Session 24 logged out. Waiting for processes to exit. Aug 13 07:10:35.971227 systemd[1]: Started sshd@24-10.0.0.69:22-10.0.0.1:52048.service - OpenSSH per-connection server daemon (10.0.0.1:52048). Aug 13 07:10:35.972298 systemd-logind[1450]: Removed session 24. Aug 13 07:10:35.996012 kubelet[2519]: E0813 07:10:35.995753 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:36.008711 sshd[4195]: Accepted publickey for core from 10.0.0.1 port 52048 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:36.010599 sshd[4195]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:36.015629 systemd-logind[1450]: New session 25 of user core. Aug 13 07:10:36.025177 systemd[1]: Started session-25.scope - Session 25 of User core. Aug 13 07:10:38.106159 containerd[1461]: time="2025-08-13T07:10:38.106113527Z" level=info msg="StopContainer for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" with timeout 30 (s)" Aug 13 07:10:38.109077 containerd[1461]: time="2025-08-13T07:10:38.109051499Z" level=info msg="Stop container \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" with signal terminated" Aug 13 07:10:38.138835 systemd[1]: cri-containerd-04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7.scope: Deactivated successfully. Aug 13 07:10:38.155393 containerd[1461]: time="2025-08-13T07:10:38.155334807Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:10:38.160627 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7-rootfs.mount: Deactivated successfully. Aug 13 07:10:38.164115 containerd[1461]: time="2025-08-13T07:10:38.164067688Z" level=info msg="StopContainer for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" with timeout 2 (s)" Aug 13 07:10:38.164334 containerd[1461]: time="2025-08-13T07:10:38.164306415Z" level=info msg="Stop container \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" with signal terminated" Aug 13 07:10:38.171414 systemd-networkd[1382]: lxc_health: Link DOWN Aug 13 07:10:38.171425 systemd-networkd[1382]: lxc_health: Lost carrier Aug 13 07:10:38.176062 containerd[1461]: time="2025-08-13T07:10:38.173436846Z" level=info msg="shim disconnected" id=04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7 namespace=k8s.io Aug 13 07:10:38.176062 containerd[1461]: time="2025-08-13T07:10:38.173483746Z" level=warning msg="cleaning up after shim disconnected" id=04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7 namespace=k8s.io Aug 13 07:10:38.176062 containerd[1461]: time="2025-08-13T07:10:38.173495319Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:38.196878 containerd[1461]: time="2025-08-13T07:10:38.196807518Z" level=info msg="StopContainer for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" returns successfully" Aug 13 07:10:38.197761 containerd[1461]: time="2025-08-13T07:10:38.197715236Z" level=info msg="StopPodSandbox for \"6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026\"" Aug 13 07:10:38.197818 containerd[1461]: time="2025-08-13T07:10:38.197791402Z" level=info msg="Container to stop \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:10:38.200536 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026-shm.mount: Deactivated successfully. Aug 13 07:10:38.201622 systemd[1]: cri-containerd-cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8.scope: Deactivated successfully. Aug 13 07:10:38.202093 systemd[1]: cri-containerd-cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8.scope: Consumed 6.957s CPU time. Aug 13 07:10:38.207246 systemd[1]: cri-containerd-6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026.scope: Deactivated successfully. Aug 13 07:10:38.224203 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8-rootfs.mount: Deactivated successfully. Aug 13 07:10:38.232952 containerd[1461]: time="2025-08-13T07:10:38.232719810Z" level=info msg="shim disconnected" id=cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8 namespace=k8s.io Aug 13 07:10:38.232952 containerd[1461]: time="2025-08-13T07:10:38.232774134Z" level=warning msg="cleaning up after shim disconnected" id=cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8 namespace=k8s.io Aug 13 07:10:38.232952 containerd[1461]: time="2025-08-13T07:10:38.232782730Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:38.233934 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026-rootfs.mount: Deactivated successfully. Aug 13 07:10:38.234897 containerd[1461]: time="2025-08-13T07:10:38.234791695Z" level=info msg="shim disconnected" id=6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026 namespace=k8s.io Aug 13 07:10:38.235107 containerd[1461]: time="2025-08-13T07:10:38.235073484Z" level=warning msg="cleaning up after shim disconnected" id=6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026 namespace=k8s.io Aug 13 07:10:38.235107 containerd[1461]: time="2025-08-13T07:10:38.235097770Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:38.251220 containerd[1461]: time="2025-08-13T07:10:38.251153234Z" level=info msg="StopContainer for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" returns successfully" Aug 13 07:10:38.251935 containerd[1461]: time="2025-08-13T07:10:38.251874133Z" level=info msg="StopPodSandbox for \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\"" Aug 13 07:10:38.251935 containerd[1461]: time="2025-08-13T07:10:38.251934269Z" level=info msg="Container to stop \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:10:38.252096 containerd[1461]: time="2025-08-13T07:10:38.251948576Z" level=info msg="Container to stop \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:10:38.252096 containerd[1461]: time="2025-08-13T07:10:38.251959578Z" level=info msg="Container to stop \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:10:38.252096 containerd[1461]: time="2025-08-13T07:10:38.251969687Z" level=info msg="Container to stop \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:10:38.252096 containerd[1461]: time="2025-08-13T07:10:38.251979124Z" level=info msg="Container to stop \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:10:38.258468 systemd[1]: cri-containerd-e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33.scope: Deactivated successfully. Aug 13 07:10:38.260809 containerd[1461]: time="2025-08-13T07:10:38.260740280Z" level=info msg="TearDown network for sandbox \"6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026\" successfully" Aug 13 07:10:38.260809 containerd[1461]: time="2025-08-13T07:10:38.260783132Z" level=info msg="StopPodSandbox for \"6828d68dc1ec9bbd0442e7363d3ea0fc3819ac7028af4d195b9a759b29e99026\" returns successfully" Aug 13 07:10:38.284882 containerd[1461]: time="2025-08-13T07:10:38.284669190Z" level=info msg="shim disconnected" id=e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33 namespace=k8s.io Aug 13 07:10:38.284882 containerd[1461]: time="2025-08-13T07:10:38.284730848Z" level=warning msg="cleaning up after shim disconnected" id=e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33 namespace=k8s.io Aug 13 07:10:38.284882 containerd[1461]: time="2025-08-13T07:10:38.284740497Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:38.299812 kubelet[2519]: I0813 07:10:38.299759 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqq2g\" (UniqueName: \"kubernetes.io/projected/ece56966-6890-43be-8d79-762183dde354-kube-api-access-rqq2g\") pod \"ece56966-6890-43be-8d79-762183dde354\" (UID: \"ece56966-6890-43be-8d79-762183dde354\") " Aug 13 07:10:38.300982 kubelet[2519]: I0813 07:10:38.299865 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ece56966-6890-43be-8d79-762183dde354-cilium-config-path\") pod \"ece56966-6890-43be-8d79-762183dde354\" (UID: \"ece56966-6890-43be-8d79-762183dde354\") " Aug 13 07:10:38.301173 containerd[1461]: time="2025-08-13T07:10:38.301125430Z" level=info msg="TearDown network for sandbox \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" successfully" Aug 13 07:10:38.301274 containerd[1461]: time="2025-08-13T07:10:38.301243607Z" level=info msg="StopPodSandbox for \"e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33\" returns successfully" Aug 13 07:10:38.303479 kubelet[2519]: I0813 07:10:38.303449 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece56966-6890-43be-8d79-762183dde354-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "ece56966-6890-43be-8d79-762183dde354" (UID: "ece56966-6890-43be-8d79-762183dde354"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Aug 13 07:10:38.303557 kubelet[2519]: I0813 07:10:38.303521 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece56966-6890-43be-8d79-762183dde354-kube-api-access-rqq2g" (OuterVolumeSpecName: "kube-api-access-rqq2g") pod "ece56966-6890-43be-8d79-762183dde354" (UID: "ece56966-6890-43be-8d79-762183dde354"). InnerVolumeSpecName "kube-api-access-rqq2g". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 07:10:38.324168 kubelet[2519]: I0813 07:10:38.323744 2519 scope.go:117] "RemoveContainer" containerID="04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7" Aug 13 07:10:38.326289 containerd[1461]: time="2025-08-13T07:10:38.326238577Z" level=info msg="RemoveContainer for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\"" Aug 13 07:10:38.330532 systemd[1]: Removed slice kubepods-besteffort-podece56966_6890_43be_8d79_762183dde354.slice - libcontainer container kubepods-besteffort-podece56966_6890_43be_8d79_762183dde354.slice. Aug 13 07:10:38.338249 containerd[1461]: time="2025-08-13T07:10:38.338211468Z" level=info msg="RemoveContainer for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" returns successfully" Aug 13 07:10:38.338536 kubelet[2519]: I0813 07:10:38.338504 2519 scope.go:117] "RemoveContainer" containerID="04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7" Aug 13 07:10:38.342706 containerd[1461]: time="2025-08-13T07:10:38.342645434Z" level=error msg="ContainerStatus for \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\": not found" Aug 13 07:10:38.342908 kubelet[2519]: E0813 07:10:38.342871 2519 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\": not found" containerID="04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7" Aug 13 07:10:38.342954 kubelet[2519]: I0813 07:10:38.342920 2519 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7"} err="failed to get container status \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\": rpc error: code = NotFound desc = an error occurred when try to find container \"04b00395ef8d4430227a9d204448731e1f629af49fbd7b45ceb4d6cfdc2822c7\": not found" Aug 13 07:10:38.342981 kubelet[2519]: I0813 07:10:38.342963 2519 scope.go:117] "RemoveContainer" containerID="cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8" Aug 13 07:10:38.344027 containerd[1461]: time="2025-08-13T07:10:38.343992010Z" level=info msg="RemoveContainer for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\"" Aug 13 07:10:38.347785 containerd[1461]: time="2025-08-13T07:10:38.347750462Z" level=info msg="RemoveContainer for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" returns successfully" Aug 13 07:10:38.347969 kubelet[2519]: I0813 07:10:38.347936 2519 scope.go:117] "RemoveContainer" containerID="6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be" Aug 13 07:10:38.348774 containerd[1461]: time="2025-08-13T07:10:38.348740306Z" level=info msg="RemoveContainer for \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\"" Aug 13 07:10:38.351898 containerd[1461]: time="2025-08-13T07:10:38.351868332Z" level=info msg="RemoveContainer for \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\" returns successfully" Aug 13 07:10:38.352055 kubelet[2519]: I0813 07:10:38.352019 2519 scope.go:117] "RemoveContainer" containerID="8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba" Aug 13 07:10:38.352950 containerd[1461]: time="2025-08-13T07:10:38.352915315Z" level=info msg="RemoveContainer for \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\"" Aug 13 07:10:38.356342 containerd[1461]: time="2025-08-13T07:10:38.356262891Z" level=info msg="RemoveContainer for \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\" returns successfully" Aug 13 07:10:38.356479 kubelet[2519]: I0813 07:10:38.356444 2519 scope.go:117] "RemoveContainer" containerID="8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77" Aug 13 07:10:38.357744 containerd[1461]: time="2025-08-13T07:10:38.357694361Z" level=info msg="RemoveContainer for \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\"" Aug 13 07:10:38.360858 containerd[1461]: time="2025-08-13T07:10:38.360811154Z" level=info msg="RemoveContainer for \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\" returns successfully" Aug 13 07:10:38.360980 kubelet[2519]: I0813 07:10:38.360960 2519 scope.go:117] "RemoveContainer" containerID="bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563" Aug 13 07:10:38.361727 containerd[1461]: time="2025-08-13T07:10:38.361704204Z" level=info msg="RemoveContainer for \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\"" Aug 13 07:10:38.364749 containerd[1461]: time="2025-08-13T07:10:38.364722219Z" level=info msg="RemoveContainer for \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\" returns successfully" Aug 13 07:10:38.364936 kubelet[2519]: I0813 07:10:38.364899 2519 scope.go:117] "RemoveContainer" containerID="cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8" Aug 13 07:10:38.365133 containerd[1461]: time="2025-08-13T07:10:38.365098880Z" level=error msg="ContainerStatus for \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\": not found" Aug 13 07:10:38.365300 kubelet[2519]: E0813 07:10:38.365257 2519 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\": not found" containerID="cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8" Aug 13 07:10:38.365339 kubelet[2519]: I0813 07:10:38.365302 2519 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8"} err="failed to get container status \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\": rpc error: code = NotFound desc = an error occurred when try to find container \"cebe591ac5f8f5e769423bd6a2c7ea975ba24bd9038f50424d9ac3ced20fbeb8\": not found" Aug 13 07:10:38.365339 kubelet[2519]: I0813 07:10:38.365333 2519 scope.go:117] "RemoveContainer" containerID="6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be" Aug 13 07:10:38.365575 containerd[1461]: time="2025-08-13T07:10:38.365533742Z" level=error msg="ContainerStatus for \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\": not found" Aug 13 07:10:38.365709 kubelet[2519]: E0813 07:10:38.365670 2519 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\": not found" containerID="6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be" Aug 13 07:10:38.365709 kubelet[2519]: I0813 07:10:38.365697 2519 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be"} err="failed to get container status \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\": rpc error: code = NotFound desc = an error occurred when try to find container \"6fef3e1821e5b74cdb54866b7b7c5d0aba75fee1a201c17c92f0feeece81a1be\": not found" Aug 13 07:10:38.365780 kubelet[2519]: I0813 07:10:38.365717 2519 scope.go:117] "RemoveContainer" containerID="8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba" Aug 13 07:10:38.365932 containerd[1461]: time="2025-08-13T07:10:38.365886286Z" level=error msg="ContainerStatus for \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\": not found" Aug 13 07:10:38.366044 kubelet[2519]: E0813 07:10:38.366021 2519 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\": not found" containerID="8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba" Aug 13 07:10:38.366075 kubelet[2519]: I0813 07:10:38.366050 2519 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba"} err="failed to get container status \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\": rpc error: code = NotFound desc = an error occurred when try to find container \"8cf83d8df01cfbe4b155afdd93633f6a0f68648e59b116850d67e979a10436ba\": not found" Aug 13 07:10:38.366075 kubelet[2519]: I0813 07:10:38.366069 2519 scope.go:117] "RemoveContainer" containerID="8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77" Aug 13 07:10:38.366240 containerd[1461]: time="2025-08-13T07:10:38.366210647Z" level=error msg="ContainerStatus for \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\": not found" Aug 13 07:10:38.366309 kubelet[2519]: E0813 07:10:38.366285 2519 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\": not found" containerID="8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77" Aug 13 07:10:38.366373 kubelet[2519]: I0813 07:10:38.366306 2519 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77"} err="failed to get container status \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\": rpc error: code = NotFound desc = an error occurred when try to find container \"8261530bf4a7ba30144362c856993aece7b39d80ae84201a28f3b1a3d261bc77\": not found" Aug 13 07:10:38.366373 kubelet[2519]: I0813 07:10:38.366321 2519 scope.go:117] "RemoveContainer" containerID="bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563" Aug 13 07:10:38.366535 containerd[1461]: time="2025-08-13T07:10:38.366500011Z" level=error msg="ContainerStatus for \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\": not found" Aug 13 07:10:38.366654 kubelet[2519]: E0813 07:10:38.366633 2519 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\": not found" containerID="bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563" Aug 13 07:10:38.366684 kubelet[2519]: I0813 07:10:38.366659 2519 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563"} err="failed to get container status \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\": rpc error: code = NotFound desc = an error occurred when try to find container \"bcb274570c261308c2c8ffe56052d1a8fe6379cdea3d9c0a6c2c0e4f71846563\": not found" Aug 13 07:10:38.400105 kubelet[2519]: I0813 07:10:38.400055 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-bpf-maps\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400105 kubelet[2519]: I0813 07:10:38.400090 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mksb5\" (UniqueName: \"kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-kube-api-access-mksb5\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400263 kubelet[2519]: I0813 07:10:38.400139 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0f16dfbc-754e-4f6b-963a-7766b242d9d2-clustermesh-secrets\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400263 kubelet[2519]: I0813 07:10:38.400155 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-run\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400263 kubelet[2519]: I0813 07:10:38.400175 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-config-path\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400263 kubelet[2519]: I0813 07:10:38.400190 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-net\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400263 kubelet[2519]: I0813 07:10:38.400204 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-etc-cni-netd\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400263 kubelet[2519]: I0813 07:10:38.400219 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-xtables-lock\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400401 kubelet[2519]: I0813 07:10:38.400215 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400401 kubelet[2519]: I0813 07:10:38.400253 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400401 kubelet[2519]: I0813 07:10:38.400232 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-lib-modules\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400401 kubelet[2519]: I0813 07:10:38.400277 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400401 kubelet[2519]: I0813 07:10:38.400298 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-kernel\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400326 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-cgroup\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400347 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cni-path\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400370 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hubble-tls\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400384 2519 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hostproc\") pod \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\" (UID: \"0f16dfbc-754e-4f6b-963a-7766b242d9d2\") " Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400427 2519 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-bpf-maps\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400442 2519 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rqq2g\" (UniqueName: \"kubernetes.io/projected/ece56966-6890-43be-8d79-762183dde354-kube-api-access-rqq2g\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.400518 kubelet[2519]: I0813 07:10:38.400453 2519 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-lib-modules\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.400674 kubelet[2519]: I0813 07:10:38.400462 2519 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ece56966-6890-43be-8d79-762183dde354-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.400674 kubelet[2519]: I0813 07:10:38.400482 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hostproc" (OuterVolumeSpecName: "hostproc") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400674 kubelet[2519]: I0813 07:10:38.400499 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400674 kubelet[2519]: I0813 07:10:38.400515 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400674 kubelet[2519]: I0813 07:10:38.400538 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cni-path" (OuterVolumeSpecName: "cni-path") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400923 kubelet[2519]: I0813 07:10:38.400691 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400923 kubelet[2519]: I0813 07:10:38.400716 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.400923 kubelet[2519]: I0813 07:10:38.400732 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Aug 13 07:10:38.403773 kubelet[2519]: I0813 07:10:38.403725 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Aug 13 07:10:38.403949 kubelet[2519]: I0813 07:10:38.403912 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 07:10:38.404271 kubelet[2519]: I0813 07:10:38.404230 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f16dfbc-754e-4f6b-963a-7766b242d9d2-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Aug 13 07:10:38.404408 kubelet[2519]: I0813 07:10:38.404283 2519 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-kube-api-access-mksb5" (OuterVolumeSpecName: "kube-api-access-mksb5") pod "0f16dfbc-754e-4f6b-963a-7766b242d9d2" (UID: "0f16dfbc-754e-4f6b-963a-7766b242d9d2"). InnerVolumeSpecName "kube-api-access-mksb5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 07:10:38.501088 kubelet[2519]: I0813 07:10:38.501047 2519 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mksb5\" (UniqueName: \"kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-kube-api-access-mksb5\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501088 kubelet[2519]: I0813 07:10:38.501083 2519 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0f16dfbc-754e-4f6b-963a-7766b242d9d2-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501088 kubelet[2519]: I0813 07:10:38.501093 2519 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-run\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501088 kubelet[2519]: I0813 07:10:38.501101 2519 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501109 2519 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501119 2519 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501127 2519 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-xtables-lock\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501135 2519 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501144 2519 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501152 2519 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-cni-path\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501160 2519 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hubble-tls\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.501275 kubelet[2519]: I0813 07:10:38.501168 2519 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0f16dfbc-754e-4f6b-963a-7766b242d9d2-hostproc\") on node \"localhost\" DevicePath \"\"" Aug 13 07:10:38.635686 systemd[1]: Removed slice kubepods-burstable-pod0f16dfbc_754e_4f6b_963a_7766b242d9d2.slice - libcontainer container kubepods-burstable-pod0f16dfbc_754e_4f6b_963a_7766b242d9d2.slice. Aug 13 07:10:38.636066 systemd[1]: kubepods-burstable-pod0f16dfbc_754e_4f6b_963a_7766b242d9d2.slice: Consumed 7.065s CPU time. Aug 13 07:10:39.043453 kubelet[2519]: E0813 07:10:39.043395 2519 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Aug 13 07:10:39.138040 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33-rootfs.mount: Deactivated successfully. Aug 13 07:10:39.138202 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e95c02a3770d75c9f1c7cce934ec02954efe1d113baba735629c219fb8ab9f33-shm.mount: Deactivated successfully. Aug 13 07:10:39.138319 systemd[1]: var-lib-kubelet-pods-ece56966\x2d6890\x2d43be\x2d8d79\x2d762183dde354-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2drqq2g.mount: Deactivated successfully. Aug 13 07:10:39.138441 systemd[1]: var-lib-kubelet-pods-0f16dfbc\x2d754e\x2d4f6b\x2d963a\x2d7766b242d9d2-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dmksb5.mount: Deactivated successfully. Aug 13 07:10:39.138549 systemd[1]: var-lib-kubelet-pods-0f16dfbc\x2d754e\x2d4f6b\x2d963a\x2d7766b242d9d2-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Aug 13 07:10:39.138660 systemd[1]: var-lib-kubelet-pods-0f16dfbc\x2d754e\x2d4f6b\x2d963a\x2d7766b242d9d2-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Aug 13 07:10:39.524173 sshd[4195]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:39.536335 systemd[1]: sshd@24-10.0.0.69:22-10.0.0.1:52048.service: Deactivated successfully. Aug 13 07:10:39.538355 systemd[1]: session-25.scope: Deactivated successfully. Aug 13 07:10:39.540148 systemd-logind[1450]: Session 25 logged out. Waiting for processes to exit. Aug 13 07:10:39.549367 systemd[1]: Started sshd@25-10.0.0.69:22-10.0.0.1:44996.service - OpenSSH per-connection server daemon (10.0.0.1:44996). Aug 13 07:10:39.550553 systemd-logind[1450]: Removed session 25. Aug 13 07:10:39.581389 sshd[4355]: Accepted publickey for core from 10.0.0.1 port 44996 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:39.583066 sshd[4355]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:39.587599 systemd-logind[1450]: New session 26 of user core. Aug 13 07:10:39.598073 systemd[1]: Started session-26.scope - Session 26 of User core. Aug 13 07:10:39.998582 kubelet[2519]: I0813 07:10:39.998510 2519 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f16dfbc-754e-4f6b-963a-7766b242d9d2" path="/var/lib/kubelet/pods/0f16dfbc-754e-4f6b-963a-7766b242d9d2/volumes" Aug 13 07:10:39.999717 kubelet[2519]: I0813 07:10:39.999688 2519 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece56966-6890-43be-8d79-762183dde354" path="/var/lib/kubelet/pods/ece56966-6890-43be-8d79-762183dde354/volumes" Aug 13 07:10:40.053041 sshd[4355]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:40.070041 systemd[1]: sshd@25-10.0.0.69:22-10.0.0.1:44996.service: Deactivated successfully. Aug 13 07:10:40.073522 systemd[1]: session-26.scope: Deactivated successfully. Aug 13 07:10:40.078181 systemd-logind[1450]: Session 26 logged out. Waiting for processes to exit. Aug 13 07:10:40.089333 systemd[1]: Started sshd@26-10.0.0.69:22-10.0.0.1:45008.service - OpenSSH per-connection server daemon (10.0.0.1:45008). Aug 13 07:10:40.093022 systemd-logind[1450]: Removed session 26. Aug 13 07:10:40.099332 systemd[1]: Created slice kubepods-burstable-pod34062e73_061f_468b_a48c_6d92de1be12b.slice - libcontainer container kubepods-burstable-pod34062e73_061f_468b_a48c_6d92de1be12b.slice. Aug 13 07:10:40.112606 kubelet[2519]: I0813 07:10:40.111646 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qwx2\" (UniqueName: \"kubernetes.io/projected/34062e73-061f-468b-a48c-6d92de1be12b-kube-api-access-9qwx2\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112606 kubelet[2519]: I0813 07:10:40.111722 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/34062e73-061f-468b-a48c-6d92de1be12b-clustermesh-secrets\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112606 kubelet[2519]: I0813 07:10:40.111741 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-host-proc-sys-kernel\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112606 kubelet[2519]: I0813 07:10:40.111754 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-lib-modules\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112606 kubelet[2519]: I0813 07:10:40.111771 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/34062e73-061f-468b-a48c-6d92de1be12b-cilium-ipsec-secrets\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112894 kubelet[2519]: I0813 07:10:40.111785 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-xtables-lock\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112894 kubelet[2519]: I0813 07:10:40.111798 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-bpf-maps\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112894 kubelet[2519]: I0813 07:10:40.111811 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-cilium-run\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112894 kubelet[2519]: I0813 07:10:40.111823 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-hostproc\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112894 kubelet[2519]: I0813 07:10:40.111836 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-cilium-cgroup\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.112894 kubelet[2519]: I0813 07:10:40.111866 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-cni-path\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.113044 kubelet[2519]: I0813 07:10:40.111880 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/34062e73-061f-468b-a48c-6d92de1be12b-cilium-config-path\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.113044 kubelet[2519]: I0813 07:10:40.111895 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/34062e73-061f-468b-a48c-6d92de1be12b-hubble-tls\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.113044 kubelet[2519]: I0813 07:10:40.111929 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-etc-cni-netd\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.113044 kubelet[2519]: I0813 07:10:40.111942 2519 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/34062e73-061f-468b-a48c-6d92de1be12b-host-proc-sys-net\") pod \"cilium-c7xv4\" (UID: \"34062e73-061f-468b-a48c-6d92de1be12b\") " pod="kube-system/cilium-c7xv4" Aug 13 07:10:40.131628 sshd[4368]: Accepted publickey for core from 10.0.0.1 port 45008 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:40.136837 sshd[4368]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:40.142080 systemd-logind[1450]: New session 27 of user core. Aug 13 07:10:40.151274 systemd[1]: Started session-27.scope - Session 27 of User core. Aug 13 07:10:40.204490 sshd[4368]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:40.229751 systemd[1]: sshd@26-10.0.0.69:22-10.0.0.1:45008.service: Deactivated successfully. Aug 13 07:10:40.231822 systemd[1]: session-27.scope: Deactivated successfully. Aug 13 07:10:40.233445 systemd-logind[1450]: Session 27 logged out. Waiting for processes to exit. Aug 13 07:10:40.246181 systemd[1]: Started sshd@27-10.0.0.69:22-10.0.0.1:45020.service - OpenSSH per-connection server daemon (10.0.0.1:45020). Aug 13 07:10:40.247048 systemd-logind[1450]: Removed session 27. Aug 13 07:10:40.279196 sshd[4380]: Accepted publickey for core from 10.0.0.1 port 45020 ssh2: RSA SHA256:CMfoLhPNmBOOiskIU7y9xMX9q9TU1tPTT3rYgwbB2Y8 Aug 13 07:10:40.281748 sshd[4380]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:10:40.287186 systemd-logind[1450]: New session 28 of user core. Aug 13 07:10:40.292058 systemd[1]: Started session-28.scope - Session 28 of User core. Aug 13 07:10:40.404179 kubelet[2519]: E0813 07:10:40.404077 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:40.404981 containerd[1461]: time="2025-08-13T07:10:40.404930662Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c7xv4,Uid:34062e73-061f-468b-a48c-6d92de1be12b,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:40.429457 containerd[1461]: time="2025-08-13T07:10:40.429316170Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:40.429457 containerd[1461]: time="2025-08-13T07:10:40.429423385Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:40.429457 containerd[1461]: time="2025-08-13T07:10:40.429442791Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:40.429630 containerd[1461]: time="2025-08-13T07:10:40.429595032Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:40.453193 systemd[1]: Started cri-containerd-9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba.scope - libcontainer container 9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba. Aug 13 07:10:40.477205 containerd[1461]: time="2025-08-13T07:10:40.477152717Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c7xv4,Uid:34062e73-061f-468b-a48c-6d92de1be12b,Namespace:kube-system,Attempt:0,} returns sandbox id \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\"" Aug 13 07:10:40.478177 kubelet[2519]: E0813 07:10:40.478146 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:40.484399 containerd[1461]: time="2025-08-13T07:10:40.484350126Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 07:10:40.497178 containerd[1461]: time="2025-08-13T07:10:40.497117533Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8\"" Aug 13 07:10:40.497859 containerd[1461]: time="2025-08-13T07:10:40.497784288Z" level=info msg="StartContainer for \"a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8\"" Aug 13 07:10:40.522984 systemd[1]: Started cri-containerd-a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8.scope - libcontainer container a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8. Aug 13 07:10:40.560538 containerd[1461]: time="2025-08-13T07:10:40.560412923Z" level=info msg="StartContainer for \"a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8\" returns successfully" Aug 13 07:10:40.571409 systemd[1]: cri-containerd-a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8.scope: Deactivated successfully. Aug 13 07:10:40.606254 containerd[1461]: time="2025-08-13T07:10:40.606161870Z" level=info msg="shim disconnected" id=a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8 namespace=k8s.io Aug 13 07:10:40.606254 containerd[1461]: time="2025-08-13T07:10:40.606238325Z" level=warning msg="cleaning up after shim disconnected" id=a6da604a9074a900fde59f1f467865d0b5f3ce7f612d349f1c9ff4f0cfd63ba8 namespace=k8s.io Aug 13 07:10:40.606254 containerd[1461]: time="2025-08-13T07:10:40.606251691Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:41.339597 kubelet[2519]: E0813 07:10:41.339551 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:41.345530 containerd[1461]: time="2025-08-13T07:10:41.344508147Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Aug 13 07:10:41.359076 containerd[1461]: time="2025-08-13T07:10:41.359017340Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74\"" Aug 13 07:10:41.359772 containerd[1461]: time="2025-08-13T07:10:41.359606746Z" level=info msg="StartContainer for \"742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74\"" Aug 13 07:10:41.395078 systemd[1]: Started cri-containerd-742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74.scope - libcontainer container 742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74. Aug 13 07:10:41.425455 containerd[1461]: time="2025-08-13T07:10:41.424618578Z" level=info msg="StartContainer for \"742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74\" returns successfully" Aug 13 07:10:41.432518 systemd[1]: cri-containerd-742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74.scope: Deactivated successfully. Aug 13 07:10:41.459555 containerd[1461]: time="2025-08-13T07:10:41.459474478Z" level=info msg="shim disconnected" id=742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74 namespace=k8s.io Aug 13 07:10:41.459555 containerd[1461]: time="2025-08-13T07:10:41.459547127Z" level=warning msg="cleaning up after shim disconnected" id=742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74 namespace=k8s.io Aug 13 07:10:41.459555 containerd[1461]: time="2025-08-13T07:10:41.459558799Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:42.221041 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-742b7bbfafad2b724a454ed1cd540cef911d523f972f887325feeca85bed2e74-rootfs.mount: Deactivated successfully. Aug 13 07:10:42.343019 kubelet[2519]: E0813 07:10:42.342979 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:42.348196 containerd[1461]: time="2025-08-13T07:10:42.348138762Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Aug 13 07:10:42.364527 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3039232163.mount: Deactivated successfully. Aug 13 07:10:42.366926 containerd[1461]: time="2025-08-13T07:10:42.366840420Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328\"" Aug 13 07:10:42.367556 containerd[1461]: time="2025-08-13T07:10:42.367526931Z" level=info msg="StartContainer for \"fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328\"" Aug 13 07:10:42.420339 systemd[1]: Started cri-containerd-fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328.scope - libcontainer container fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328. Aug 13 07:10:42.453791 containerd[1461]: time="2025-08-13T07:10:42.453736730Z" level=info msg="StartContainer for \"fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328\" returns successfully" Aug 13 07:10:42.455942 systemd[1]: cri-containerd-fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328.scope: Deactivated successfully. Aug 13 07:10:42.483340 containerd[1461]: time="2025-08-13T07:10:42.483189009Z" level=info msg="shim disconnected" id=fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328 namespace=k8s.io Aug 13 07:10:42.483340 containerd[1461]: time="2025-08-13T07:10:42.483260116Z" level=warning msg="cleaning up after shim disconnected" id=fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328 namespace=k8s.io Aug 13 07:10:42.483340 containerd[1461]: time="2025-08-13T07:10:42.483272529Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:43.221430 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fdfef874849afa6a0883bf88df1902b5f4bcc77b0e0bc60f3e68b69de26ec328-rootfs.mount: Deactivated successfully. Aug 13 07:10:43.347697 kubelet[2519]: E0813 07:10:43.347645 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:43.353881 containerd[1461]: time="2025-08-13T07:10:43.353509084Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Aug 13 07:10:43.368245 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount233993971.mount: Deactivated successfully. Aug 13 07:10:43.377349 containerd[1461]: time="2025-08-13T07:10:43.377288007Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9\"" Aug 13 07:10:43.377958 containerd[1461]: time="2025-08-13T07:10:43.377917098Z" level=info msg="StartContainer for \"e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9\"" Aug 13 07:10:43.410990 systemd[1]: Started cri-containerd-e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9.scope - libcontainer container e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9. Aug 13 07:10:43.434812 systemd[1]: cri-containerd-e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9.scope: Deactivated successfully. Aug 13 07:10:43.439382 containerd[1461]: time="2025-08-13T07:10:43.439335967Z" level=info msg="StartContainer for \"e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9\" returns successfully" Aug 13 07:10:43.445028 containerd[1461]: time="2025-08-13T07:10:43.441230273Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34062e73_061f_468b_a48c_6d92de1be12b.slice/cri-containerd-e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9.scope/memory.events\": no such file or directory" Aug 13 07:10:43.471454 containerd[1461]: time="2025-08-13T07:10:43.471376813Z" level=info msg="shim disconnected" id=e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9 namespace=k8s.io Aug 13 07:10:43.471454 containerd[1461]: time="2025-08-13T07:10:43.471445684Z" level=warning msg="cleaning up after shim disconnected" id=e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9 namespace=k8s.io Aug 13 07:10:43.471454 containerd[1461]: time="2025-08-13T07:10:43.471454190Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:44.044730 kubelet[2519]: E0813 07:10:44.044673 2519 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Aug 13 07:10:44.221171 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e65902062ce0f761ee59d24cb92498f12242ecacb212d8eaad33852d6395e4c9-rootfs.mount: Deactivated successfully. Aug 13 07:10:44.352884 kubelet[2519]: E0813 07:10:44.352714 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:44.379395 containerd[1461]: time="2025-08-13T07:10:44.379316880Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Aug 13 07:10:44.395091 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3595277431.mount: Deactivated successfully. Aug 13 07:10:44.398661 containerd[1461]: time="2025-08-13T07:10:44.398615084Z" level=info msg="CreateContainer within sandbox \"9abab99c51df5d5037b3b4ffcd777e28525a727b8baefe21ef2f0ed67cd21dba\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"51e47aec15ee58859c95d6e9eb39c1bccb9914a0be77c223801de4e17d6470be\"" Aug 13 07:10:44.399196 containerd[1461]: time="2025-08-13T07:10:44.399161648Z" level=info msg="StartContainer for \"51e47aec15ee58859c95d6e9eb39c1bccb9914a0be77c223801de4e17d6470be\"" Aug 13 07:10:44.440062 systemd[1]: Started cri-containerd-51e47aec15ee58859c95d6e9eb39c1bccb9914a0be77c223801de4e17d6470be.scope - libcontainer container 51e47aec15ee58859c95d6e9eb39c1bccb9914a0be77c223801de4e17d6470be. Aug 13 07:10:44.480202 containerd[1461]: time="2025-08-13T07:10:44.480150738Z" level=info msg="StartContainer for \"51e47aec15ee58859c95d6e9eb39c1bccb9914a0be77c223801de4e17d6470be\" returns successfully" Aug 13 07:10:44.927935 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Aug 13 07:10:45.357909 kubelet[2519]: E0813 07:10:45.357716 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:45.379159 kubelet[2519]: I0813 07:10:45.379083 2519 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-c7xv4" podStartSLOduration=5.379059594 podStartE2EDuration="5.379059594s" podCreationTimestamp="2025-08-13 07:10:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:45.378563979 +0000 UTC m=+91.484253511" watchObservedRunningTime="2025-08-13 07:10:45.379059594 +0000 UTC m=+91.484749126" Aug 13 07:10:46.140267 kubelet[2519]: I0813 07:10:46.139509 2519 setters.go:618] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-08-13T07:10:46Z","lastTransitionTime":"2025-08-13T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Aug 13 07:10:46.405460 kubelet[2519]: E0813 07:10:46.405422 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:46.995901 kubelet[2519]: E0813 07:10:46.995802 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:48.237626 systemd-networkd[1382]: lxc_health: Link UP Aug 13 07:10:48.244380 systemd-networkd[1382]: lxc_health: Gained carrier Aug 13 07:10:48.406368 kubelet[2519]: E0813 07:10:48.405821 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:49.365569 kubelet[2519]: E0813 07:10:49.365529 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:50.218245 systemd-networkd[1382]: lxc_health: Gained IPv6LL Aug 13 07:10:50.366870 kubelet[2519]: E0813 07:10:50.366759 2519 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 07:10:51.215981 systemd[1]: run-containerd-runc-k8s.io-51e47aec15ee58859c95d6e9eb39c1bccb9914a0be77c223801de4e17d6470be-runc.sZcOMw.mount: Deactivated successfully. Aug 13 07:10:55.527312 sshd[4380]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:55.531232 systemd[1]: sshd@27-10.0.0.69:22-10.0.0.1:45020.service: Deactivated successfully. Aug 13 07:10:55.533147 systemd[1]: session-28.scope: Deactivated successfully. Aug 13 07:10:55.533764 systemd-logind[1450]: Session 28 logged out. Waiting for processes to exit. Aug 13 07:10:55.534818 systemd-logind[1450]: Removed session 28.