Apr 13 23:44:40.377266 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Apr 13 18:40:27 -00 2026 Apr 13 23:44:40.377296 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c1ba97db2f6278922cfc5bd0ca74b4bb573fca2c3aed19c121a34271e693e156 Apr 13 23:44:40.377310 kernel: BIOS-provided physical RAM map: Apr 13 23:44:40.377318 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Apr 13 23:44:40.377326 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Apr 13 23:44:40.377334 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Apr 13 23:44:40.377344 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Apr 13 23:44:40.377351 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Apr 13 23:44:40.377359 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable Apr 13 23:44:40.377366 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS Apr 13 23:44:40.377376 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable Apr 13 23:44:40.377384 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009c9eefff] reserved Apr 13 23:44:40.377391 kernel: BIOS-e820: [mem 0x000000009c9ef000-0x000000009caeefff] type 20 Apr 13 23:44:40.377399 kernel: BIOS-e820: [mem 0x000000009caef000-0x000000009cb6efff] reserved Apr 13 23:44:40.377407 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data Apr 13 23:44:40.377415 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Apr 13 23:44:40.377425 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable Apr 13 23:44:40.377432 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved Apr 13 23:44:40.377439 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Apr 13 23:44:40.377445 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Apr 13 23:44:40.377452 kernel: NX (Execute Disable) protection: active Apr 13 23:44:40.377459 kernel: APIC: Static calls initialized Apr 13 23:44:40.377467 kernel: efi: EFI v2.7 by EDK II Apr 13 23:44:40.377475 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b674118 Apr 13 23:44:40.377482 kernel: SMBIOS 2.8 present. Apr 13 23:44:40.377489 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 0.0.0 02/06/2015 Apr 13 23:44:40.377496 kernel: Hypervisor detected: KVM Apr 13 23:44:40.377505 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Apr 13 23:44:40.377512 kernel: kvm-clock: using sched offset of 7367110888 cycles Apr 13 23:44:40.377520 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 13 23:44:40.377528 kernel: tsc: Detected 2793.438 MHz processor Apr 13 23:44:40.377536 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Apr 13 23:44:40.377544 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Apr 13 23:44:40.377551 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x10000000000 Apr 13 23:44:40.377558 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Apr 13 23:44:40.377566 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Apr 13 23:44:40.377575 kernel: Using GB pages for direct mapping Apr 13 23:44:40.377582 kernel: Secure boot disabled Apr 13 23:44:40.377590 kernel: ACPI: Early table checksum verification disabled Apr 13 23:44:40.377598 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Apr 13 23:44:40.377610 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Apr 13 23:44:40.377619 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 13 23:44:40.377628 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 13 23:44:40.377639 kernel: ACPI: FACS 0x000000009CBDD000 000040 Apr 13 23:44:40.377648 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 13 23:44:40.377796 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 13 23:44:40.377805 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 13 23:44:40.377814 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 13 23:44:40.377824 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Apr 13 23:44:40.377832 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Apr 13 23:44:40.377846 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Apr 13 23:44:40.377853 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Apr 13 23:44:40.377861 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Apr 13 23:44:40.377886 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Apr 13 23:44:40.377895 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Apr 13 23:44:40.377904 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Apr 13 23:44:40.377912 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Apr 13 23:44:40.377922 kernel: No NUMA configuration found Apr 13 23:44:40.377930 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] Apr 13 23:44:40.377940 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] Apr 13 23:44:40.377948 kernel: Zone ranges: Apr 13 23:44:40.377956 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Apr 13 23:44:40.377964 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] Apr 13 23:44:40.377971 kernel: Normal empty Apr 13 23:44:40.377979 kernel: Movable zone start for each node Apr 13 23:44:40.377986 kernel: Early memory node ranges Apr 13 23:44:40.377994 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Apr 13 23:44:40.378001 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Apr 13 23:44:40.378009 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Apr 13 23:44:40.378021 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] Apr 13 23:44:40.378029 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] Apr 13 23:44:40.378036 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] Apr 13 23:44:40.378044 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] Apr 13 23:44:40.378052 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 13 23:44:40.378060 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Apr 13 23:44:40.378068 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Apr 13 23:44:40.378076 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 13 23:44:40.378085 kernel: On node 0, zone DMA: 240 pages in unavailable ranges Apr 13 23:44:40.378097 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Apr 13 23:44:40.378107 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges Apr 13 23:44:40.378115 kernel: ACPI: PM-Timer IO Port: 0x608 Apr 13 23:44:40.378124 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Apr 13 23:44:40.378133 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Apr 13 23:44:40.378141 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Apr 13 23:44:40.378150 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Apr 13 23:44:40.378158 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Apr 13 23:44:40.378167 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Apr 13 23:44:40.378178 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Apr 13 23:44:40.378188 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Apr 13 23:44:40.378197 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Apr 13 23:44:40.378206 kernel: TSC deadline timer available Apr 13 23:44:40.378215 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Apr 13 23:44:40.378224 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Apr 13 23:44:40.378233 kernel: kvm-guest: KVM setup pv remote TLB flush Apr 13 23:44:40.378241 kernel: kvm-guest: setup PV sched yield Apr 13 23:44:40.378250 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Apr 13 23:44:40.378258 kernel: Booting paravirtualized kernel on KVM Apr 13 23:44:40.378269 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Apr 13 23:44:40.378278 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Apr 13 23:44:40.378286 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u524288 Apr 13 23:44:40.378295 kernel: pcpu-alloc: s196328 r8192 d28952 u524288 alloc=1*2097152 Apr 13 23:44:40.378303 kernel: pcpu-alloc: [0] 0 1 2 3 Apr 13 23:44:40.378311 kernel: kvm-guest: PV spinlocks enabled Apr 13 23:44:40.378319 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Apr 13 23:44:40.378328 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c1ba97db2f6278922cfc5bd0ca74b4bb573fca2c3aed19c121a34271e693e156 Apr 13 23:44:40.378338 kernel: random: crng init done Apr 13 23:44:40.378346 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 13 23:44:40.378354 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 13 23:44:40.378361 kernel: Fallback order for Node 0: 0 Apr 13 23:44:40.378369 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 Apr 13 23:44:40.378376 kernel: Policy zone: DMA32 Apr 13 23:44:40.378384 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 13 23:44:40.378392 kernel: Memory: 2394676K/2567000K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42896K init, 2300K bss, 172120K reserved, 0K cma-reserved) Apr 13 23:44:40.378401 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Apr 13 23:44:40.378412 kernel: ftrace: allocating 37996 entries in 149 pages Apr 13 23:44:40.378420 kernel: ftrace: allocated 149 pages with 4 groups Apr 13 23:44:40.378428 kernel: Dynamic Preempt: voluntary Apr 13 23:44:40.378436 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 13 23:44:40.378453 kernel: rcu: RCU event tracing is enabled. Apr 13 23:44:40.378464 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Apr 13 23:44:40.378473 kernel: Trampoline variant of Tasks RCU enabled. Apr 13 23:44:40.378481 kernel: Rude variant of Tasks RCU enabled. Apr 13 23:44:40.378490 kernel: Tracing variant of Tasks RCU enabled. Apr 13 23:44:40.378499 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 13 23:44:40.378509 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Apr 13 23:44:40.378519 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Apr 13 23:44:40.378531 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 13 23:44:40.378540 kernel: Console: colour dummy device 80x25 Apr 13 23:44:40.378549 kernel: printk: console [ttyS0] enabled Apr 13 23:44:40.378559 kernel: ACPI: Core revision 20230628 Apr 13 23:44:40.378568 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Apr 13 23:44:40.378580 kernel: APIC: Switch to symmetric I/O mode setup Apr 13 23:44:40.378588 kernel: x2apic enabled Apr 13 23:44:40.378597 kernel: APIC: Switched APIC routing to: physical x2apic Apr 13 23:44:40.378605 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Apr 13 23:44:40.378614 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Apr 13 23:44:40.378623 kernel: kvm-guest: setup PV IPIs Apr 13 23:44:40.378631 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Apr 13 23:44:40.378640 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x284409db922, max_idle_ns: 440795228871 ns Apr 13 23:44:40.378649 kernel: Calibrating delay loop (skipped) preset value.. 5586.87 BogoMIPS (lpj=2793438) Apr 13 23:44:40.378793 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Apr 13 23:44:40.378803 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Apr 13 23:44:40.378813 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Apr 13 23:44:40.378823 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Apr 13 23:44:40.378833 kernel: Spectre V2 : Mitigation: Retpolines Apr 13 23:44:40.378843 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Apr 13 23:44:40.378852 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Apr 13 23:44:40.378861 kernel: RETBleed: Vulnerable Apr 13 23:44:40.379640 kernel: Speculative Store Bypass: Vulnerable Apr 13 23:44:40.379680 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Apr 13 23:44:40.379691 kernel: GDS: Unknown: Dependent on hypervisor status Apr 13 23:44:40.380205 kernel: active return thunk: its_return_thunk Apr 13 23:44:40.380241 kernel: ITS: Mitigation: Aligned branch/return thunks Apr 13 23:44:40.380254 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Apr 13 23:44:40.380265 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Apr 13 23:44:40.380276 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Apr 13 23:44:40.380286 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Apr 13 23:44:40.380308 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Apr 13 23:44:40.380319 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Apr 13 23:44:40.380330 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Apr 13 23:44:40.380340 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Apr 13 23:44:40.380351 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Apr 13 23:44:40.380361 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Apr 13 23:44:40.380372 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Apr 13 23:44:40.380383 kernel: Freeing SMP alternatives memory: 32K Apr 13 23:44:40.380394 kernel: pid_max: default: 32768 minimum: 301 Apr 13 23:44:40.380410 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 13 23:44:40.380420 kernel: landlock: Up and running. Apr 13 23:44:40.380431 kernel: SELinux: Initializing. Apr 13 23:44:40.380441 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 13 23:44:40.380452 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 13 23:44:40.380463 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8370C CPU @ 2.80GHz (family: 0x6, model: 0x6a, stepping: 0x6) Apr 13 23:44:40.380473 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 13 23:44:40.380484 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 13 23:44:40.380495 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 13 23:44:40.380508 kernel: Performance Events: unsupported p6 CPU model 106 no PMU driver, software events only. Apr 13 23:44:40.380518 kernel: signal: max sigframe size: 3632 Apr 13 23:44:40.380528 kernel: rcu: Hierarchical SRCU implementation. Apr 13 23:44:40.380539 kernel: rcu: Max phase no-delay instances is 400. Apr 13 23:44:40.380550 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Apr 13 23:44:40.380560 kernel: smp: Bringing up secondary CPUs ... Apr 13 23:44:40.380570 kernel: smpboot: x86: Booting SMP configuration: Apr 13 23:44:40.380580 kernel: .... node #0, CPUs: #1 #2 #3 Apr 13 23:44:40.380591 kernel: smp: Brought up 1 node, 4 CPUs Apr 13 23:44:40.380603 kernel: smpboot: Max logical packages: 1 Apr 13 23:44:40.380614 kernel: smpboot: Total of 4 processors activated (22347.50 BogoMIPS) Apr 13 23:44:40.380625 kernel: devtmpfs: initialized Apr 13 23:44:40.380635 kernel: x86/mm: Memory block size: 128MB Apr 13 23:44:40.380646 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Apr 13 23:44:40.380685 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Apr 13 23:44:40.380696 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) Apr 13 23:44:40.380706 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Apr 13 23:44:40.380717 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Apr 13 23:44:40.380730 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 13 23:44:40.380741 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Apr 13 23:44:40.380751 kernel: pinctrl core: initialized pinctrl subsystem Apr 13 23:44:40.380762 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 13 23:44:40.380773 kernel: audit: initializing netlink subsys (disabled) Apr 13 23:44:40.380783 kernel: audit: type=2000 audit(1776123878.091:1): state=initialized audit_enabled=0 res=1 Apr 13 23:44:40.380793 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 13 23:44:40.380804 kernel: thermal_sys: Registered thermal governor 'user_space' Apr 13 23:44:40.380814 kernel: cpuidle: using governor menu Apr 13 23:44:40.380828 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 13 23:44:40.380838 kernel: dca service started, version 1.12.1 Apr 13 23:44:40.380849 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Apr 13 23:44:40.380859 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Apr 13 23:44:40.380885 kernel: PCI: Using configuration type 1 for base access Apr 13 23:44:40.380896 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Apr 13 23:44:40.380907 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 13 23:44:40.380917 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Apr 13 23:44:40.380928 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 13 23:44:40.380941 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Apr 13 23:44:40.380952 kernel: ACPI: Added _OSI(Module Device) Apr 13 23:44:40.380962 kernel: ACPI: Added _OSI(Processor Device) Apr 13 23:44:40.380973 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 13 23:44:40.380983 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 13 23:44:40.380994 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Apr 13 23:44:40.381004 kernel: ACPI: Interpreter enabled Apr 13 23:44:40.381012 kernel: ACPI: PM: (supports S0 S3 S5) Apr 13 23:44:40.381021 kernel: ACPI: Using IOAPIC for interrupt routing Apr 13 23:44:40.381031 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Apr 13 23:44:40.381040 kernel: PCI: Using E820 reservations for host bridge windows Apr 13 23:44:40.381048 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Apr 13 23:44:40.381057 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 13 23:44:40.381279 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 13 23:44:40.381394 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Apr 13 23:44:40.381487 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Apr 13 23:44:40.381506 kernel: PCI host bridge to bus 0000:00 Apr 13 23:44:40.381594 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Apr 13 23:44:40.381790 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Apr 13 23:44:40.381968 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Apr 13 23:44:40.382070 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Apr 13 23:44:40.382149 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Apr 13 23:44:40.382228 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0xfffffffff window] Apr 13 23:44:40.382309 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 13 23:44:40.382408 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Apr 13 23:44:40.382503 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Apr 13 23:44:40.382587 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Apr 13 23:44:40.382742 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Apr 13 23:44:40.382837 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Apr 13 23:44:40.382946 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Apr 13 23:44:40.383020 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Apr 13 23:44:40.383173 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Apr 13 23:44:40.383240 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Apr 13 23:44:40.383297 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Apr 13 23:44:40.383364 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] Apr 13 23:44:40.383440 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Apr 13 23:44:40.383524 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Apr 13 23:44:40.383597 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Apr 13 23:44:40.383857 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] Apr 13 23:44:40.383967 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Apr 13 23:44:40.384039 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Apr 13 23:44:40.384110 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Apr 13 23:44:40.384166 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] Apr 13 23:44:40.384240 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Apr 13 23:44:40.384304 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Apr 13 23:44:40.384360 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Apr 13 23:44:40.384436 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Apr 13 23:44:40.384508 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Apr 13 23:44:40.384577 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Apr 13 23:44:40.384734 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Apr 13 23:44:40.384819 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Apr 13 23:44:40.384829 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Apr 13 23:44:40.384837 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Apr 13 23:44:40.384845 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Apr 13 23:44:40.384854 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Apr 13 23:44:40.384862 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Apr 13 23:44:40.384929 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Apr 13 23:44:40.384937 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Apr 13 23:44:40.384955 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Apr 13 23:44:40.384963 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Apr 13 23:44:40.384971 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Apr 13 23:44:40.384979 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Apr 13 23:44:40.384987 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Apr 13 23:44:40.384994 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Apr 13 23:44:40.385003 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Apr 13 23:44:40.385012 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Apr 13 23:44:40.385021 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Apr 13 23:44:40.385033 kernel: iommu: Default domain type: Translated Apr 13 23:44:40.385043 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Apr 13 23:44:40.385052 kernel: efivars: Registered efivars operations Apr 13 23:44:40.385060 kernel: PCI: Using ACPI for IRQ routing Apr 13 23:44:40.385067 kernel: PCI: pci_cache_line_size set to 64 bytes Apr 13 23:44:40.385076 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Apr 13 23:44:40.385085 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] Apr 13 23:44:40.385094 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] Apr 13 23:44:40.385103 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] Apr 13 23:44:40.385209 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Apr 13 23:44:40.385297 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Apr 13 23:44:40.385484 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Apr 13 23:44:40.385504 kernel: vgaarb: loaded Apr 13 23:44:40.385516 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Apr 13 23:44:40.385527 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Apr 13 23:44:40.385539 kernel: clocksource: Switched to clocksource kvm-clock Apr 13 23:44:40.385550 kernel: VFS: Disk quotas dquot_6.6.0 Apr 13 23:44:40.385560 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 13 23:44:40.385575 kernel: pnp: PnP ACPI init Apr 13 23:44:40.385821 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Apr 13 23:44:40.385842 kernel: pnp: PnP ACPI: found 6 devices Apr 13 23:44:40.385880 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Apr 13 23:44:40.385892 kernel: NET: Registered PF_INET protocol family Apr 13 23:44:40.385903 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 13 23:44:40.385913 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 13 23:44:40.385924 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 13 23:44:40.385939 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 13 23:44:40.385948 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 13 23:44:40.385957 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 13 23:44:40.385966 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 13 23:44:40.385975 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 13 23:44:40.385983 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 13 23:44:40.385992 kernel: NET: Registered PF_XDP protocol family Apr 13 23:44:40.386090 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Apr 13 23:44:40.386184 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Apr 13 23:44:40.386277 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Apr 13 23:44:40.386360 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Apr 13 23:44:40.386440 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Apr 13 23:44:40.386523 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Apr 13 23:44:40.386604 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Apr 13 23:44:40.386740 kernel: pci_bus 0000:00: resource 9 [mem 0x800000000-0xfffffffff window] Apr 13 23:44:40.386757 kernel: PCI: CLS 0 bytes, default 64 Apr 13 23:44:40.386768 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Apr 13 23:44:40.386782 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x284409db922, max_idle_ns: 440795228871 ns Apr 13 23:44:40.386793 kernel: Initialise system trusted keyrings Apr 13 23:44:40.386804 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 13 23:44:40.386815 kernel: Key type asymmetric registered Apr 13 23:44:40.386826 kernel: Asymmetric key parser 'x509' registered Apr 13 23:44:40.386836 kernel: hrtimer: interrupt took 14036304 ns Apr 13 23:44:40.386847 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Apr 13 23:44:40.386857 kernel: io scheduler mq-deadline registered Apr 13 23:44:40.386887 kernel: io scheduler kyber registered Apr 13 23:44:40.386899 kernel: io scheduler bfq registered Apr 13 23:44:40.386909 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Apr 13 23:44:40.386920 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Apr 13 23:44:40.386929 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Apr 13 23:44:40.386938 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Apr 13 23:44:40.386946 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 13 23:44:40.386955 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Apr 13 23:44:40.386963 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Apr 13 23:44:40.386975 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Apr 13 23:44:40.386984 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Apr 13 23:44:40.387110 kernel: rtc_cmos 00:04: RTC can wake from S4 Apr 13 23:44:40.387206 kernel: rtc_cmos 00:04: registered as rtc0 Apr 13 23:44:40.387287 kernel: rtc_cmos 00:04: setting system clock to 2026-04-13T23:44:39 UTC (1776123879) Apr 13 23:44:40.387307 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Apr 13 23:44:40.387408 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Apr 13 23:44:40.387422 kernel: intel_pstate: CPU model not supported Apr 13 23:44:40.387439 kernel: efifb: probing for efifb Apr 13 23:44:40.387450 kernel: efifb: framebuffer at 0xc0000000, using 1408k, total 1408k Apr 13 23:44:40.387461 kernel: efifb: mode is 800x600x24, linelength=2400, pages=1 Apr 13 23:44:40.387472 kernel: efifb: scrolling: redraw Apr 13 23:44:40.387485 kernel: efifb: Truecolor: size=0:8:8:8, shift=0:16:8:0 Apr 13 23:44:40.387496 kernel: Console: switching to colour frame buffer device 100x37 Apr 13 23:44:40.387526 kernel: fb0: EFI VGA frame buffer device Apr 13 23:44:40.387540 kernel: pstore: Using crash dump compression: deflate Apr 13 23:44:40.387550 kernel: pstore: Registered efi_pstore as persistent store backend Apr 13 23:44:40.387563 kernel: NET: Registered PF_INET6 protocol family Apr 13 23:44:40.387577 kernel: Segment Routing with IPv6 Apr 13 23:44:40.387588 kernel: In-situ OAM (IOAM) with IPv6 Apr 13 23:44:40.387599 kernel: NET: Registered PF_PACKET protocol family Apr 13 23:44:40.387610 kernel: Key type dns_resolver registered Apr 13 23:44:40.387620 kernel: IPI shorthand broadcast: enabled Apr 13 23:44:40.387631 kernel: sched_clock: Marking stable (1411137497, 383603560)->(2170872424, -376131367) Apr 13 23:44:40.387642 kernel: registered taskstats version 1 Apr 13 23:44:40.387803 kernel: Loading compiled-in X.509 certificates Apr 13 23:44:40.387815 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 51221ce98a81ccf90ef3d16403b42695603c5d00' Apr 13 23:44:40.387829 kernel: Key type .fscrypt registered Apr 13 23:44:40.387838 kernel: Key type fscrypt-provisioning registered Apr 13 23:44:40.387847 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 13 23:44:40.387856 kernel: ima: Allocated hash algorithm: sha1 Apr 13 23:44:40.387885 kernel: ima: No architecture policies found Apr 13 23:44:40.387895 kernel: clk: Disabling unused clocks Apr 13 23:44:40.387905 kernel: Freeing unused kernel image (initmem) memory: 42896K Apr 13 23:44:40.387915 kernel: Write protecting the kernel read-only data: 36864k Apr 13 23:44:40.387926 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Apr 13 23:44:40.387940 kernel: Run /init as init process Apr 13 23:44:40.387949 kernel: with arguments: Apr 13 23:44:40.387961 kernel: /init Apr 13 23:44:40.387970 kernel: with environment: Apr 13 23:44:40.387978 kernel: HOME=/ Apr 13 23:44:40.387987 kernel: TERM=linux Apr 13 23:44:40.388000 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 13 23:44:40.388013 systemd[1]: Detected virtualization kvm. Apr 13 23:44:40.388025 systemd[1]: Detected architecture x86-64. Apr 13 23:44:40.388034 systemd[1]: Running in initrd. Apr 13 23:44:40.388168 systemd[1]: No hostname configured, using default hostname. Apr 13 23:44:40.388179 systemd[1]: Hostname set to . Apr 13 23:44:40.388198 systemd[1]: Initializing machine ID from VM UUID. Apr 13 23:44:40.388209 systemd[1]: Queued start job for default target initrd.target. Apr 13 23:44:40.388219 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 13 23:44:40.388230 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 13 23:44:40.388241 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 13 23:44:40.388252 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 13 23:44:40.388263 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 13 23:44:40.388274 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 13 23:44:40.388290 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 13 23:44:40.388301 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 13 23:44:40.388312 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 13 23:44:40.388322 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 13 23:44:40.388333 systemd[1]: Reached target paths.target - Path Units. Apr 13 23:44:40.388344 systemd[1]: Reached target slices.target - Slice Units. Apr 13 23:44:40.388354 systemd[1]: Reached target swap.target - Swaps. Apr 13 23:44:40.388367 systemd[1]: Reached target timers.target - Timer Units. Apr 13 23:44:40.388378 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 13 23:44:40.388388 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 13 23:44:40.388398 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 13 23:44:40.388409 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 13 23:44:40.388420 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 13 23:44:40.388430 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 13 23:44:40.388441 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 13 23:44:40.388453 systemd[1]: Reached target sockets.target - Socket Units. Apr 13 23:44:40.388465 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 13 23:44:40.388476 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 13 23:44:40.388486 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 13 23:44:40.388496 systemd[1]: Starting systemd-fsck-usr.service... Apr 13 23:44:40.388507 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 13 23:44:40.388517 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 13 23:44:40.388527 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:40.388572 systemd-journald[194]: Collecting audit messages is disabled. Apr 13 23:44:40.388599 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 13 23:44:40.388609 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 13 23:44:40.388620 systemd-journald[194]: Journal started Apr 13 23:44:40.388646 systemd-journald[194]: Runtime Journal (/run/log/journal/13404a3bf12d4f779f7874cd10d1bee9) is 6.0M, max 48.3M, 42.2M free. Apr 13 23:44:40.394710 systemd[1]: Started systemd-journald.service - Journal Service. Apr 13 23:44:40.397259 systemd[1]: Finished systemd-fsck-usr.service. Apr 13 23:44:40.411094 systemd-modules-load[195]: Inserted module 'overlay' Apr 13 23:44:40.412003 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 13 23:44:40.421808 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 13 23:44:40.424225 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:40.432240 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 13 23:44:40.443297 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 13 23:44:40.450568 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 13 23:44:40.466819 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 13 23:44:40.469797 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 13 23:44:40.470486 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 13 23:44:40.489767 kernel: Bridge firewalling registered Apr 13 23:44:40.492470 systemd-modules-load[195]: Inserted module 'br_netfilter' Apr 13 23:44:40.493793 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 13 23:44:40.512193 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 13 23:44:40.545961 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 13 23:44:40.557298 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 13 23:44:40.564397 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 13 23:44:40.577977 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 13 23:44:40.590428 dracut-cmdline[229]: dracut-dracut-053 Apr 13 23:44:40.598400 dracut-cmdline[229]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c1ba97db2f6278922cfc5bd0ca74b4bb573fca2c3aed19c121a34271e693e156 Apr 13 23:44:40.628628 systemd-resolved[233]: Positive Trust Anchors: Apr 13 23:44:40.628775 systemd-resolved[233]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 13 23:44:40.628811 systemd-resolved[233]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 13 23:44:40.632520 systemd-resolved[233]: Defaulting to hostname 'linux'. Apr 13 23:44:40.633963 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 13 23:44:40.653143 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 13 23:44:40.806759 kernel: SCSI subsystem initialized Apr 13 23:44:40.888943 kernel: Loading iSCSI transport class v2.0-870. Apr 13 23:44:40.920740 kernel: iscsi: registered transport (tcp) Apr 13 23:44:40.957337 kernel: iscsi: registered transport (qla4xxx) Apr 13 23:44:40.957414 kernel: QLogic iSCSI HBA Driver Apr 13 23:44:41.093547 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 13 23:44:41.113297 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 13 23:44:41.166571 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 13 23:44:41.166650 kernel: device-mapper: uevent: version 1.0.3 Apr 13 23:44:41.169370 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 13 23:44:41.300938 kernel: raid6: avx512x4 gen() 36067 MB/s Apr 13 23:44:41.318887 kernel: raid6: avx512x2 gen() 32815 MB/s Apr 13 23:44:41.336825 kernel: raid6: avx512x1 gen() 23329 MB/s Apr 13 23:44:41.353750 kernel: raid6: avx2x4 gen() 28215 MB/s Apr 13 23:44:41.373914 kernel: raid6: avx2x2 gen() 30903 MB/s Apr 13 23:44:41.391950 kernel: raid6: avx2x1 gen() 16964 MB/s Apr 13 23:44:41.395527 kernel: raid6: using algorithm avx512x4 gen() 36067 MB/s Apr 13 23:44:41.409955 kernel: raid6: .... xor() 4893 MB/s, rmw enabled Apr 13 23:44:41.410073 kernel: raid6: using avx512x2 recovery algorithm Apr 13 23:44:41.462731 kernel: xor: automatically using best checksumming function avx Apr 13 23:44:41.727722 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 13 23:44:41.741562 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 13 23:44:41.752181 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 13 23:44:41.768813 systemd-udevd[415]: Using default interface naming scheme 'v255'. Apr 13 23:44:41.776567 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 13 23:44:41.817531 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 13 23:44:41.834027 dracut-pre-trigger[416]: rd.md=0: removing MD RAID activation Apr 13 23:44:41.909794 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 13 23:44:41.955358 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 13 23:44:42.016328 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 13 23:44:42.029202 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 13 23:44:42.041341 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 13 23:44:42.049145 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 13 23:44:42.053801 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Apr 13 23:44:42.072452 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Apr 13 23:44:42.057207 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 13 23:44:42.069352 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 13 23:44:42.083140 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 13 23:44:42.083180 kernel: GPT:9289727 != 19775487 Apr 13 23:44:42.083189 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 13 23:44:42.083201 kernel: GPT:9289727 != 19775487 Apr 13 23:44:42.083208 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 13 23:44:42.083215 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 13 23:44:42.087940 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 13 23:44:42.093804 kernel: cryptd: max_cpu_qlen set to 1000 Apr 13 23:44:42.139774 kernel: libata version 3.00 loaded. Apr 13 23:44:42.141138 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 13 23:44:42.165700 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (469) Apr 13 23:44:42.165753 kernel: BTRFS: device fsid de1edd48-4571-4695-92f0-7af6e33c4e3d devid 1 transid 31 /dev/vda3 scanned by (udev-worker) (471) Apr 13 23:44:42.160472 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 13 23:44:42.160784 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 13 23:44:42.173446 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 13 23:44:42.179361 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 13 23:44:42.182273 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:42.189589 kernel: ahci 0000:00:1f.2: version 3.0 Apr 13 23:44:42.190015 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Apr 13 23:44:42.190035 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Apr 13 23:44:42.190286 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:42.200956 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Apr 13 23:44:42.201131 kernel: scsi host0: ahci Apr 13 23:44:42.201221 kernel: AVX2 version of gcm_enc/dec engaged. Apr 13 23:44:42.201229 kernel: scsi host1: ahci Apr 13 23:44:42.201298 kernel: scsi host2: ahci Apr 13 23:44:42.201373 kernel: AES CTR mode by8 optimization enabled Apr 13 23:44:42.201385 kernel: scsi host3: ahci Apr 13 23:44:42.203767 kernel: scsi host4: ahci Apr 13 23:44:42.211344 kernel: scsi host5: ahci Apr 13 23:44:42.215970 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:42.289814 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Apr 13 23:44:42.289848 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Apr 13 23:44:42.289860 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Apr 13 23:44:42.289891 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Apr 13 23:44:42.289907 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Apr 13 23:44:42.289918 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Apr 13 23:44:42.296089 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Apr 13 23:44:42.310095 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Apr 13 23:44:42.349422 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Apr 13 23:44:42.354227 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Apr 13 23:44:42.370600 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Apr 13 23:44:42.394912 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 13 23:44:42.398093 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 13 23:44:42.398155 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:42.414258 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 13 23:44:42.414285 disk-uuid[554]: Primary Header is updated. Apr 13 23:44:42.414285 disk-uuid[554]: Secondary Entries is updated. Apr 13 23:44:42.414285 disk-uuid[554]: Secondary Header is updated. Apr 13 23:44:42.404375 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:42.410156 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:42.508079 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 13 23:44:42.517490 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:42.529999 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 13 23:44:42.577524 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 13 23:44:42.589672 kernel: ata2: SATA link down (SStatus 0 SControl 300) Apr 13 23:44:42.598692 kernel: ata4: SATA link down (SStatus 0 SControl 300) Apr 13 23:44:42.598759 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Apr 13 23:44:42.601896 kernel: ata5: SATA link down (SStatus 0 SControl 300) Apr 13 23:44:42.601950 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Apr 13 23:44:42.605165 kernel: ata3.00: applying bridge limits Apr 13 23:44:42.606166 kernel: ata1: SATA link down (SStatus 0 SControl 300) Apr 13 23:44:42.607686 kernel: ata3.00: configured for UDMA/100 Apr 13 23:44:42.609738 kernel: ata6: SATA link down (SStatus 0 SControl 300) Apr 13 23:44:42.614907 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 13 23:44:42.684631 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Apr 13 23:44:42.685281 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 13 23:44:42.700752 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Apr 13 23:44:43.551795 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 13 23:44:43.551889 disk-uuid[556]: The operation has completed successfully. Apr 13 23:44:43.579856 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 13 23:44:43.580117 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 13 23:44:43.618003 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 13 23:44:43.653030 sh[598]: Success Apr 13 23:44:43.760788 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Apr 13 23:44:43.813551 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 13 23:44:43.858228 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 13 23:44:43.859901 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 13 23:44:43.917723 kernel: BTRFS info (device dm-0): first mount of filesystem de1edd48-4571-4695-92f0-7af6e33c4e3d Apr 13 23:44:43.917917 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Apr 13 23:44:43.917967 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 13 23:44:43.958161 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 13 23:44:43.979337 kernel: BTRFS info (device dm-0): using free space tree Apr 13 23:44:44.005999 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 13 23:44:44.009972 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 13 23:44:44.026409 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 13 23:44:44.031821 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 13 23:44:44.044355 kernel: BTRFS info (device vda6): first mount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 13 23:44:44.044415 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 13 23:44:44.044425 kernel: BTRFS info (device vda6): using free space tree Apr 13 23:44:44.049814 kernel: BTRFS info (device vda6): auto enabling async discard Apr 13 23:44:44.060637 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 13 23:44:44.065118 kernel: BTRFS info (device vda6): last unmount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 13 23:44:44.074174 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 13 23:44:44.088014 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 13 23:44:44.224316 ignition[694]: Ignition 2.19.0 Apr 13 23:44:44.224339 ignition[694]: Stage: fetch-offline Apr 13 23:44:44.224379 ignition[694]: no configs at "/usr/lib/ignition/base.d" Apr 13 23:44:44.224388 ignition[694]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 13 23:44:44.224486 ignition[694]: parsed url from cmdline: "" Apr 13 23:44:44.224490 ignition[694]: no config URL provided Apr 13 23:44:44.224495 ignition[694]: reading system config file "/usr/lib/ignition/user.ign" Apr 13 23:44:44.224503 ignition[694]: no config at "/usr/lib/ignition/user.ign" Apr 13 23:44:44.224533 ignition[694]: op(1): [started] loading QEMU firmware config module Apr 13 23:44:44.224539 ignition[694]: op(1): executing: "modprobe" "qemu_fw_cfg" Apr 13 23:44:44.243511 ignition[694]: op(1): [finished] loading QEMU firmware config module Apr 13 23:44:44.246627 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 13 23:44:44.254893 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 13 23:44:44.274162 systemd-networkd[786]: lo: Link UP Apr 13 23:44:44.274185 systemd-networkd[786]: lo: Gained carrier Apr 13 23:44:44.275610 systemd-networkd[786]: Enumeration completed Apr 13 23:44:44.275785 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 13 23:44:44.276603 systemd-networkd[786]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 13 23:44:44.276607 systemd-networkd[786]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 13 23:44:44.279113 systemd-networkd[786]: eth0: Link UP Apr 13 23:44:44.279117 systemd-networkd[786]: eth0: Gained carrier Apr 13 23:44:44.279129 systemd-networkd[786]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 13 23:44:44.279251 systemd[1]: Reached target network.target - Network. Apr 13 23:44:44.307850 systemd-networkd[786]: eth0: DHCPv4 address 10.0.0.25/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 13 23:44:44.473436 ignition[694]: parsing config with SHA512: 6195d8321114e9996e9efe82cb59bd417e24cdbcd23a68933135e7ca924efaf31b829f7fcdc87995285b06ac9b7f9b2f0464e6f19d38d9a13033eb348b84ca09 Apr 13 23:44:44.482855 unknown[694]: fetched base config from "system" Apr 13 23:44:44.482886 unknown[694]: fetched user config from "qemu" Apr 13 23:44:44.483533 ignition[694]: fetch-offline: fetch-offline passed Apr 13 23:44:44.486514 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 13 23:44:44.483613 ignition[694]: Ignition finished successfully Apr 13 23:44:44.499703 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Apr 13 23:44:44.515932 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 13 23:44:44.599644 ignition[791]: Ignition 2.19.0 Apr 13 23:44:44.599721 ignition[791]: Stage: kargs Apr 13 23:44:44.599948 ignition[791]: no configs at "/usr/lib/ignition/base.d" Apr 13 23:44:44.603820 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 13 23:44:44.599958 ignition[791]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 13 23:44:44.618373 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 13 23:44:44.601123 ignition[791]: kargs: kargs passed Apr 13 23:44:44.601185 ignition[791]: Ignition finished successfully Apr 13 23:44:44.648712 ignition[799]: Ignition 2.19.0 Apr 13 23:44:44.648732 ignition[799]: Stage: disks Apr 13 23:44:44.648959 ignition[799]: no configs at "/usr/lib/ignition/base.d" Apr 13 23:44:44.648968 ignition[799]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 13 23:44:44.650439 ignition[799]: disks: disks passed Apr 13 23:44:44.654116 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 13 23:44:44.650520 ignition[799]: Ignition finished successfully Apr 13 23:44:44.662306 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 13 23:44:44.665242 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 13 23:44:44.679689 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 13 23:44:44.681968 systemd[1]: Reached target sysinit.target - System Initialization. Apr 13 23:44:44.685971 systemd[1]: Reached target basic.target - Basic System. Apr 13 23:44:44.704177 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 13 23:44:44.781922 systemd-fsck[809]: ROOT: clean, 14/553520 files, 52654/553472 blocks Apr 13 23:44:44.798272 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 13 23:44:44.821036 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 13 23:44:45.010941 kernel: EXT4-fs (vda9): mounted filesystem e02793bf-3e0d-4c7e-b11a-92c664da7ce3 r/w with ordered data mode. Quota mode: none. Apr 13 23:44:45.010985 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 13 23:44:45.014184 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 13 23:44:45.030218 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 13 23:44:45.038949 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 13 23:44:45.043094 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Apr 13 23:44:45.043159 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 13 23:44:45.043179 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 13 23:44:45.062379 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (817) Apr 13 23:44:45.068783 kernel: BTRFS info (device vda6): first mount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 13 23:44:45.068843 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 13 23:44:45.068858 kernel: BTRFS info (device vda6): using free space tree Apr 13 23:44:45.076683 kernel: BTRFS info (device vda6): auto enabling async discard Apr 13 23:44:45.081183 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 13 23:44:45.086311 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 13 23:44:45.141222 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 13 23:44:45.200564 initrd-setup-root[841]: cut: /sysroot/etc/passwd: No such file or directory Apr 13 23:44:45.207601 initrd-setup-root[848]: cut: /sysroot/etc/group: No such file or directory Apr 13 23:44:45.229153 initrd-setup-root[855]: cut: /sysroot/etc/shadow: No such file or directory Apr 13 23:44:45.235035 initrd-setup-root[862]: cut: /sysroot/etc/gshadow: No such file or directory Apr 13 23:44:45.419452 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 13 23:44:45.437997 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 13 23:44:45.441431 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 13 23:44:45.448826 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 13 23:44:45.451715 kernel: BTRFS info (device vda6): last unmount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 13 23:44:45.490774 ignition[931]: INFO : Ignition 2.19.0 Apr 13 23:44:45.493329 ignition[931]: INFO : Stage: mount Apr 13 23:44:45.493329 ignition[931]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 13 23:44:45.493329 ignition[931]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 13 23:44:45.503408 ignition[931]: INFO : mount: mount passed Apr 13 23:44:45.503408 ignition[931]: INFO : Ignition finished successfully Apr 13 23:44:45.500632 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 13 23:44:45.503787 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 13 23:44:45.517050 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 13 23:44:45.527912 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 13 23:44:45.572801 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (944) Apr 13 23:44:45.599381 kernel: BTRFS info (device vda6): first mount of filesystem 7dd1319a-da93-42af-ac3b-f04d4587a8af Apr 13 23:44:45.599459 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 13 23:44:45.599469 kernel: BTRFS info (device vda6): using free space tree Apr 13 23:44:45.616830 kernel: BTRFS info (device vda6): auto enabling async discard Apr 13 23:44:45.634288 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 13 23:44:45.679390 ignition[961]: INFO : Ignition 2.19.0 Apr 13 23:44:45.679390 ignition[961]: INFO : Stage: files Apr 13 23:44:45.683229 ignition[961]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 13 23:44:45.683229 ignition[961]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 13 23:44:45.683229 ignition[961]: DEBUG : files: compiled without relabeling support, skipping Apr 13 23:44:45.693847 ignition[961]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 13 23:44:45.693847 ignition[961]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 13 23:44:45.702846 ignition[961]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 13 23:44:45.706774 ignition[961]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 13 23:44:45.711005 ignition[961]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 13 23:44:45.710246 unknown[961]: wrote ssh authorized keys file for user: core Apr 13 23:44:45.730621 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 13 23:44:45.730621 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Apr 13 23:44:45.830519 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 13 23:44:45.853929 systemd-networkd[786]: eth0: Gained IPv6LL Apr 13 23:44:46.012453 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 13 23:44:46.012453 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 13 23:44:46.029721 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-x86-64.raw: attempt #1 Apr 13 23:44:46.205629 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 13 23:44:47.686249 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 13 23:44:47.686249 ignition[961]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Apr 13 23:44:47.694801 ignition[961]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Apr 13 23:44:47.806541 ignition[961]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Apr 13 23:44:47.840294 ignition[961]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Apr 13 23:44:47.847104 ignition[961]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Apr 13 23:44:47.847104 ignition[961]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Apr 13 23:44:47.847104 ignition[961]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Apr 13 23:44:47.847104 ignition[961]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 13 23:44:47.847104 ignition[961]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 13 23:44:47.847104 ignition[961]: INFO : files: files passed Apr 13 23:44:47.847104 ignition[961]: INFO : Ignition finished successfully Apr 13 23:44:47.844029 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 13 23:44:47.864405 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 13 23:44:47.868938 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 13 23:44:47.884135 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 13 23:44:47.884276 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 13 23:44:47.893918 initrd-setup-root-after-ignition[988]: grep: /sysroot/oem/oem-release: No such file or directory Apr 13 23:44:47.896774 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 13 23:44:47.896774 initrd-setup-root-after-ignition[990]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 13 23:44:47.903256 initrd-setup-root-after-ignition[995]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 13 23:44:47.900775 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 13 23:44:47.907028 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 13 23:44:47.984455 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 13 23:44:48.057973 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 13 23:44:48.058100 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 13 23:44:48.061923 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 13 23:44:48.062452 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 13 23:44:48.077936 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 13 23:44:48.094914 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 13 23:44:48.117929 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 13 23:44:48.142395 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 13 23:44:48.165312 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 13 23:44:48.168186 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 13 23:44:48.171953 systemd[1]: Stopped target timers.target - Timer Units. Apr 13 23:44:48.175857 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 13 23:44:48.176380 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 13 23:44:48.180445 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 13 23:44:48.184644 systemd[1]: Stopped target basic.target - Basic System. Apr 13 23:44:48.197315 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 13 23:44:48.204534 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 13 23:44:48.208069 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 13 23:44:48.212996 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 13 23:44:48.217459 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 13 23:44:48.230075 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 13 23:44:48.234952 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 13 23:44:48.240540 systemd[1]: Stopped target swap.target - Swaps. Apr 13 23:44:48.244382 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 13 23:44:48.244772 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 13 23:44:48.249839 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 13 23:44:48.254241 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 13 23:44:48.266190 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 13 23:44:48.268804 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 13 23:44:48.275109 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 13 23:44:48.275299 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 13 23:44:48.281272 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 13 23:44:48.281889 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 13 23:44:48.291882 systemd[1]: Stopped target paths.target - Path Units. Apr 13 23:44:48.292372 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 13 23:44:48.295039 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 13 23:44:48.299713 systemd[1]: Stopped target slices.target - Slice Units. Apr 13 23:44:48.306281 systemd[1]: Stopped target sockets.target - Socket Units. Apr 13 23:44:48.309119 systemd[1]: iscsid.socket: Deactivated successfully. Apr 13 23:44:48.309426 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 13 23:44:48.315972 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 13 23:44:48.316540 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 13 23:44:48.369540 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 13 23:44:48.369784 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 13 23:44:48.374613 systemd[1]: ignition-files.service: Deactivated successfully. Apr 13 23:44:48.375180 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 13 23:44:48.405709 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 13 23:44:48.409201 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 13 23:44:48.409622 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 13 23:44:48.417573 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 13 23:44:48.427415 ignition[1015]: INFO : Ignition 2.19.0 Apr 13 23:44:48.427415 ignition[1015]: INFO : Stage: umount Apr 13 23:44:48.427415 ignition[1015]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 13 23:44:48.427415 ignition[1015]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 13 23:44:48.427415 ignition[1015]: INFO : umount: umount passed Apr 13 23:44:48.427415 ignition[1015]: INFO : Ignition finished successfully Apr 13 23:44:48.419697 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 13 23:44:48.419989 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 13 23:44:48.421995 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 13 23:44:48.422084 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 13 23:44:48.430534 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 13 23:44:48.430779 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 13 23:44:48.439992 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 13 23:44:48.440126 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 13 23:44:48.445812 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 13 23:44:48.449181 systemd[1]: Stopped target network.target - Network. Apr 13 23:44:48.454526 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 13 23:44:48.454604 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 13 23:44:48.456083 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 13 23:44:48.456134 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 13 23:44:48.466744 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 13 23:44:48.466830 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 13 23:44:48.467028 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 13 23:44:48.467070 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 13 23:44:48.479143 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 13 23:44:48.483770 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 13 23:44:48.497844 systemd-networkd[786]: eth0: DHCPv6 lease lost Apr 13 23:44:48.500757 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 13 23:44:48.502264 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 13 23:44:48.507955 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 13 23:44:48.508096 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 13 23:44:48.518383 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 13 23:44:48.518441 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 13 23:44:48.581268 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 13 23:44:48.582416 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 13 23:44:48.582491 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 13 23:44:48.587417 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 13 23:44:48.587478 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 13 23:44:48.595505 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 13 23:44:48.595975 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 13 23:44:48.603464 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 13 23:44:48.603534 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 13 23:44:48.608164 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 13 23:44:48.611847 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 13 23:44:48.612167 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 13 23:44:48.631310 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 13 23:44:48.631472 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 13 23:44:48.638325 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 13 23:44:48.638461 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 13 23:44:48.645551 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 13 23:44:48.645798 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 13 23:44:48.652129 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 13 23:44:48.652238 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 13 23:44:48.655762 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 13 23:44:48.655809 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 13 23:44:48.664598 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 13 23:44:48.664688 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 13 23:44:48.674106 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 13 23:44:48.674212 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 13 23:44:48.678037 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 13 23:44:48.678179 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 13 23:44:48.701080 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 13 23:44:48.701351 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 13 23:44:48.701426 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 13 23:44:48.715410 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 13 23:44:48.715488 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:48.769610 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 13 23:44:48.770342 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 13 23:44:48.773325 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 13 23:44:48.795280 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 13 23:44:48.812845 systemd[1]: Switching root. Apr 13 23:44:48.858568 systemd-journald[194]: Journal stopped Apr 13 23:44:50.767525 systemd-journald[194]: Received SIGTERM from PID 1 (systemd). Apr 13 23:44:50.767610 kernel: SELinux: policy capability network_peer_controls=1 Apr 13 23:44:50.767631 kernel: SELinux: policy capability open_perms=1 Apr 13 23:44:50.767648 kernel: SELinux: policy capability extended_socket_class=1 Apr 13 23:44:50.767695 kernel: SELinux: policy capability always_check_network=0 Apr 13 23:44:50.767711 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 13 23:44:50.767730 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 13 23:44:50.767741 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 13 23:44:50.767752 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 13 23:44:50.767763 kernel: audit: type=1403 audit(1776123889.070:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 13 23:44:50.767776 systemd[1]: Successfully loaded SELinux policy in 72.032ms. Apr 13 23:44:50.767792 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.310ms. Apr 13 23:44:50.767805 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 13 23:44:50.767819 systemd[1]: Detected virtualization kvm. Apr 13 23:44:50.767833 systemd[1]: Detected architecture x86-64. Apr 13 23:44:50.767848 systemd[1]: Detected first boot. Apr 13 23:44:50.767878 systemd[1]: Initializing machine ID from VM UUID. Apr 13 23:44:50.767890 zram_generator::config[1060]: No configuration found. Apr 13 23:44:50.767904 systemd[1]: Populated /etc with preset unit settings. Apr 13 23:44:50.767917 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 13 23:44:50.767931 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 13 23:44:50.767944 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 13 23:44:50.767959 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 13 23:44:50.767978 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 13 23:44:50.767992 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 13 23:44:50.768011 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 13 23:44:50.768024 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 13 23:44:50.768036 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 13 23:44:50.768048 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 13 23:44:50.768061 systemd[1]: Created slice user.slice - User and Session Slice. Apr 13 23:44:50.768074 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 13 23:44:50.768090 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 13 23:44:50.768105 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 13 23:44:50.768118 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 13 23:44:50.768133 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 13 23:44:50.768146 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 13 23:44:50.768158 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Apr 13 23:44:50.768169 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 13 23:44:50.768181 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 13 23:44:50.768193 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 13 23:44:50.768208 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 13 23:44:50.768220 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 13 23:44:50.768232 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 13 23:44:50.768246 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 13 23:44:50.768261 systemd[1]: Reached target slices.target - Slice Units. Apr 13 23:44:50.768274 systemd[1]: Reached target swap.target - Swaps. Apr 13 23:44:50.768289 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 13 23:44:50.768302 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 13 23:44:50.768317 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 13 23:44:50.768329 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 13 23:44:50.768342 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 13 23:44:50.768355 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 13 23:44:50.768367 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 13 23:44:50.768379 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 13 23:44:50.768391 systemd[1]: Mounting media.mount - External Media Directory... Apr 13 23:44:50.768404 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:50.768416 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 13 23:44:50.768434 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 13 23:44:50.768447 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 13 23:44:50.768462 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 13 23:44:50.768474 systemd[1]: Reached target machines.target - Containers. Apr 13 23:44:50.768487 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 13 23:44:50.768499 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 13 23:44:50.768511 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 13 23:44:50.768523 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 13 23:44:50.768538 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 13 23:44:50.768550 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 13 23:44:50.768563 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 13 23:44:50.768577 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 13 23:44:50.768591 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 13 23:44:50.768604 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 13 23:44:50.768616 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 13 23:44:50.768628 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 13 23:44:50.768640 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 13 23:44:50.768704 kernel: fuse: init (API version 7.39) Apr 13 23:44:50.768725 systemd[1]: Stopped systemd-fsck-usr.service. Apr 13 23:44:50.768738 kernel: loop: module loaded Apr 13 23:44:50.768750 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 13 23:44:50.768763 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 13 23:44:50.768776 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 13 23:44:50.768788 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 13 23:44:50.768827 systemd-journald[1144]: Collecting audit messages is disabled. Apr 13 23:44:50.769011 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 13 23:44:50.769040 systemd-journald[1144]: Journal started Apr 13 23:44:50.769069 systemd-journald[1144]: Runtime Journal (/run/log/journal/13404a3bf12d4f779f7874cd10d1bee9) is 6.0M, max 48.3M, 42.2M free. Apr 13 23:44:50.088395 systemd[1]: Queued start job for default target multi-user.target. Apr 13 23:44:50.175388 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Apr 13 23:44:50.177896 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 13 23:44:50.771730 systemd[1]: verity-setup.service: Deactivated successfully. Apr 13 23:44:50.773722 systemd[1]: Stopped verity-setup.service. Apr 13 23:44:50.782694 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:50.782768 kernel: ACPI: bus type drm_connector registered Apr 13 23:44:50.784777 systemd[1]: Started systemd-journald.service - Journal Service. Apr 13 23:44:50.787338 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 13 23:44:50.794896 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 13 23:44:50.797519 systemd[1]: Mounted media.mount - External Media Directory. Apr 13 23:44:50.799826 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 13 23:44:50.802359 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 13 23:44:50.804800 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 13 23:44:50.808152 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 13 23:44:50.813012 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 13 23:44:50.842031 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 13 23:44:50.842496 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 13 23:44:50.845126 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 13 23:44:50.845526 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 13 23:44:50.848288 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 13 23:44:50.849392 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 13 23:44:50.851607 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 13 23:44:50.852639 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 13 23:44:50.855578 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 13 23:44:50.856358 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 13 23:44:50.859588 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 13 23:44:50.859788 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 13 23:44:50.862560 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 13 23:44:50.865426 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 13 23:44:50.868587 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 13 23:44:50.885695 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 13 23:44:50.902992 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 13 23:44:50.913119 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 13 23:44:50.946776 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 13 23:44:50.946837 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 13 23:44:50.950001 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 13 23:44:50.961015 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 13 23:44:50.967220 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 13 23:44:50.970602 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 13 23:44:50.975137 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 13 23:44:50.980718 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 13 23:44:50.984126 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 13 23:44:50.986961 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 13 23:44:50.991331 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 13 23:44:51.003176 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 13 23:44:51.016724 systemd-journald[1144]: Time spent on flushing to /var/log/journal/13404a3bf12d4f779f7874cd10d1bee9 is 167.687ms for 994 entries. Apr 13 23:44:51.016724 systemd-journald[1144]: System Journal (/var/log/journal/13404a3bf12d4f779f7874cd10d1bee9) is 8.0M, max 195.6M, 187.6M free. Apr 13 23:44:51.265030 systemd-journald[1144]: Received client request to flush runtime journal. Apr 13 23:44:51.265229 kernel: loop0: detected capacity change from 0 to 142488 Apr 13 23:44:51.058604 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 13 23:44:51.072368 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 13 23:44:51.077163 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 13 23:44:51.079386 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 13 23:44:51.081496 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 13 23:44:51.083939 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 13 23:44:51.089005 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 13 23:44:51.102974 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 13 23:44:51.273288 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 13 23:44:51.288490 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 13 23:44:51.318585 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 13 23:44:51.699707 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 13 23:44:51.698626 udevadm[1186]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 13 23:44:51.704481 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 13 23:44:51.785468 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 13 23:44:51.790555 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 13 23:44:51.790737 kernel: loop1: detected capacity change from 0 to 228704 Apr 13 23:44:51.799483 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 13 23:44:51.819199 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 13 23:44:51.875723 kernel: loop2: detected capacity change from 0 to 140768 Apr 13 23:44:51.905892 systemd-tmpfiles[1193]: ACLs are not supported, ignoring. Apr 13 23:44:51.905912 systemd-tmpfiles[1193]: ACLs are not supported, ignoring. Apr 13 23:44:51.943156 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 13 23:44:52.013720 kernel: loop3: detected capacity change from 0 to 142488 Apr 13 23:44:52.050888 kernel: loop4: detected capacity change from 0 to 228704 Apr 13 23:44:52.085710 kernel: loop5: detected capacity change from 0 to 140768 Apr 13 23:44:52.122970 (sd-merge)[1198]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Apr 13 23:44:52.123539 (sd-merge)[1198]: Merged extensions into '/usr'. Apr 13 23:44:52.139650 systemd[1]: Reloading requested from client PID 1174 ('systemd-sysext') (unit systemd-sysext.service)... Apr 13 23:44:52.140360 systemd[1]: Reloading... Apr 13 23:44:52.617832 zram_generator::config[1224]: No configuration found. Apr 13 23:44:52.970551 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 13 23:44:53.304816 ldconfig[1169]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 13 23:44:53.320801 systemd[1]: Reloading finished in 1179 ms. Apr 13 23:44:53.375752 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 13 23:44:53.381232 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 13 23:44:53.404484 systemd[1]: Starting ensure-sysext.service... Apr 13 23:44:53.420585 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 13 23:44:53.481412 systemd[1]: Reloading requested from client PID 1261 ('systemctl') (unit ensure-sysext.service)... Apr 13 23:44:53.481436 systemd[1]: Reloading... Apr 13 23:44:53.536294 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 13 23:44:53.536775 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 13 23:44:53.539077 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 13 23:44:53.539455 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Apr 13 23:44:53.539506 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Apr 13 23:44:53.545825 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Apr 13 23:44:53.547028 systemd-tmpfiles[1262]: Skipping /boot Apr 13 23:44:53.566861 zram_generator::config[1287]: No configuration found. Apr 13 23:44:53.569540 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Apr 13 23:44:53.569812 systemd-tmpfiles[1262]: Skipping /boot Apr 13 23:44:53.938717 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 13 23:44:54.090879 systemd[1]: Reloading finished in 609 ms. Apr 13 23:44:54.175592 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 13 23:44:54.223622 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 13 23:44:54.232954 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 13 23:44:54.238264 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 13 23:44:54.264705 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 13 23:44:54.281920 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 13 23:44:54.296272 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:54.296549 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 13 23:44:54.314199 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 13 23:44:54.375106 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 13 23:44:54.383479 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 13 23:44:54.385926 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 13 23:44:54.399440 augenrules[1350]: No rules Apr 13 23:44:54.401990 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 13 23:44:54.404680 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:54.407827 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 13 23:44:54.413311 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 13 23:44:54.424471 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 13 23:44:54.424623 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 13 23:44:54.428689 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 13 23:44:54.428807 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 13 23:44:54.433384 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 13 23:44:54.433504 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 13 23:44:54.451142 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:54.451281 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 13 23:44:54.462691 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 13 23:44:54.491272 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 13 23:44:54.496955 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 13 23:44:54.499982 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 13 23:44:54.500201 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:54.501495 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 13 23:44:54.507938 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 13 23:44:54.514506 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 13 23:44:54.550205 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 13 23:44:54.550781 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 13 23:44:54.555515 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 13 23:44:54.556245 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 13 23:44:54.565018 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 13 23:44:54.565478 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 13 23:44:54.579749 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:54.580331 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 13 23:44:54.593516 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 13 23:44:54.599090 systemd-resolved[1336]: Positive Trust Anchors: Apr 13 23:44:54.599103 systemd-resolved[1336]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 13 23:44:54.599132 systemd-resolved[1336]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 13 23:44:54.601049 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 13 23:44:54.607791 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 13 23:44:54.614678 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 13 23:44:54.617451 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 13 23:44:54.618074 systemd-resolved[1336]: Defaulting to hostname 'linux'. Apr 13 23:44:54.618194 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 13 23:44:54.618291 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 13 23:44:54.634016 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 13 23:44:54.638003 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 13 23:44:54.638523 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 13 23:44:54.659630 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 13 23:44:54.659800 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 13 23:44:54.670970 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 13 23:44:54.671459 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 13 23:44:54.675447 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 13 23:44:54.678115 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 13 23:44:54.692634 systemd[1]: Finished ensure-sysext.service. Apr 13 23:44:54.714216 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 13 23:44:54.756104 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 13 23:44:54.756228 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 13 23:44:54.774257 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 13 23:44:54.963602 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 13 23:44:54.967867 systemd[1]: Reached target time-set.target - System Time Set. Apr 13 23:44:55.291250 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 13 23:44:55.308968 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 13 23:44:55.322166 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 13 23:44:55.378497 systemd-udevd[1385]: Using default interface naming scheme 'v255'. Apr 13 23:44:55.420495 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 13 23:44:55.494803 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 13 23:44:55.515999 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 13 23:44:55.638748 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1392) Apr 13 23:44:55.662447 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Apr 13 23:44:55.671200 systemd-networkd[1394]: lo: Link UP Apr 13 23:44:55.671208 systemd-networkd[1394]: lo: Gained carrier Apr 13 23:44:55.672412 systemd-networkd[1394]: Enumeration completed Apr 13 23:44:55.677958 systemd-networkd[1394]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 13 23:44:55.677966 systemd-networkd[1394]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 13 23:44:55.679434 systemd-networkd[1394]: eth0: Link UP Apr 13 23:44:55.679437 systemd-networkd[1394]: eth0: Gained carrier Apr 13 23:44:55.679454 systemd-networkd[1394]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 13 23:44:55.830359 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 13 23:44:55.859081 systemd[1]: Reached target network.target - Network. Apr 13 23:44:55.961382 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 13 23:44:55.975792 systemd-networkd[1394]: eth0: DHCPv4 address 10.0.0.25/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 13 23:44:55.981405 systemd-timesyncd[1382]: Network configuration changed, trying to establish connection. Apr 13 23:44:56.507594 systemd-timesyncd[1382]: Contacted time server 10.0.0.1:123 (10.0.0.1). Apr 13 23:44:56.507638 systemd-timesyncd[1382]: Initial clock synchronization to Mon 2026-04-13 23:44:56.507476 UTC. Apr 13 23:44:56.530571 systemd-resolved[1336]: Clock change detected. Flushing caches. Apr 13 23:44:56.606637 systemd-networkd[1394]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 13 23:44:56.613068 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Apr 13 23:44:56.619009 kernel: ACPI: button: Power Button [PWRF] Apr 13 23:44:56.628874 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Apr 13 23:44:56.640287 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 13 23:44:56.659748 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Apr 13 23:44:56.660482 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Apr 13 23:44:56.660624 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Apr 13 23:44:56.660748 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Apr 13 23:44:56.667037 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Apr 13 23:44:56.681480 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 13 23:44:56.772461 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:56.793549 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 13 23:44:56.795959 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:56.821664 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 13 23:44:56.834721 kernel: mousedev: PS/2 mouse device common for all mice Apr 13 23:44:57.356030 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 13 23:44:57.409613 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 13 23:44:57.428553 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 13 23:44:57.531877 lvm[1434]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 13 23:44:57.581629 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 13 23:44:57.587477 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 13 23:44:57.592829 systemd[1]: Reached target sysinit.target - System Initialization. Apr 13 23:44:57.598791 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 13 23:44:57.605943 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 13 23:44:57.612588 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 13 23:44:57.619934 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 13 23:44:57.624437 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 13 23:44:57.630859 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 13 23:44:57.632693 systemd[1]: Reached target paths.target - Path Units. Apr 13 23:44:57.637713 systemd[1]: Reached target timers.target - Timer Units. Apr 13 23:44:57.706636 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 13 23:44:57.722499 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 13 23:44:57.742788 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 13 23:44:57.757262 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 13 23:44:57.762049 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 13 23:44:57.767019 systemd[1]: Reached target sockets.target - Socket Units. Apr 13 23:44:57.769686 systemd[1]: Reached target basic.target - Basic System. Apr 13 23:44:57.772040 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 13 23:44:57.772073 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 13 23:44:57.787758 systemd[1]: Starting containerd.service - containerd container runtime... Apr 13 23:44:57.794339 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 13 23:44:57.800130 lvm[1438]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 13 23:44:57.802684 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 13 23:44:57.808454 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 13 23:44:57.821070 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 13 23:44:57.828004 jq[1441]: false Apr 13 23:44:57.830890 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 13 23:44:57.897628 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 13 23:44:57.912564 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 13 23:44:57.916518 extend-filesystems[1442]: Found loop3 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found loop4 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found loop5 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found sr0 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda1 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda2 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda3 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found usr Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda4 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda6 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda7 Apr 13 23:44:57.916518 extend-filesystems[1442]: Found vda9 Apr 13 23:44:57.935519 dbus-daemon[1440]: [system] SELinux support is enabled Apr 13 23:44:57.959032 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1405) Apr 13 23:44:57.959054 extend-filesystems[1442]: Checking size of /dev/vda9 Apr 13 23:44:57.959054 extend-filesystems[1442]: Resized partition /dev/vda9 Apr 13 23:44:57.979932 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Apr 13 23:44:57.928143 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 13 23:44:57.980113 extend-filesystems[1458]: resize2fs 1.47.1 (20-May-2024) Apr 13 23:44:57.959151 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 13 23:44:57.974198 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 13 23:44:57.976499 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 13 23:44:57.983534 systemd[1]: Starting update-engine.service - Update Engine... Apr 13 23:44:57.999221 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 13 23:44:58.002676 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 13 23:44:58.022058 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 13 23:44:58.029535 jq[1464]: true Apr 13 23:44:58.035503 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 13 23:44:58.036625 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 13 23:44:58.037436 systemd[1]: motdgen.service: Deactivated successfully. Apr 13 23:44:58.037676 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 13 23:44:58.117581 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 13 23:44:58.118733 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 13 23:44:58.132198 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Apr 13 23:44:58.176712 update_engine[1462]: I20260413 23:44:58.157564 1462 main.cc:92] Flatcar Update Engine starting Apr 13 23:44:58.176712 update_engine[1462]: I20260413 23:44:58.163543 1462 update_check_scheduler.cc:74] Next update check in 9m39s Apr 13 23:44:58.165800 (ntainerd)[1468]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 13 23:44:58.179212 jq[1467]: true Apr 13 23:44:58.179440 extend-filesystems[1458]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Apr 13 23:44:58.179440 extend-filesystems[1458]: old_desc_blocks = 1, new_desc_blocks = 1 Apr 13 23:44:58.179440 extend-filesystems[1458]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Apr 13 23:44:58.177257 systemd-logind[1457]: Watching system buttons on /dev/input/event1 (Power Button) Apr 13 23:44:58.214562 extend-filesystems[1442]: Resized filesystem in /dev/vda9 Apr 13 23:44:58.177277 systemd-logind[1457]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 13 23:44:58.180417 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 13 23:44:58.180590 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 13 23:44:58.189728 systemd-logind[1457]: New seat seat0. Apr 13 23:44:58.226412 systemd-networkd[1394]: eth0: Gained IPv6LL Apr 13 23:44:58.228705 systemd[1]: Started systemd-logind.service - User Login Management. Apr 13 23:44:58.236060 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 13 23:44:58.327668 systemd[1]: Started update-engine.service - Update Engine. Apr 13 23:44:58.337686 tar[1466]: linux-amd64/LICENSE Apr 13 23:44:58.337686 tar[1466]: linux-amd64/helm Apr 13 23:44:58.337056 systemd[1]: Reached target network-online.target - Network is Online. Apr 13 23:44:58.343269 bash[1495]: Updated "/home/core/.ssh/authorized_keys" Apr 13 23:44:58.354051 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Apr 13 23:44:58.378678 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:44:58.406666 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 13 23:44:58.430311 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 13 23:44:58.431663 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 13 23:44:58.436787 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 13 23:44:58.571651 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 13 23:44:58.707607 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 13 23:44:58.712131 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 13 23:44:58.738606 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Apr 13 23:44:58.759736 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 13 23:44:58.827686 systemd[1]: coreos-metadata.service: Deactivated successfully. Apr 13 23:44:58.829843 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Apr 13 23:44:58.897124 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 13 23:44:58.913056 locksmithd[1500]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 13 23:44:59.420574 sshd_keygen[1463]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 13 23:44:59.556591 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 13 23:44:59.575908 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 13 23:44:59.599096 systemd[1]: issuegen.service: Deactivated successfully. Apr 13 23:44:59.599383 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 13 23:44:59.637570 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 13 23:44:59.671872 containerd[1468]: time="2026-04-13T23:44:59.671596205Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 13 23:44:59.800336 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 13 23:44:59.816902 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 13 23:44:59.841576 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Apr 13 23:44:59.868491 systemd[1]: Reached target getty.target - Login Prompts. Apr 13 23:44:59.907147 containerd[1468]: time="2026-04-13T23:44:59.905214679Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.916617 containerd[1468]: time="2026-04-13T23:44:59.916447200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 13 23:44:59.916617 containerd[1468]: time="2026-04-13T23:44:59.916507358Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 13 23:44:59.916617 containerd[1468]: time="2026-04-13T23:44:59.916598880Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 13 23:44:59.918207 containerd[1468]: time="2026-04-13T23:44:59.916864803Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 13 23:44:59.918207 containerd[1468]: time="2026-04-13T23:44:59.916888299Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.921329 containerd[1468]: time="2026-04-13T23:44:59.921086261Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 13 23:44:59.921329 containerd[1468]: time="2026-04-13T23:44:59.921235611Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.923656 containerd[1468]: time="2026-04-13T23:44:59.921772632Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 13 23:44:59.923656 containerd[1468]: time="2026-04-13T23:44:59.921806506Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.923656 containerd[1468]: time="2026-04-13T23:44:59.921854855Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 13 23:44:59.923656 containerd[1468]: time="2026-04-13T23:44:59.921867085Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.926791 containerd[1468]: time="2026-04-13T23:44:59.923105797Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.935801 containerd[1468]: time="2026-04-13T23:44:59.934121200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 13 23:44:59.965560 containerd[1468]: time="2026-04-13T23:44:59.939380233Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 13 23:45:00.015593 containerd[1468]: time="2026-04-13T23:45:00.010497752Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 13 23:45:00.021003 containerd[1468]: time="2026-04-13T23:45:00.020007791Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 13 23:45:00.022920 containerd[1468]: time="2026-04-13T23:45:00.022241024Z" level=info msg="metadata content store policy set" policy=shared Apr 13 23:45:00.057455 containerd[1468]: time="2026-04-13T23:45:00.051940528Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 13 23:45:00.109307 containerd[1468]: time="2026-04-13T23:45:00.061426481Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 13 23:45:00.109307 containerd[1468]: time="2026-04-13T23:45:00.106824203Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 13 23:45:00.148543 containerd[1468]: time="2026-04-13T23:45:00.131876439Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 13 23:45:00.154414 containerd[1468]: time="2026-04-13T23:45:00.150963992Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 13 23:45:00.193542 containerd[1468]: time="2026-04-13T23:45:00.191925374Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 13 23:45:00.205276 containerd[1468]: time="2026-04-13T23:45:00.204132935Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 13 23:45:00.210248 containerd[1468]: time="2026-04-13T23:45:00.209095182Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.211401051Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.211551274Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.211657868Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.211710990Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.211749253Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.211768802Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.213656053Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.214282325Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.214332442Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.215113 containerd[1468]: time="2026-04-13T23:45:00.214348440Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 13 23:45:00.217470 containerd[1468]: time="2026-04-13T23:45:00.217197876Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.222141 containerd[1468]: time="2026-04-13T23:45:00.221715617Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.224018080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.246826330Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.247025823Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.247046927Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.247062065Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.247078755Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.247096320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.228943 containerd[1468]: time="2026-04-13T23:45:00.248623061Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.252437 containerd[1468]: time="2026-04-13T23:45:00.252240344Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.252654 containerd[1468]: time="2026-04-13T23:45:00.252634628Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.252796 containerd[1468]: time="2026-04-13T23:45:00.252781669Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.252924 containerd[1468]: time="2026-04-13T23:45:00.252910737Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 13 23:45:00.253090 containerd[1468]: time="2026-04-13T23:45:00.253076177Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.253145 containerd[1468]: time="2026-04-13T23:45:00.253136487Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.253226 containerd[1468]: time="2026-04-13T23:45:00.253213448Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 13 23:45:00.253461 containerd[1468]: time="2026-04-13T23:45:00.253445635Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 13 23:45:00.255269 containerd[1468]: time="2026-04-13T23:45:00.254957724Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 13 23:45:00.255269 containerd[1468]: time="2026-04-13T23:45:00.255191290Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 13 23:45:00.255269 containerd[1468]: time="2026-04-13T23:45:00.255215764Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 13 23:45:00.255269 containerd[1468]: time="2026-04-13T23:45:00.255227937Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.255269 containerd[1468]: time="2026-04-13T23:45:00.255333389Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 13 23:45:00.256854 containerd[1468]: time="2026-04-13T23:45:00.255956516Z" level=info msg="NRI interface is disabled by configuration." Apr 13 23:45:00.257961 containerd[1468]: time="2026-04-13T23:45:00.257121608Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 13 23:45:00.267801 containerd[1468]: time="2026-04-13T23:45:00.266689513Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 13 23:45:00.267801 containerd[1468]: time="2026-04-13T23:45:00.266931999Z" level=info msg="Connect containerd service" Apr 13 23:45:00.267801 containerd[1468]: time="2026-04-13T23:45:00.267177128Z" level=info msg="using legacy CRI server" Apr 13 23:45:00.267801 containerd[1468]: time="2026-04-13T23:45:00.267190250Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 13 23:45:00.272609 containerd[1468]: time="2026-04-13T23:45:00.272346004Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 13 23:45:00.308126 containerd[1468]: time="2026-04-13T23:45:00.306949803Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 13 23:45:00.336762 containerd[1468]: time="2026-04-13T23:45:00.331478494Z" level=info msg="Start subscribing containerd event" Apr 13 23:45:00.336762 containerd[1468]: time="2026-04-13T23:45:00.333587923Z" level=info msg="Start recovering state" Apr 13 23:45:00.397625 containerd[1468]: time="2026-04-13T23:45:00.338863938Z" level=info msg="Start event monitor" Apr 13 23:45:00.397625 containerd[1468]: time="2026-04-13T23:45:00.395530720Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 13 23:45:00.397625 containerd[1468]: time="2026-04-13T23:45:00.397460765Z" level=info msg="Start snapshots syncer" Apr 13 23:45:00.397625 containerd[1468]: time="2026-04-13T23:45:00.397709015Z" level=info msg="Start cni network conf syncer for default" Apr 13 23:45:00.397625 containerd[1468]: time="2026-04-13T23:45:00.397744032Z" level=info msg="Start streaming server" Apr 13 23:45:00.398285 containerd[1468]: time="2026-04-13T23:45:00.397810392Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 13 23:45:00.398285 containerd[1468]: time="2026-04-13T23:45:00.397936954Z" level=info msg="containerd successfully booted in 0.728921s" Apr 13 23:45:00.398368 systemd[1]: Started containerd.service - containerd container runtime. Apr 13 23:45:00.630805 tar[1466]: linux-amd64/README.md Apr 13 23:45:00.656667 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 13 23:45:02.515781 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:45:02.596578 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 13 23:45:02.609827 systemd[1]: Startup finished in 1.689s (kernel) + 9.147s (initrd) + 13.088s (userspace) = 23.925s. Apr 13 23:45:02.612423 (kubelet)[1553]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:45:07.277372 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 13 23:45:07.302946 systemd[1]: Started sshd@0-10.0.0.25:22-10.0.0.1:54544.service - OpenSSH per-connection server daemon (10.0.0.1:54544). Apr 13 23:45:07.577853 sshd[1565]: Accepted publickey for core from 10.0.0.1 port 54544 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:07.586156 sshd[1565]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:07.683862 systemd-logind[1457]: New session 1 of user core. Apr 13 23:45:07.692405 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 13 23:45:07.720273 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 13 23:45:07.826348 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 13 23:45:07.847491 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 13 23:45:07.919231 (systemd)[1569]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 13 23:45:08.202653 systemd[1569]: Queued start job for default target default.target. Apr 13 23:45:08.221935 systemd[1569]: Created slice app.slice - User Application Slice. Apr 13 23:45:08.222041 systemd[1569]: Reached target paths.target - Paths. Apr 13 23:45:08.222057 systemd[1569]: Reached target timers.target - Timers. Apr 13 23:45:08.231726 systemd[1569]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 13 23:45:08.321701 systemd[1569]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 13 23:45:08.321821 systemd[1569]: Reached target sockets.target - Sockets. Apr 13 23:45:08.321834 systemd[1569]: Reached target basic.target - Basic System. Apr 13 23:45:08.321903 systemd[1569]: Reached target default.target - Main User Target. Apr 13 23:45:08.321933 systemd[1569]: Startup finished in 360ms. Apr 13 23:45:08.322375 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 13 23:45:08.337537 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 13 23:45:08.507015 systemd[1]: Started sshd@1-10.0.0.25:22-10.0.0.1:54558.service - OpenSSH per-connection server daemon (10.0.0.1:54558). Apr 13 23:45:08.619442 sshd[1580]: Accepted publickey for core from 10.0.0.1 port 54558 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:08.630864 sshd[1580]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:08.669849 kubelet[1553]: E0413 23:45:08.666797 1553 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:45:08.689710 systemd-logind[1457]: New session 2 of user core. Apr 13 23:45:08.690085 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:45:08.690208 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:45:08.691277 systemd[1]: kubelet.service: Consumed 6.822s CPU time. Apr 13 23:45:08.709193 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 13 23:45:08.910686 sshd[1580]: pam_unix(sshd:session): session closed for user core Apr 13 23:45:08.978838 systemd[1]: sshd@1-10.0.0.25:22-10.0.0.1:54558.service: Deactivated successfully. Apr 13 23:45:08.986966 systemd[1]: session-2.scope: Deactivated successfully. Apr 13 23:45:08.989198 systemd-logind[1457]: Session 2 logged out. Waiting for processes to exit. Apr 13 23:45:09.015486 systemd[1]: Started sshd@2-10.0.0.25:22-10.0.0.1:54564.service - OpenSSH per-connection server daemon (10.0.0.1:54564). Apr 13 23:45:09.024105 systemd-logind[1457]: Removed session 2. Apr 13 23:45:09.316739 sshd[1588]: Accepted publickey for core from 10.0.0.1 port 54564 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:09.332167 sshd[1588]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:09.399290 systemd-logind[1457]: New session 3 of user core. Apr 13 23:45:09.417956 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 13 23:45:09.534379 sshd[1588]: pam_unix(sshd:session): session closed for user core Apr 13 23:45:09.569107 systemd[1]: sshd@2-10.0.0.25:22-10.0.0.1:54564.service: Deactivated successfully. Apr 13 23:45:09.584963 systemd[1]: session-3.scope: Deactivated successfully. Apr 13 23:45:09.603059 systemd-logind[1457]: Session 3 logged out. Waiting for processes to exit. Apr 13 23:45:09.635093 systemd[1]: Started sshd@3-10.0.0.25:22-10.0.0.1:54570.service - OpenSSH per-connection server daemon (10.0.0.1:54570). Apr 13 23:45:09.693609 systemd-logind[1457]: Removed session 3. Apr 13 23:45:10.019563 sshd[1595]: Accepted publickey for core from 10.0.0.1 port 54570 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:10.027523 sshd[1595]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:10.097894 systemd-logind[1457]: New session 4 of user core. Apr 13 23:45:10.127448 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 13 23:45:10.241335 sshd[1595]: pam_unix(sshd:session): session closed for user core Apr 13 23:45:10.299620 systemd[1]: sshd@3-10.0.0.25:22-10.0.0.1:54570.service: Deactivated successfully. Apr 13 23:45:10.306884 systemd[1]: session-4.scope: Deactivated successfully. Apr 13 23:45:10.321890 systemd-logind[1457]: Session 4 logged out. Waiting for processes to exit. Apr 13 23:45:10.362819 systemd[1]: Started sshd@4-10.0.0.25:22-10.0.0.1:54572.service - OpenSSH per-connection server daemon (10.0.0.1:54572). Apr 13 23:45:10.378058 systemd-logind[1457]: Removed session 4. Apr 13 23:45:10.685006 sshd[1602]: Accepted publickey for core from 10.0.0.1 port 54572 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:10.699686 sshd[1602]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:10.817608 systemd-logind[1457]: New session 5 of user core. Apr 13 23:45:10.831704 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 13 23:45:11.046582 sudo[1605]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 13 23:45:11.046936 sudo[1605]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 13 23:45:11.105911 sudo[1605]: pam_unix(sudo:session): session closed for user root Apr 13 23:45:11.117821 sshd[1602]: pam_unix(sshd:session): session closed for user core Apr 13 23:45:11.201391 systemd[1]: sshd@4-10.0.0.25:22-10.0.0.1:54572.service: Deactivated successfully. Apr 13 23:45:11.236355 systemd[1]: session-5.scope: Deactivated successfully. Apr 13 23:45:11.269077 systemd-logind[1457]: Session 5 logged out. Waiting for processes to exit. Apr 13 23:45:11.305714 systemd[1]: Started sshd@5-10.0.0.25:22-10.0.0.1:54582.service - OpenSSH per-connection server daemon (10.0.0.1:54582). Apr 13 23:45:11.329812 systemd-logind[1457]: Removed session 5. Apr 13 23:45:11.524733 sshd[1610]: Accepted publickey for core from 10.0.0.1 port 54582 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:11.564554 sshd[1610]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:11.597530 systemd-logind[1457]: New session 6 of user core. Apr 13 23:45:11.635848 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 13 23:45:11.789506 sudo[1614]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 13 23:45:11.789888 sudo[1614]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 13 23:45:11.838258 sudo[1614]: pam_unix(sudo:session): session closed for user root Apr 13 23:45:11.933206 sudo[1613]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 13 23:45:11.934420 sudo[1613]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 13 23:45:12.022270 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 13 23:45:12.106596 auditctl[1617]: No rules Apr 13 23:45:12.107967 systemd[1]: audit-rules.service: Deactivated successfully. Apr 13 23:45:12.108167 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 13 23:45:12.156427 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 13 23:45:12.307380 augenrules[1635]: No rules Apr 13 23:45:12.314619 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 13 23:45:12.333583 sudo[1613]: pam_unix(sudo:session): session closed for user root Apr 13 23:45:12.342560 sshd[1610]: pam_unix(sshd:session): session closed for user core Apr 13 23:45:12.367688 systemd[1]: sshd@5-10.0.0.25:22-10.0.0.1:54582.service: Deactivated successfully. Apr 13 23:45:12.375460 systemd[1]: session-6.scope: Deactivated successfully. Apr 13 23:45:12.396756 systemd-logind[1457]: Session 6 logged out. Waiting for processes to exit. Apr 13 23:45:12.498217 systemd[1]: Started sshd@6-10.0.0.25:22-10.0.0.1:54586.service - OpenSSH per-connection server daemon (10.0.0.1:54586). Apr 13 23:45:12.506582 systemd-logind[1457]: Removed session 6. Apr 13 23:45:12.795464 sshd[1643]: Accepted publickey for core from 10.0.0.1 port 54586 ssh2: RSA SHA256:L16zK+ubCZNTurpOZzyaV2jyctPe8ubVYVI0iU3AHjQ Apr 13 23:45:12.817131 sshd[1643]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 13 23:45:12.940003 systemd-logind[1457]: New session 7 of user core. Apr 13 23:45:12.965388 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 13 23:45:13.102640 sudo[1646]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 13 23:45:13.105606 sudo[1646]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 13 23:45:15.955566 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 13 23:45:15.995160 (dockerd)[1665]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 13 23:45:18.731143 dockerd[1665]: time="2026-04-13T23:45:18.726918737Z" level=info msg="Starting up" Apr 13 23:45:18.752799 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 13 23:45:18.817732 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:45:19.914159 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:45:19.960563 (kubelet)[1693]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:45:20.082814 dockerd[1665]: time="2026-04-13T23:45:20.081072171Z" level=info msg="Loading containers: start." Apr 13 23:45:21.478110 kernel: Initializing XFRM netlink socket Apr 13 23:45:22.213719 systemd-networkd[1394]: docker0: Link UP Apr 13 23:45:22.432249 dockerd[1665]: time="2026-04-13T23:45:22.431871443Z" level=info msg="Loading containers: done." Apr 13 23:45:22.562815 kubelet[1693]: E0413 23:45:22.562645 1693 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:45:22.570881 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2778493764-merged.mount: Deactivated successfully. Apr 13 23:45:22.587216 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:45:22.587871 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:45:22.589362 systemd[1]: kubelet.service: Consumed 2.490s CPU time. Apr 13 23:45:22.596768 dockerd[1665]: time="2026-04-13T23:45:22.591123589Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 13 23:45:22.629287 dockerd[1665]: time="2026-04-13T23:45:22.628919540Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 13 23:45:22.635591 dockerd[1665]: time="2026-04-13T23:45:22.635109838Z" level=info msg="Daemon has completed initialization" Apr 13 23:45:23.496815 dockerd[1665]: time="2026-04-13T23:45:23.494516532Z" level=info msg="API listen on /run/docker.sock" Apr 13 23:45:23.497966 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 13 23:45:30.252958 containerd[1468]: time="2026-04-13T23:45:30.251580743Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.10\"" Apr 13 23:45:32.760368 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 13 23:45:32.788475 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:45:32.874896 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount328130829.mount: Deactivated successfully. Apr 13 23:45:33.474936 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:45:33.512208 (kubelet)[1843]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:45:36.496130 kubelet[1843]: E0413 23:45:36.493659 1843 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:45:36.504784 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:45:36.504943 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:45:36.508552 systemd[1]: kubelet.service: Consumed 2.429s CPU time. Apr 13 23:45:42.540771 containerd[1468]: time="2026-04-13T23:45:42.535947907Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:42.581132 containerd[1468]: time="2026-04-13T23:45:42.540270779Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.10: active requests=0, bytes read=29988857" Apr 13 23:45:42.588826 containerd[1468]: time="2026-04-13T23:45:42.587720977Z" level=info msg="ImageCreate event name:\"sha256:e1586f2f8635ddb8eb665e8155e4aadb66d9ca499906c11db63a79ae66456b74\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:42.719280 containerd[1468]: time="2026-04-13T23:45:42.718844320Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:bbff81e41af4bfca88a1d05a066a48e12e2689c534d073a8c688e3ad6c8701e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:42.731149 containerd[1468]: time="2026-04-13T23:45:42.729928137Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.10\" with image id \"sha256:e1586f2f8635ddb8eb665e8155e4aadb66d9ca499906c11db63a79ae66456b74\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:bbff81e41af4bfca88a1d05a066a48e12e2689c534d073a8c688e3ad6c8701e3\", size \"29986018\" in 12.476491715s" Apr 13 23:45:42.734688 containerd[1468]: time="2026-04-13T23:45:42.732858777Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.10\" returns image reference \"sha256:e1586f2f8635ddb8eb665e8155e4aadb66d9ca499906c11db63a79ae66456b74\"" Apr 13 23:45:42.785545 containerd[1468]: time="2026-04-13T23:45:42.785447326Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.10\"" Apr 13 23:45:43.710127 update_engine[1462]: I20260413 23:45:43.708998 1462 update_attempter.cc:509] Updating boot flags... Apr 13 23:45:43.852214 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1908) Apr 13 23:45:46.701220 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Apr 13 23:45:46.736430 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:45:47.287828 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:45:47.304756 (kubelet)[1925]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:45:51.644442 containerd[1468]: time="2026-04-13T23:45:51.642707265Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:51.693348 containerd[1468]: time="2026-04-13T23:45:51.682959581Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.10: active requests=0, bytes read=26021841" Apr 13 23:45:51.784880 containerd[1468]: time="2026-04-13T23:45:51.784698444Z" level=info msg="ImageCreate event name:\"sha256:26db35ccbf4330e5ada4a2786276aac158e92aced08cecce6cb614146e224230\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:51.995752 kubelet[1925]: E0413 23:45:51.995289 1925 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:45:52.005350 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:45:52.005800 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:45:52.007356 systemd[1]: kubelet.service: Consumed 3.769s CPU time. Apr 13 23:45:52.092292 containerd[1468]: time="2026-04-13T23:45:52.092034057Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:b0880d6ee19f2b9148d3d37008c5ee9fc73976e8edad4d0709f11d32ab3ee709\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:52.164066 containerd[1468]: time="2026-04-13T23:45:52.163844949Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.10\" with image id \"sha256:26db35ccbf4330e5ada4a2786276aac158e92aced08cecce6cb614146e224230\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:b0880d6ee19f2b9148d3d37008c5ee9fc73976e8edad4d0709f11d32ab3ee709\", size \"27552094\" in 9.378350076s" Apr 13 23:45:52.164066 containerd[1468]: time="2026-04-13T23:45:52.163912341Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.10\" returns image reference \"sha256:26db35ccbf4330e5ada4a2786276aac158e92aced08cecce6cb614146e224230\"" Apr 13 23:45:52.207076 containerd[1468]: time="2026-04-13T23:45:52.206108955Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.10\"" Apr 13 23:45:59.815900 containerd[1468]: time="2026-04-13T23:45:59.814822822Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:45:59.821387 containerd[1468]: time="2026-04-13T23:45:59.820857261Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.10: active requests=0, bytes read=20162685" Apr 13 23:45:59.836497 containerd[1468]: time="2026-04-13T23:45:59.835556434Z" level=info msg="ImageCreate event name:\"sha256:7f5d3f3b598c23877c138d7739627d8f0160b0a91d321108e9b5affad54f85f7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:00.121023 containerd[1468]: time="2026-04-13T23:46:00.120189473Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:dc1a1aec3bb0ed126b1adff795935124f719969356b24a159fc1a2a0883b89bc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:00.236705 containerd[1468]: time="2026-04-13T23:46:00.235950174Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.10\" with image id \"sha256:7f5d3f3b598c23877c138d7739627d8f0160b0a91d321108e9b5affad54f85f7\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:dc1a1aec3bb0ed126b1adff795935124f719969356b24a159fc1a2a0883b89bc\", size \"21692956\" in 8.029114951s" Apr 13 23:46:00.238124 containerd[1468]: time="2026-04-13T23:46:00.237926490Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.10\" returns image reference \"sha256:7f5d3f3b598c23877c138d7739627d8f0160b0a91d321108e9b5affad54f85f7\"" Apr 13 23:46:00.282853 containerd[1468]: time="2026-04-13T23:46:00.282523362Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.10\"" Apr 13 23:46:02.244259 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Apr 13 23:46:02.322131 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:46:02.925101 (kubelet)[1950]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:46:02.927643 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:46:06.979682 kubelet[1950]: E0413 23:46:06.977952 1950 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:46:06.990154 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:46:06.991645 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:46:06.994710 systemd[1]: kubelet.service: Consumed 3.453s CPU time. Apr 13 23:46:13.179731 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2696210892.mount: Deactivated successfully. Apr 13 23:46:17.221821 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Apr 13 23:46:17.267431 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:46:17.934957 (kubelet)[1970]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:46:17.937283 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:46:18.715615 containerd[1468]: time="2026-04-13T23:46:18.714606906Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:18.727118 containerd[1468]: time="2026-04-13T23:46:18.725155974Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.10: active requests=0, bytes read=31828657" Apr 13 23:46:18.741135 containerd[1468]: time="2026-04-13T23:46:18.737921325Z" level=info msg="ImageCreate event name:\"sha256:bed75257625288e2a7e106a7fe6bf8373eaa2bc2b14805d32033c7655e882f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:18.981897 containerd[1468]: time="2026-04-13T23:46:18.980692453Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e8151e38ef22f032dba686cc1bba5a3e525dedbe2d549fa44e653fe79426e261\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:18.992897 containerd[1468]: time="2026-04-13T23:46:18.992708649Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.10\" with image id \"sha256:bed75257625288e2a7e106a7fe6bf8373eaa2bc2b14805d32033c7655e882f76\", repo tag \"registry.k8s.io/kube-proxy:v1.33.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:e8151e38ef22f032dba686cc1bba5a3e525dedbe2d549fa44e653fe79426e261\", size \"31827782\" in 18.708366245s" Apr 13 23:46:18.996531 containerd[1468]: time="2026-04-13T23:46:18.992812639Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.10\" returns image reference \"sha256:bed75257625288e2a7e106a7fe6bf8373eaa2bc2b14805d32033c7655e882f76\"" Apr 13 23:46:19.030352 containerd[1468]: time="2026-04-13T23:46:19.028890648Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Apr 13 23:46:21.935211 kubelet[1970]: E0413 23:46:21.933120 1970 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:46:21.940642 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:46:21.941152 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:46:21.951542 systemd[1]: kubelet.service: Consumed 3.422s CPU time. Apr 13 23:46:22.299071 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3126370592.mount: Deactivated successfully. Apr 13 23:46:32.231794 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Apr 13 23:46:32.290837 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:46:33.313591 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:46:33.315024 (kubelet)[2036]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:46:41.608007 kubelet[2036]: E0413 23:46:41.604649 2036 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:46:41.620542 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:46:41.624901 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:46:41.665988 systemd[1]: kubelet.service: Consumed 6.573s CPU time. Apr 13 23:46:42.213861 containerd[1468]: time="2026-04-13T23:46:42.213112124Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:42.215621 containerd[1468]: time="2026-04-13T23:46:42.214612924Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20941714" Apr 13 23:46:42.319641 containerd[1468]: time="2026-04-13T23:46:42.318219969Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:42.904372 containerd[1468]: time="2026-04-13T23:46:42.904125555Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:42.973540 containerd[1468]: time="2026-04-13T23:46:42.973371420Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 23.940724628s" Apr 13 23:46:42.973540 containerd[1468]: time="2026-04-13T23:46:42.973509725Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Apr 13 23:46:43.003341 containerd[1468]: time="2026-04-13T23:46:43.002876380Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 13 23:46:46.127845 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3666844151.mount: Deactivated successfully. Apr 13 23:46:46.134403 containerd[1468]: time="2026-04-13T23:46:46.127083723Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:46.134403 containerd[1468]: time="2026-04-13T23:46:46.132826121Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321070" Apr 13 23:46:46.165663 containerd[1468]: time="2026-04-13T23:46:46.165309626Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:46.489358 containerd[1468]: time="2026-04-13T23:46:46.470605730Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:46:46.626047 containerd[1468]: time="2026-04-13T23:46:46.624936104Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 3.621606496s" Apr 13 23:46:46.631857 containerd[1468]: time="2026-04-13T23:46:46.628109228Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Apr 13 23:46:46.700435 containerd[1468]: time="2026-04-13T23:46:46.695658391Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Apr 13 23:46:50.421709 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2912540591.mount: Deactivated successfully. Apr 13 23:46:51.719798 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Apr 13 23:46:51.900129 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:46:53.595274 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:46:53.601275 (kubelet)[2071]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:47:02.374779 kubelet[2071]: E0413 23:47:02.369355 2071 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:47:02.390241 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:47:02.391268 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:47:02.400931 systemd[1]: kubelet.service: Consumed 6.808s CPU time. Apr 13 23:47:12.499824 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Apr 13 23:47:12.565412 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:47:13.882123 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:47:13.882929 (kubelet)[2088]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:47:19.828710 kubelet[2088]: E0413 23:47:19.827721 2088 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:47:19.851671 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:47:19.854894 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:47:19.856512 systemd[1]: kubelet.service: Consumed 4.430s CPU time. Apr 13 23:47:26.914540 containerd[1468]: time="2026-04-13T23:47:26.912607173Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:47:26.919158 containerd[1468]: time="2026-04-13T23:47:26.918653902Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=23718278" Apr 13 23:47:27.110772 containerd[1468]: time="2026-04-13T23:47:27.109714342Z" level=info msg="ImageCreate event name:\"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:47:27.724595 containerd[1468]: time="2026-04-13T23:47:27.723949602Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 13 23:47:27.918945 containerd[1468]: time="2026-04-13T23:47:27.914474814Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"23716032\" in 41.214703894s" Apr 13 23:47:27.932366 containerd[1468]: time="2026-04-13T23:47:27.931818602Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\"" Apr 13 23:47:29.976512 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Apr 13 23:47:30.141673 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:47:31.307503 (kubelet)[2163]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:47:31.307894 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:47:35.452472 kubelet[2163]: E0413 23:47:35.439532 2163 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:47:35.465953 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:47:35.475357 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:47:35.482687 systemd[1]: kubelet.service: Consumed 3.158s CPU time. Apr 13 23:47:45.798576 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Apr 13 23:47:45.994882 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:47:47.393568 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:47:47.412387 (kubelet)[2192]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:47:52.758679 kubelet[2192]: E0413 23:47:52.751774 2192 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:47:52.778814 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:47:52.785528 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:47:52.796576 systemd[1]: kubelet.service: Consumed 4.207s CPU time, 107.9M memory peak, 0B memory swap peak. Apr 13 23:48:02.939761 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Apr 13 23:48:02.993744 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:48:04.587805 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:48:04.692949 (kubelet)[2211]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 13 23:48:16.634581 kubelet[2211]: E0413 23:48:16.618302 2211 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 13 23:48:16.662093 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 13 23:48:16.676706 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 13 23:48:16.721002 systemd[1]: kubelet.service: Consumed 9.307s CPU time. Apr 13 23:48:18.360886 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:48:18.417485 systemd[1]: kubelet.service: Consumed 9.307s CPU time. Apr 13 23:48:18.826851 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:48:19.816263 systemd[1]: Reloading requested from client PID 2229 ('systemctl') (unit session-7.scope)... Apr 13 23:48:19.819218 systemd[1]: Reloading... Apr 13 23:48:27.731118 zram_generator::config[2268]: No configuration found. Apr 13 23:48:34.279587 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 13 23:48:38.138785 systemd[1]: Reloading finished in 18313 ms. Apr 13 23:48:39.714436 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:48:39.817600 (kubelet)[2306]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 13 23:48:40.295873 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:48:40.383044 systemd[1]: kubelet.service: Deactivated successfully. Apr 13 23:48:40.395203 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:48:40.418541 systemd[1]: kubelet.service: Consumed 1.196s CPU time, 44.1M memory peak, 0B memory swap peak. Apr 13 23:48:40.707170 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:48:42.803629 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:48:42.807050 (kubelet)[2322]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 13 23:48:51.182505 kubelet[2322]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 13 23:48:51.182505 kubelet[2322]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 13 23:48:51.182505 kubelet[2322]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 13 23:48:51.224256 kubelet[2322]: I0413 23:48:51.206841 2322 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 13 23:48:55.684178 kubelet[2322]: I0413 23:48:55.683269 2322 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 13 23:48:55.690364 kubelet[2322]: I0413 23:48:55.685740 2322 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 13 23:48:55.776384 kubelet[2322]: I0413 23:48:55.775876 2322 server.go:956] "Client rotation is on, will bootstrap in background" Apr 13 23:48:56.868893 kubelet[2322]: E0413 23:48:56.841887 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:48:57.697493 kubelet[2322]: I0413 23:48:57.696368 2322 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 13 23:48:58.427047 kubelet[2322]: E0413 23:48:58.425346 2322 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 13 23:48:58.429958 kubelet[2322]: I0413 23:48:58.428236 2322 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 13 23:48:59.188759 kubelet[2322]: E0413 23:48:59.186420 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:48:59.312103 kubelet[2322]: I0413 23:48:59.308587 2322 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 13 23:48:59.430367 kubelet[2322]: I0413 23:48:59.425458 2322 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 13 23:48:59.493873 kubelet[2322]: I0413 23:48:59.432772 2322 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 13 23:48:59.506457 kubelet[2322]: I0413 23:48:59.503778 2322 topology_manager.go:138] "Creating topology manager with none policy" Apr 13 23:48:59.518641 kubelet[2322]: I0413 23:48:59.515609 2322 container_manager_linux.go:303] "Creating device plugin manager" Apr 13 23:48:59.554694 kubelet[2322]: I0413 23:48:59.552204 2322 state_mem.go:36] "Initialized new in-memory state store" Apr 13 23:48:59.635015 kubelet[2322]: I0413 23:48:59.632356 2322 kubelet.go:480] "Attempting to sync node with API server" Apr 13 23:48:59.639615 kubelet[2322]: I0413 23:48:59.637758 2322 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 13 23:48:59.652992 kubelet[2322]: I0413 23:48:59.649569 2322 kubelet.go:386] "Adding apiserver pod source" Apr 13 23:48:59.676487 kubelet[2322]: I0413 23:48:59.668251 2322 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 13 23:48:59.796440 kubelet[2322]: E0413 23:48:59.794790 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:48:59.796440 kubelet[2322]: E0413 23:48:59.796197 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:48:59.821627 kubelet[2322]: I0413 23:48:59.820415 2322 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 13 23:48:59.991575 kubelet[2322]: I0413 23:48:59.989165 2322 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 13 23:49:00.000051 kubelet[2322]: W0413 23:48:59.999619 2322 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 13 23:49:00.242949 kubelet[2322]: I0413 23:49:00.239834 2322 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 13 23:49:00.242949 kubelet[2322]: I0413 23:49:00.318656 2322 server.go:1289] "Started kubelet" Apr 13 23:49:00.382487 kubelet[2322]: I0413 23:49:00.369747 2322 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 13 23:49:00.406155 kubelet[2322]: I0413 23:49:00.401308 2322 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 13 23:49:00.409205 kubelet[2322]: E0413 23:49:00.372362 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.25:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:49:00.418266 kubelet[2322]: I0413 23:49:00.418081 2322 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 13 23:49:00.540896 kubelet[2322]: I0413 23:49:00.540402 2322 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 13 23:49:00.544251 kubelet[2322]: I0413 23:49:00.543789 2322 server.go:317] "Adding debug handlers to kubelet server" Apr 13 23:49:00.553243 kubelet[2322]: I0413 23:49:00.552838 2322 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 13 23:49:00.602107 kubelet[2322]: I0413 23:49:00.600306 2322 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 13 23:49:00.610893 kubelet[2322]: E0413 23:49:00.606508 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:00.672212 kubelet[2322]: I0413 23:49:00.668147 2322 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 13 23:49:00.676004 kubelet[2322]: I0413 23:49:00.675781 2322 reconciler.go:26] "Reconciler: start to sync state" Apr 13 23:49:00.683304 kubelet[2322]: E0413 23:49:00.682998 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="200ms" Apr 13 23:49:00.683747 kubelet[2322]: I0413 23:49:00.683604 2322 factory.go:223] Registration of the systemd container factory successfully Apr 13 23:49:00.691136 kubelet[2322]: I0413 23:49:00.686996 2322 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 13 23:49:00.705069 kubelet[2322]: E0413 23:49:00.699471 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:49:00.759372 kubelet[2322]: E0413 23:49:00.758195 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:00.780269 kubelet[2322]: E0413 23:49:00.779992 2322 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 13 23:49:00.875370 kubelet[2322]: E0413 23:49:00.873822 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:00.886198 kubelet[2322]: I0413 23:49:00.886048 2322 factory.go:223] Registration of the containerd container factory successfully Apr 13 23:49:00.983835 kubelet[2322]: E0413 23:49:00.980201 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="400ms" Apr 13 23:49:00.990872 kubelet[2322]: E0413 23:49:00.990371 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.005447 kubelet[2322]: I0413 23:49:01.005117 2322 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 13 23:49:01.021369 kubelet[2322]: I0413 23:49:01.016910 2322 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 13 23:49:01.023965 kubelet[2322]: I0413 23:49:01.023078 2322 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 13 23:49:01.036140 kubelet[2322]: I0413 23:49:01.035164 2322 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 13 23:49:01.036140 kubelet[2322]: I0413 23:49:01.035365 2322 kubelet.go:2436] "Starting kubelet main sync loop" Apr 13 23:49:01.036140 kubelet[2322]: E0413 23:49:01.035485 2322 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 13 23:49:01.133691 kubelet[2322]: E0413 23:49:01.125809 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.133691 kubelet[2322]: E0413 23:49:01.125924 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:49:01.174103 kubelet[2322]: E0413 23:49:01.173504 2322 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 13 23:49:01.187171 kubelet[2322]: E0413 23:49:01.186856 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:49:01.231205 kubelet[2322]: E0413 23:49:01.230084 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.363830 kubelet[2322]: E0413 23:49:01.359901 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.379657 kubelet[2322]: E0413 23:49:01.379446 2322 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:49:01.380827 kubelet[2322]: E0413 23:49:01.379719 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:49:01.465632 kubelet[2322]: E0413 23:49:01.434411 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="800ms" Apr 13 23:49:01.481824 kubelet[2322]: E0413 23:49:01.473090 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.597355 kubelet[2322]: E0413 23:49:01.595301 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.739271 kubelet[2322]: E0413 23:49:01.736328 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.797647 kubelet[2322]: E0413 23:49:01.794940 2322 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:49:01.873551 kubelet[2322]: E0413 23:49:01.869235 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.986156 kubelet[2322]: E0413 23:49:01.985671 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:01.998412 kubelet[2322]: I0413 23:49:01.994847 2322 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 13 23:49:02.005816 kubelet[2322]: I0413 23:49:02.004082 2322 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 13 23:49:02.018114 kubelet[2322]: I0413 23:49:02.015521 2322 state_mem.go:36] "Initialized new in-memory state store" Apr 13 23:49:02.124520 kubelet[2322]: E0413 23:49:02.113857 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.135528 kubelet[2322]: I0413 23:49:02.131789 2322 policy_none.go:49] "None policy: Start" Apr 13 23:49:02.140520 kubelet[2322]: I0413 23:49:02.137258 2322 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 13 23:49:02.161948 kubelet[2322]: I0413 23:49:02.159721 2322 state_mem.go:35] "Initializing new in-memory state store" Apr 13 23:49:02.208093 kubelet[2322]: E0413 23:49:02.204228 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:49:02.228769 kubelet[2322]: E0413 23:49:02.225496 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.304143 kubelet[2322]: E0413 23:49:02.303451 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="1.6s" Apr 13 23:49:02.332491 kubelet[2322]: E0413 23:49:02.332193 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.425846 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 13 23:49:02.442357 kubelet[2322]: E0413 23:49:02.441932 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.548420 kubelet[2322]: E0413 23:49:02.545946 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.619220 kubelet[2322]: E0413 23:49:02.612204 2322 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:49:02.683662 kubelet[2322]: E0413 23:49:02.681382 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.716966 kubelet[2322]: E0413 23:49:02.715179 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:49:02.729517 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 13 23:49:02.791353 kubelet[2322]: E0413 23:49:02.788782 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:02.931089 kubelet[2322]: E0413 23:49:02.924267 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:03.024107 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 13 23:49:03.041565 kubelet[2322]: E0413 23:49:03.041198 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:03.182443 kubelet[2322]: E0413 23:49:03.181330 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:49:03.208167 kubelet[2322]: E0413 23:49:03.204595 2322 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 13 23:49:03.212433 kubelet[2322]: I0413 23:49:03.212003 2322 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 13 23:49:03.215408 kubelet[2322]: I0413 23:49:03.215043 2322 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 13 23:49:03.222181 kubelet[2322]: I0413 23:49:03.221896 2322 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 13 23:49:03.230523 kubelet[2322]: E0413 23:49:03.228340 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:49:03.294195 kubelet[2322]: E0413 23:49:03.293746 2322 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 13 23:49:03.310260 kubelet[2322]: E0413 23:49:03.308934 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:49:03.477749 kubelet[2322]: I0413 23:49:03.476124 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:03.521522 kubelet[2322]: E0413 23:49:03.515926 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:03.820839 kubelet[2322]: E0413 23:49:03.819116 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:49:03.913374 kubelet[2322]: I0413 23:49:03.906523 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:03.938635 kubelet[2322]: E0413 23:49:03.929207 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="3.2s" Apr 13 23:49:03.938635 kubelet[2322]: E0413 23:49:03.929131 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:04.335336 kubelet[2322]: E0413 23:49:04.334847 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:49:04.401343 kubelet[2322]: I0413 23:49:04.399285 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:04.409151 kubelet[2322]: I0413 23:49:04.408903 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3ffefc5deebb452093137d8633d90466-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3ffefc5deebb452093137d8633d90466\") " pod="kube-system/kube-apiserver-localhost" Apr 13 23:49:04.409151 kubelet[2322]: I0413 23:49:04.409151 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3ffefc5deebb452093137d8633d90466-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3ffefc5deebb452093137d8633d90466\") " pod="kube-system/kube-apiserver-localhost" Apr 13 23:49:04.410591 kubelet[2322]: I0413 23:49:04.409181 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3ffefc5deebb452093137d8633d90466-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3ffefc5deebb452093137d8633d90466\") " pod="kube-system/kube-apiserver-localhost" Apr 13 23:49:04.416667 kubelet[2322]: E0413 23:49:04.414779 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:04.541116 kubelet[2322]: I0413 23:49:04.537173 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:49:04.593993 kubelet[2322]: I0413 23:49:04.586928 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:49:04.604927 kubelet[2322]: I0413 23:49:04.601905 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:49:04.604927 kubelet[2322]: I0413 23:49:04.602231 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:49:04.604927 kubelet[2322]: I0413 23:49:04.603844 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:49:04.723549 kubelet[2322]: I0413 23:49:04.709795 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/39798d73a6894e44ae801eb773bf9a39-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"39798d73a6894e44ae801eb773bf9a39\") " pod="kube-system/kube-scheduler-localhost" Apr 13 23:49:04.947005 kubelet[2322]: E0413 23:49:04.946275 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:49:04.947501 systemd[1]: Created slice kubepods-burstable-pod3ffefc5deebb452093137d8633d90466.slice - libcontainer container kubepods-burstable-pod3ffefc5deebb452093137d8633d90466.slice. Apr 13 23:49:05.105160 kubelet[2322]: E0413 23:49:05.103928 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:05.145125 kubelet[2322]: E0413 23:49:05.142632 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:05.191486 containerd[1468]: time="2026-04-13T23:49:05.191078767Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3ffefc5deebb452093137d8633d90466,Namespace:kube-system,Attempt:0,}" Apr 13 23:49:05.201560 systemd[1]: Created slice kubepods-burstable-podebf8e820819e4b80bc03d078b9ba80f5.slice - libcontainer container kubepods-burstable-podebf8e820819e4b80bc03d078b9ba80f5.slice. Apr 13 23:49:05.331928 kubelet[2322]: E0413 23:49:05.328464 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:05.346639 kubelet[2322]: E0413 23:49:05.342294 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:05.350636 kubelet[2322]: I0413 23:49:05.350338 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:05.358140 kubelet[2322]: E0413 23:49:05.357699 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:05.365281 containerd[1468]: time="2026-04-13T23:49:05.365119979Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:ebf8e820819e4b80bc03d078b9ba80f5,Namespace:kube-system,Attempt:0,}" Apr 13 23:49:05.415039 systemd[1]: Created slice kubepods-burstable-pod39798d73a6894e44ae801eb773bf9a39.slice - libcontainer container kubepods-burstable-pod39798d73a6894e44ae801eb773bf9a39.slice. Apr 13 23:49:05.485896 kubelet[2322]: E0413 23:49:05.478718 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:49:05.568104 kubelet[2322]: E0413 23:49:05.565666 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:05.639415 kubelet[2322]: E0413 23:49:05.613041 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:05.682094 containerd[1468]: time="2026-04-13T23:49:05.680234681Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:39798d73a6894e44ae801eb773bf9a39,Namespace:kube-system,Attempt:0,}" Apr 13 23:49:06.197013 kubelet[2322]: E0413 23:49:06.184772 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.25:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:49:07.157251 kubelet[2322]: E0413 23:49:07.156871 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="6.4s" Apr 13 23:49:07.163088 kubelet[2322]: I0413 23:49:07.156963 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:07.301685 kubelet[2322]: E0413 23:49:07.295948 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:07.622434 kubelet[2322]: E0413 23:49:07.596474 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:49:09.391055 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount377308963.mount: Deactivated successfully. Apr 13 23:49:09.520824 kubelet[2322]: E0413 23:49:09.518794 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:49:09.632895 containerd[1468]: time="2026-04-13T23:49:09.632088780Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 13 23:49:09.695139 containerd[1468]: time="2026-04-13T23:49:09.694704272Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=311988" Apr 13 23:49:09.785247 containerd[1468]: time="2026-04-13T23:49:09.783288204Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 13 23:49:09.809254 containerd[1468]: time="2026-04-13T23:49:09.808349711Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 13 23:49:09.859077 containerd[1468]: time="2026-04-13T23:49:09.854351422Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 13 23:49:09.884509 containerd[1468]: time="2026-04-13T23:49:09.884202572Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 13 23:49:10.199019 containerd[1468]: time="2026-04-13T23:49:10.126678055Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 13 23:49:10.285875 containerd[1468]: time="2026-04-13T23:49:10.285391601Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 4.603472368s" Apr 13 23:49:10.321920 containerd[1468]: time="2026-04-13T23:49:10.321306745Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 13 23:49:10.403326 containerd[1468]: time="2026-04-13T23:49:10.399237514Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 5.032473304s" Apr 13 23:49:10.416843 containerd[1468]: time="2026-04-13T23:49:10.412410075Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 5.216818574s" Apr 13 23:49:10.496425 kubelet[2322]: E0413 23:49:10.493331 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:49:10.680953 kubelet[2322]: I0413 23:49:10.677609 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:10.687432 kubelet[2322]: E0413 23:49:10.668223 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:49:10.892415 kubelet[2322]: E0413 23:49:10.886917 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:11.153199 containerd[1468]: time="2026-04-13T23:49:11.151304592Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 13 23:49:11.159947 containerd[1468]: time="2026-04-13T23:49:11.156803139Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 13 23:49:11.180431 containerd[1468]: time="2026-04-13T23:49:11.176163258Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 13 23:49:11.187326 containerd[1468]: time="2026-04-13T23:49:11.181043578Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 13 23:49:11.206956 containerd[1468]: time="2026-04-13T23:49:11.206493800Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 13 23:49:11.206956 containerd[1468]: time="2026-04-13T23:49:11.206782394Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 13 23:49:11.206956 containerd[1468]: time="2026-04-13T23:49:11.206798783Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 13 23:49:11.207953 containerd[1468]: time="2026-04-13T23:49:11.207615059Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 13 23:49:11.271492 containerd[1468]: time="2026-04-13T23:49:11.238720086Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 13 23:49:11.271492 containerd[1468]: time="2026-04-13T23:49:11.239702030Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 13 23:49:11.271492 containerd[1468]: time="2026-04-13T23:49:11.239718911Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 13 23:49:11.274948 containerd[1468]: time="2026-04-13T23:49:11.274332733Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 13 23:49:11.612707 kubelet[2322]: E0413 23:49:11.609310 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:49:11.742789 systemd[1]: Started cri-containerd-14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210.scope - libcontainer container 14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210. Apr 13 23:49:11.766126 systemd[1]: Started cri-containerd-464186f617a7e92d2ff5d3ace6ac415dd36daa9d811ca27383da08d6f78e0f5b.scope - libcontainer container 464186f617a7e92d2ff5d3ace6ac415dd36daa9d811ca27383da08d6f78e0f5b. Apr 13 23:49:11.802199 systemd[1]: Started cri-containerd-aec21221d5f2199c2f70d861387febe1830c44719b6e5cf16243d723e13dad81.scope - libcontainer container aec21221d5f2199c2f70d861387febe1830c44719b6e5cf16243d723e13dad81. Apr 13 23:49:12.362894 containerd[1468]: time="2026-04-13T23:49:12.362696239Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:ebf8e820819e4b80bc03d078b9ba80f5,Namespace:kube-system,Attempt:0,} returns sandbox id \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\"" Apr 13 23:49:12.413598 containerd[1468]: time="2026-04-13T23:49:12.411938888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3ffefc5deebb452093137d8633d90466,Namespace:kube-system,Attempt:0,} returns sandbox id \"464186f617a7e92d2ff5d3ace6ac415dd36daa9d811ca27383da08d6f78e0f5b\"" Apr 13 23:49:12.423331 kubelet[2322]: E0413 23:49:12.423104 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:12.578863 kubelet[2322]: E0413 23:49:12.578257 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:12.621966 containerd[1468]: time="2026-04-13T23:49:12.612359582Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:39798d73a6894e44ae801eb773bf9a39,Namespace:kube-system,Attempt:0,} returns sandbox id \"aec21221d5f2199c2f70d861387febe1830c44719b6e5cf16243d723e13dad81\"" Apr 13 23:49:12.814106 kubelet[2322]: E0413 23:49:12.807539 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:12.917948 containerd[1468]: time="2026-04-13T23:49:12.916050577Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 13 23:49:12.969481 containerd[1468]: time="2026-04-13T23:49:12.961760744Z" level=info msg="CreateContainer within sandbox \"464186f617a7e92d2ff5d3ace6ac415dd36daa9d811ca27383da08d6f78e0f5b\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 13 23:49:13.479120 kubelet[2322]: E0413 23:49:13.415610 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:49:13.482257 containerd[1468]: time="2026-04-13T23:49:13.471681391Z" level=info msg="CreateContainer within sandbox \"aec21221d5f2199c2f70d861387febe1830c44719b6e5cf16243d723e13dad81\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 13 23:49:13.650559 kubelet[2322]: E0413 23:49:13.650260 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.25:6443: connect: connection refused" interval="7s" Apr 13 23:49:13.659450 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount18188264.mount: Deactivated successfully. Apr 13 23:49:13.858709 containerd[1468]: time="2026-04-13T23:49:13.854835293Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71\"" Apr 13 23:49:13.893957 containerd[1468]: time="2026-04-13T23:49:13.888953585Z" level=info msg="CreateContainer within sandbox \"464186f617a7e92d2ff5d3ace6ac415dd36daa9d811ca27383da08d6f78e0f5b\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"e54945a1bfd0258cf7db1259dd6ee5206b39c872285f52d1f236699ec8573c38\"" Apr 13 23:49:13.999118 containerd[1468]: time="2026-04-13T23:49:13.995256322Z" level=info msg="StartContainer for \"06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71\"" Apr 13 23:49:14.000053 containerd[1468]: time="2026-04-13T23:49:13.999914137Z" level=info msg="StartContainer for \"e54945a1bfd0258cf7db1259dd6ee5206b39c872285f52d1f236699ec8573c38\"" Apr 13 23:49:14.123264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3076073964.mount: Deactivated successfully. Apr 13 23:49:14.242119 containerd[1468]: time="2026-04-13T23:49:14.239351285Z" level=info msg="CreateContainer within sandbox \"aec21221d5f2199c2f70d861387febe1830c44719b6e5cf16243d723e13dad81\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"de2aec49366b77bdce86bcd450ae6f023d6693e0b68349e16e2a236ef3a6c264\"" Apr 13 23:49:14.259792 containerd[1468]: time="2026-04-13T23:49:14.259617627Z" level=info msg="StartContainer for \"de2aec49366b77bdce86bcd450ae6f023d6693e0b68349e16e2a236ef3a6c264\"" Apr 13 23:49:14.602306 systemd[1]: Started cri-containerd-06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71.scope - libcontainer container 06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71. Apr 13 23:49:14.671781 systemd[1]: Started cri-containerd-e54945a1bfd0258cf7db1259dd6ee5206b39c872285f52d1f236699ec8573c38.scope - libcontainer container e54945a1bfd0258cf7db1259dd6ee5206b39c872285f52d1f236699ec8573c38. Apr 13 23:49:14.848730 systemd[1]: run-containerd-runc-k8s.io-de2aec49366b77bdce86bcd450ae6f023d6693e0b68349e16e2a236ef3a6c264-runc.PsCh6X.mount: Deactivated successfully. Apr 13 23:49:15.033829 systemd[1]: Started cri-containerd-de2aec49366b77bdce86bcd450ae6f023d6693e0b68349e16e2a236ef3a6c264.scope - libcontainer container de2aec49366b77bdce86bcd450ae6f023d6693e0b68349e16e2a236ef3a6c264. Apr 13 23:49:15.839781 containerd[1468]: time="2026-04-13T23:49:15.838832417Z" level=info msg="StartContainer for \"06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71\" returns successfully" Apr 13 23:49:16.200962 containerd[1468]: time="2026-04-13T23:49:16.200309977Z" level=info msg="StartContainer for \"de2aec49366b77bdce86bcd450ae6f023d6693e0b68349e16e2a236ef3a6c264\" returns successfully" Apr 13 23:49:16.316757 kubelet[2322]: E0413 23:49:16.316570 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.25:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:49:16.321473 containerd[1468]: time="2026-04-13T23:49:16.321332432Z" level=info msg="StartContainer for \"e54945a1bfd0258cf7db1259dd6ee5206b39c872285f52d1f236699ec8573c38\" returns successfully" Apr 13 23:49:16.321797 kubelet[2322]: E0413 23:49:16.236890 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.25:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:49:17.605413 kubelet[2322]: I0413 23:49:17.601804 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:17.751923 kubelet[2322]: E0413 23:49:17.734412 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": dial tcp 10.0.0.25:6443: connect: connection refused" node="localhost" Apr 13 23:49:18.708582 kubelet[2322]: E0413 23:49:18.708301 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:18.711851 kubelet[2322]: E0413 23:49:18.711626 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:20.057617 kubelet[2322]: E0413 23:49:20.057490 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:20.059522 kubelet[2322]: E0413 23:49:20.057903 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:20.276287 kubelet[2322]: E0413 23:49:20.266141 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:20.339197 kubelet[2322]: E0413 23:49:20.337917 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:22.499428 kubelet[2322]: E0413 23:49:22.433428 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:22.574592 kubelet[2322]: E0413 23:49:22.571492 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:22.575871 kubelet[2322]: E0413 23:49:22.574699 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:22.585141 kubelet[2322]: E0413 23:49:22.580675 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:22.585141 kubelet[2322]: E0413 23:49:22.579798 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:22.601558 kubelet[2322]: E0413 23:49:22.601086 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:23.621192 kubelet[2322]: E0413 23:49:23.616940 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:49:23.766169 kubelet[2322]: E0413 23:49:23.753798 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:23.822557 kubelet[2322]: E0413 23:49:23.819253 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:24.797014 kubelet[2322]: E0413 23:49:24.796863 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:25.013462 kubelet[2322]: E0413 23:49:25.013304 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:25.032180 kubelet[2322]: I0413 23:49:25.030783 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:25.326689 kubelet[2322]: E0413 23:49:25.313886 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:25.403787 kubelet[2322]: E0413 23:49:25.401796 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:26.491336 kubelet[2322]: E0413 23:49:26.490899 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:26.509121 kubelet[2322]: E0413 23:49:26.508730 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:26.526781 kubelet[2322]: E0413 23:49:26.524838 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:26.535090 kubelet[2322]: E0413 23:49:26.534332 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:27.897513 kubelet[2322]: E0413 23:49:27.891809 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:27.934738 kubelet[2322]: E0413 23:49:27.932528 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:30.665122 kubelet[2322]: E0413 23:49:30.664749 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="7s" Apr 13 23:49:31.160193 kubelet[2322]: E0413 23:49:31.150255 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:49:32.934388 kubelet[2322]: E0413 23:49:32.928825 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:49:33.322509 kubelet[2322]: E0413 23:49:33.319832 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:49:33.699816 kubelet[2322]: E0413 23:49:33.690461 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:49:35.165647 kubelet[2322]: E0413 23:49:35.165257 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 13 23:49:36.407150 kubelet[2322]: E0413 23:49:36.393526 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:49:39.333606 kubelet[2322]: E0413 23:49:39.331485 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:49:39.338402 kubelet[2322]: E0413 23:49:39.337877 2322 certificate_manager.go:461] "Reached backoff limit, still unable to rotate certs" err="timed out waiting for the condition" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:49:42.359384 kubelet[2322]: I0413 23:49:42.358071 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:49:42.660929 kubelet[2322]: E0413 23:49:42.657652 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:49:42.737426 kubelet[2322]: E0413 23:49:42.730943 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:49:43.725178 kubelet[2322]: E0413 23:49:43.723558 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:49:46.818214 kubelet[2322]: E0413 23:49:46.816148 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:49:47.710136 kubelet[2322]: E0413 23:49:47.708445 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="7s" Apr 13 23:49:52.468294 kubelet[2322]: E0413 23:49:52.462299 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 13 23:49:53.789475 kubelet[2322]: E0413 23:49:53.788029 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:49:55.173365 kubelet[2322]: E0413 23:49:55.158413 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:49:56.607218 kubelet[2322]: E0413 23:49:56.601533 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:49:59.708582 kubelet[2322]: I0413 23:49:59.706588 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:50:00.354803 kubelet[2322]: E0413 23:50:00.349568 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:50:03.817134 kubelet[2322]: E0413 23:50:03.805605 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:50:04.846249 kubelet[2322]: E0413 23:50:04.845119 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="7s" Apr 13 23:50:07.820630 kubelet[2322]: E0413 23:50:07.818501 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:50:09.814272 kubelet[2322]: E0413 23:50:09.813485 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 13 23:50:14.327538 kubelet[2322]: E0413 23:50:14.084107 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:50:17.092535 kubelet[2322]: E0413 23:50:16.905655 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:50:17.258072 kubelet[2322]: I0413 23:50:17.257943 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:50:21.806524 kubelet[2322]: E0413 23:50:21.796199 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:50:21.964120 kubelet[2322]: E0413 23:50:21.958457 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="7s" Apr 13 23:50:24.425282 kubelet[2322]: E0413 23:50:24.424156 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:50:27.319353 kubelet[2322]: E0413 23:50:27.311810 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 13 23:50:34.424099 kubelet[2322]: I0413 23:50:34.415801 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:50:34.491388 kubelet[2322]: E0413 23:50:34.433774 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:50:37.303788 kubelet[2322]: E0413 23:50:37.267616 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:50:38.496720 kubelet[2322]: E0413 23:50:38.488298 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 13 23:50:39.017303 kubelet[2322]: E0413 23:50:39.012740 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="7s" Apr 13 23:50:42.362476 kubelet[2322]: E0413 23:50:42.340318 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:50:42.416364 kubelet[2322]: E0413 23:50:42.393905 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:50:44.042214 kubelet[2322]: E0413 23:50:44.039719 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.25:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 13 23:50:44.503099 kubelet[2322]: E0413 23:50:44.501409 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:50:44.535737 kubelet[2322]: E0413 23:50:44.518516 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 13 23:50:47.369401 kubelet[2322]: E0413 23:50:47.364329 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:50:47.384218 kubelet[2322]: E0413 23:50:47.381806 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:50:51.739928 kubelet[2322]: I0413 23:50:51.736151 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:50:53.374624 kubelet[2322]: E0413 23:50:53.373631 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 13 23:50:53.538638 kubelet[2322]: E0413 23:50:53.441850 2322 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 13 23:50:54.524687 kubelet[2322]: E0413 23:50:54.523677 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:50:56.093299 kubelet[2322]: E0413 23:50:56.079871 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="7s" Apr 13 23:50:57.386101 kubelet[2322]: E0413 23:50:57.375575 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.25:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:51:00.605468 kubelet[2322]: E0413 23:51:00.596374 2322 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 13 23:51:01.802829 kubelet[2322]: E0413 23:51:01.798801 2322 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.25:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 13 23:51:02.312085 kubelet[2322]: E0413 23:51:02.309642 2322 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 13 23:51:02.331116 kubelet[2322]: E0413 23:51:02.330372 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:51:04.541317 kubelet[2322]: E0413 23:51:04.538174 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:51:08.913158 kubelet[2322]: I0413 23:51:08.912591 2322 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:51:09.002550 kubelet[2322]: E0413 23:51:09.001942 2322 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Apr 13 23:51:09.426036 kubelet[2322]: I0413 23:51:09.422099 2322 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Apr 13 23:51:09.438547 kubelet[2322]: E0413 23:51:09.435606 2322 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Apr 13 23:51:09.575643 kubelet[2322]: E0413 23:51:09.574482 2322 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18a60f7ecb745e9c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,LastTimestamp:2026-04-13 23:49:00.29793654 +0000 UTC m=+17.310196685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:51:10.964113 kubelet[2322]: E0413 23:51:10.937932 2322 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18a60f7ee8093a55 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-13 23:49:00.777454165 +0000 UTC m=+17.789714275,LastTimestamp:2026-04-13 23:49:00.777454165 +0000 UTC m=+17.789714275,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 13 23:51:12.172103 kubelet[2322]: E0413 23:51:12.170571 2322 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 13 23:51:12.278959 kubelet[2322]: E0413 23:51:12.272930 2322 kubelet_node_status.go:460] "Node not becoming ready in time after startup" Apr 13 23:51:13.443428 kubelet[2322]: E0413 23:51:13.442416 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:14.623092 kubelet[2322]: E0413 23:51:14.621883 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:51:18.612306 kubelet[2322]: E0413 23:51:18.601099 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:22.404401 kubelet[2322]: E0413 23:51:22.402884 2322 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Apr 13 23:51:23.706076 kubelet[2322]: E0413 23:51:23.704409 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:24.641018 kubelet[2322]: E0413 23:51:24.636379 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:51:28.936196 kubelet[2322]: E0413 23:51:28.933959 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:33.868572 kubelet[2322]: E0413 23:51:33.868332 2322 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Apr 13 23:51:33.982636 kubelet[2322]: E0413 23:51:33.982099 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:34.663299 kubelet[2322]: E0413 23:51:34.659827 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 13 23:51:39.048143 kubelet[2322]: I0413 23:51:39.044555 2322 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Apr 13 23:51:39.122511 kubelet[2322]: E0413 23:51:39.117494 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:39.424726 kubelet[2322]: I0413 23:51:39.422084 2322 apiserver.go:52] "Watching apiserver" Apr 13 23:51:39.674138 kubelet[2322]: I0413 23:51:39.672926 2322 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 13 23:51:41.238131 kubelet[2322]: I0413 23:51:41.236497 2322 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Apr 13 23:51:41.442873 kubelet[2322]: E0413 23:51:41.442079 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:51:41.965434 kubelet[2322]: E0413 23:51:41.950167 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:51:42.012320 kubelet[2322]: I0413 23:51:42.005662 2322 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Apr 13 23:51:42.895255 kubelet[2322]: E0413 23:51:42.891459 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:51:44.278064 kubelet[2322]: E0413 23:51:44.273650 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:49.323450 kubelet[2322]: E0413 23:51:49.319550 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:53.911561 kubelet[2322]: I0413 23:51:53.910650 2322 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=12.906230142 podStartE2EDuration="12.906230142s" podCreationTimestamp="2026-04-13 23:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-13 23:51:52.828849707 +0000 UTC m=+189.841109820" watchObservedRunningTime="2026-04-13 23:51:53.906230142 +0000 UTC m=+190.918490252" Apr 13 23:51:54.417887 kubelet[2322]: E0413 23:51:54.413576 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:51:55.410073 kubelet[2322]: I0413 23:51:55.409720 2322 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=13.409410624 podStartE2EDuration="13.409410624s" podCreationTimestamp="2026-04-13 23:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-13 23:51:54.035769662 +0000 UTC m=+191.048029792" watchObservedRunningTime="2026-04-13 23:51:55.409410624 +0000 UTC m=+192.421670725" Apr 13 23:51:55.414769 kubelet[2322]: I0413 23:51:55.413648 2322 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=15.413401931 podStartE2EDuration="15.413401931s" podCreationTimestamp="2026-04-13 23:51:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-13 23:51:55.390116885 +0000 UTC m=+192.402377003" watchObservedRunningTime="2026-04-13 23:51:55.413401931 +0000 UTC m=+192.425662036" Apr 13 23:51:59.614483 kubelet[2322]: E0413 23:51:59.613664 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:04.770956 kubelet[2322]: E0413 23:52:04.768784 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:09.945643 kubelet[2322]: E0413 23:52:09.941223 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:15.067546 kubelet[2322]: E0413 23:52:15.066918 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:20.155932 kubelet[2322]: E0413 23:52:20.154802 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:25.236943 kubelet[2322]: E0413 23:52:25.232443 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:30.415178 kubelet[2322]: E0413 23:52:30.413765 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:35.578102 kubelet[2322]: E0413 23:52:35.522805 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:40.749880 kubelet[2322]: E0413 23:52:40.740844 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:42.486827 systemd[1]: cri-containerd-06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71.scope: Deactivated successfully. Apr 13 23:52:42.519331 systemd[1]: cri-containerd-06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71.scope: Consumed 13.355s CPU time. Apr 13 23:52:43.420409 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71-rootfs.mount: Deactivated successfully. Apr 13 23:52:43.703248 containerd[1468]: time="2026-04-13T23:52:43.701454616Z" level=info msg="shim disconnected" id=06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71 namespace=k8s.io Apr 13 23:52:43.739616 containerd[1468]: time="2026-04-13T23:52:43.739102334Z" level=warning msg="cleaning up after shim disconnected" id=06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71 namespace=k8s.io Apr 13 23:52:43.739616 containerd[1468]: time="2026-04-13T23:52:43.739548904Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 13 23:52:44.192207 containerd[1468]: time="2026-04-13T23:52:44.191480695Z" level=warning msg="cleanup warnings time=\"2026-04-13T23:52:44Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 13 23:52:45.467025 kubelet[2322]: I0413 23:52:45.466490 2322 scope.go:117] "RemoveContainer" containerID="06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71" Apr 13 23:52:45.474078 kubelet[2322]: E0413 23:52:45.471549 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:52:45.911706 kubelet[2322]: E0413 23:52:45.908708 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:45.951093 containerd[1468]: time="2026-04-13T23:52:45.947111115Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 13 23:52:46.417210 containerd[1468]: time="2026-04-13T23:52:46.416387607Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b\"" Apr 13 23:52:46.509311 containerd[1468]: time="2026-04-13T23:52:46.505564197Z" level=info msg="StartContainer for \"a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b\"" Apr 13 23:52:47.409908 systemd[1]: run-containerd-runc-k8s.io-a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b-runc.QR9Vma.mount: Deactivated successfully. Apr 13 23:52:47.499412 systemd[1]: Started cri-containerd-a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b.scope - libcontainer container a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b. Apr 13 23:52:48.409183 containerd[1468]: time="2026-04-13T23:52:48.408932606Z" level=info msg="StartContainer for \"a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b\" returns successfully" Apr 13 23:52:50.172731 kubelet[2322]: E0413 23:52:50.171837 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:52:50.987848 kubelet[2322]: E0413 23:52:50.986661 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:54.090542 kubelet[2322]: E0413 23:52:54.090024 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:52:56.051449 kubelet[2322]: E0413 23:52:56.051131 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:52:56.295345 kubelet[2322]: E0413 23:52:56.290115 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:01.171624 kubelet[2322]: E0413 23:53:01.170704 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:06.250069 kubelet[2322]: E0413 23:53:06.245360 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:06.951847 kubelet[2322]: E0413 23:53:06.946278 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:08.125164 kubelet[2322]: E0413 23:53:08.122173 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:11.522905 kubelet[2322]: E0413 23:53:11.514653 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:16.676605 kubelet[2322]: E0413 23:53:16.668613 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:17.431934 kubelet[2322]: E0413 23:53:17.431377 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:21.784045 kubelet[2322]: E0413 23:53:21.776468 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:26.920088 kubelet[2322]: E0413 23:53:26.913201 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:28.441799 systemd[1]: cri-containerd-a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b.scope: Deactivated successfully. Apr 13 23:53:28.443338 systemd[1]: cri-containerd-a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b.scope: Consumed 8.999s CPU time. Apr 13 23:53:29.883098 containerd[1468]: time="2026-04-13T23:53:29.879238336Z" level=info msg="shim disconnected" id=a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b namespace=k8s.io Apr 13 23:53:29.892060 containerd[1468]: time="2026-04-13T23:53:29.889005294Z" level=warning msg="cleaning up after shim disconnected" id=a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b namespace=k8s.io Apr 13 23:53:29.892060 containerd[1468]: time="2026-04-13T23:53:29.890147734Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 13 23:53:29.901263 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b-rootfs.mount: Deactivated successfully. Apr 13 23:53:31.627853 kubelet[2322]: I0413 23:53:31.627579 2322 scope.go:117] "RemoveContainer" containerID="06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71" Apr 13 23:53:31.752338 kubelet[2322]: I0413 23:53:31.749878 2322 scope.go:117] "RemoveContainer" containerID="a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b" Apr 13 23:53:31.759942 kubelet[2322]: E0413 23:53:31.757199 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:31.790213 kubelet[2322]: E0413 23:53:31.786848 2322 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-controller-manager pod=kube-controller-manager-localhost_kube-system(ebf8e820819e4b80bc03d078b9ba80f5)\"" pod="kube-system/kube-controller-manager-localhost" podUID="ebf8e820819e4b80bc03d078b9ba80f5" Apr 13 23:53:32.005928 kubelet[2322]: E0413 23:53:31.996756 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:32.168487 containerd[1468]: time="2026-04-13T23:53:32.166317637Z" level=info msg="RemoveContainer for \"06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71\"" Apr 13 23:53:32.409603 containerd[1468]: time="2026-04-13T23:53:32.389069919Z" level=info msg="RemoveContainer for \"06db1f540094b75fb7cf9424524f779b56d0732fd10e53005a4fc280457ccb71\" returns successfully" Apr 13 23:53:37.270102 kubelet[2322]: E0413 23:53:37.266511 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:40.925404 kubelet[2322]: I0413 23:53:40.924418 2322 scope.go:117] "RemoveContainer" containerID="a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b" Apr 13 23:53:40.936919 kubelet[2322]: E0413 23:53:40.936022 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:41.290069 containerd[1468]: time="2026-04-13T23:53:41.288954362Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:2,}" Apr 13 23:53:41.883866 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3080316403.mount: Deactivated successfully. Apr 13 23:53:42.018055 containerd[1468]: time="2026-04-13T23:53:42.014173676Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:2,} returns container id \"460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2\"" Apr 13 23:53:42.054180 containerd[1468]: time="2026-04-13T23:53:42.045591937Z" level=info msg="StartContainer for \"460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2\"" Apr 13 23:53:42.661136 kubelet[2322]: E0413 23:53:42.655964 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:43.843367 systemd[1]: Started cri-containerd-460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2.scope - libcontainer container 460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2. Apr 13 23:53:45.430451 containerd[1468]: time="2026-04-13T23:53:45.430154040Z" level=info msg="StartContainer for \"460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2\" returns successfully" Apr 13 23:53:45.838392 kubelet[2322]: E0413 23:53:45.837634 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:47.160332 kubelet[2322]: E0413 23:53:47.158793 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:48.045076 kubelet[2322]: E0413 23:53:48.043349 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:49.964670 kubelet[2322]: E0413 23:53:49.955908 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:53.770188 kubelet[2322]: E0413 23:53:53.678849 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:54.482163 kubelet[2322]: E0413 23:53:54.479276 2322 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.335s" Apr 13 23:53:56.342494 kubelet[2322]: E0413 23:53:56.339248 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:53:58.990461 kubelet[2322]: E0413 23:53:58.988694 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:53:59.695123 kubelet[2322]: E0413 23:53:59.692332 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:54:04.113857 kubelet[2322]: E0413 23:54:04.113422 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:54:09.292110 kubelet[2322]: E0413 23:54:09.288589 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:54:11.695810 systemd[1]: Reloading requested from client PID 2764 ('systemctl') (unit session-7.scope)... Apr 13 23:54:11.696433 systemd[1]: Reloading... Apr 13 23:54:14.414175 kubelet[2322]: E0413 23:54:14.411919 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:54:16.827053 zram_generator::config[2800]: No configuration found. Apr 13 23:54:19.609781 kubelet[2322]: E0413 23:54:19.609457 2322 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:54:19.798319 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 13 23:54:22.077630 systemd[1]: Reloading finished in 10380 ms. Apr 13 23:54:23.597928 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:54:23.674051 systemd[1]: kubelet.service: Deactivated successfully. Apr 13 23:54:23.677773 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:54:23.680610 systemd[1]: kubelet.service: Consumed 3min 1.670s CPU time, 147.6M memory peak, 0B memory swap peak. Apr 13 23:54:23.816962 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 13 23:54:27.018687 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 13 23:54:27.169537 (kubelet)[2848]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 13 23:54:37.701172 update_engine[1462]: I20260413 23:54:37.700340 1462 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Apr 13 23:54:37.728801 update_engine[1462]: I20260413 23:54:37.707333 1462 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Apr 13 23:54:37.728801 update_engine[1462]: I20260413 23:54:37.720851 1462 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.776883 1462 omaha_request_params.cc:62] Current group set to lts Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.781720 1462 update_attempter.cc:499] Already updated boot flags. Skipping. Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.781850 1462 update_attempter.cc:643] Scheduling an action processor start. Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.782037 1462 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.782967 1462 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.783357 1462 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.783371 1462 omaha_request_action.cc:272] Request: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: Apr 13 23:54:37.788360 update_engine[1462]: I20260413 23:54:37.783396 1462 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 13 23:54:37.802925 locksmithd[1500]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Apr 13 23:54:37.811579 update_engine[1462]: I20260413 23:54:37.811304 1462 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 13 23:54:37.819158 update_engine[1462]: I20260413 23:54:37.818677 1462 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 13 23:54:37.838732 update_engine[1462]: E20260413 23:54:37.837806 1462 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 13 23:54:37.872701 update_engine[1462]: I20260413 23:54:37.869462 1462 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Apr 13 23:54:46.484155 kubelet[2848]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 13 23:54:46.484155 kubelet[2848]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 13 23:54:46.484155 kubelet[2848]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 13 23:54:46.512795 kubelet[2848]: I0413 23:54:46.487907 2848 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 13 23:54:47.698788 update_engine[1462]: I20260413 23:54:47.695582 1462 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 13 23:54:47.700757 update_engine[1462]: I20260413 23:54:47.700562 1462 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 13 23:54:47.702455 update_engine[1462]: I20260413 23:54:47.702088 1462 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 13 23:54:47.720440 update_engine[1462]: E20260413 23:54:47.718096 1462 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 13 23:54:47.727081 update_engine[1462]: I20260413 23:54:47.725862 1462 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Apr 13 23:54:48.355222 kubelet[2848]: I0413 23:54:48.340704 2848 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 13 23:54:48.361554 kubelet[2848]: I0413 23:54:48.359183 2848 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 13 23:54:48.486991 kubelet[2848]: I0413 23:54:48.484515 2848 server.go:956] "Client rotation is on, will bootstrap in background" Apr 13 23:54:49.028940 kubelet[2848]: I0413 23:54:49.023535 2848 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 13 23:54:50.862592 kubelet[2848]: I0413 23:54:50.859822 2848 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 13 23:54:53.301655 kubelet[2848]: E0413 23:54:53.294149 2848 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 13 23:54:53.314644 kubelet[2848]: I0413 23:54:53.307443 2848 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 13 23:54:54.534378 kubelet[2848]: I0413 23:54:54.534000 2848 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 13 23:54:54.567593 kubelet[2848]: I0413 23:54:54.560196 2848 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 13 23:54:54.611425 kubelet[2848]: I0413 23:54:54.565642 2848 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 13 23:54:54.640391 kubelet[2848]: I0413 23:54:54.614938 2848 topology_manager.go:138] "Creating topology manager with none policy" Apr 13 23:54:54.640391 kubelet[2848]: I0413 23:54:54.619903 2848 container_manager_linux.go:303] "Creating device plugin manager" Apr 13 23:54:54.712635 kubelet[2848]: I0413 23:54:54.710443 2848 state_mem.go:36] "Initialized new in-memory state store" Apr 13 23:54:54.744674 kubelet[2848]: I0413 23:54:54.741943 2848 kubelet.go:480] "Attempting to sync node with API server" Apr 13 23:54:54.755297 kubelet[2848]: I0413 23:54:54.750859 2848 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 13 23:54:54.776139 kubelet[2848]: I0413 23:54:54.775907 2848 kubelet.go:386] "Adding apiserver pod source" Apr 13 23:54:54.777187 kubelet[2848]: I0413 23:54:54.776392 2848 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 13 23:54:55.264477 kubelet[2848]: I0413 23:54:55.252935 2848 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 13 23:54:55.503152 kubelet[2848]: I0413 23:54:55.498666 2848 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 13 23:54:56.328223 kubelet[2848]: I0413 23:54:56.322904 2848 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 13 23:54:56.360662 kubelet[2848]: I0413 23:54:56.360475 2848 server.go:1289] "Started kubelet" Apr 13 23:54:56.512250 kubelet[2848]: I0413 23:54:56.439877 2848 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 13 23:54:56.560742 kubelet[2848]: I0413 23:54:56.522278 2848 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 13 23:54:56.683551 kubelet[2848]: I0413 23:54:56.682590 2848 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 13 23:54:56.973966 kubelet[2848]: I0413 23:54:56.972809 2848 apiserver.go:52] "Watching apiserver" Apr 13 23:54:57.164049 kubelet[2848]: I0413 23:54:57.158335 2848 server.go:317] "Adding debug handlers to kubelet server" Apr 13 23:54:57.696136 update_engine[1462]: I20260413 23:54:57.693965 1462 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 13 23:54:57.708081 update_engine[1462]: I20260413 23:54:57.700852 1462 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 13 23:54:57.712123 update_engine[1462]: I20260413 23:54:57.708079 1462 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 13 23:54:57.720793 update_engine[1462]: E20260413 23:54:57.720381 1462 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 13 23:54:57.720793 update_engine[1462]: I20260413 23:54:57.720712 1462 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Apr 13 23:54:57.971599 kubelet[2848]: E0413 23:54:57.955677 2848 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 13 23:54:58.396606 kubelet[2848]: I0413 23:54:58.389658 2848 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 13 23:54:58.434544 kubelet[2848]: I0413 23:54:58.431271 2848 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 13 23:54:58.520145 kubelet[2848]: I0413 23:54:58.519070 2848 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 13 23:54:58.762962 kubelet[2848]: I0413 23:54:58.757828 2848 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 13 23:54:59.268662 kubelet[2848]: I0413 23:54:59.265727 2848 reconciler.go:26] "Reconciler: start to sync state" Apr 13 23:54:59.440246 kubelet[2848]: I0413 23:54:59.436956 2848 factory.go:223] Registration of the systemd container factory successfully Apr 13 23:54:59.599685 kubelet[2848]: I0413 23:54:59.538689 2848 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 13 23:54:59.749574 kubelet[2848]: I0413 23:54:59.742379 2848 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 13 23:54:59.885528 kubelet[2848]: I0413 23:54:59.882904 2848 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 13 23:54:59.889107 kubelet[2848]: I0413 23:54:59.888314 2848 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 13 23:54:59.894020 kubelet[2848]: I0413 23:54:59.893406 2848 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 13 23:54:59.894020 kubelet[2848]: I0413 23:54:59.893889 2848 kubelet.go:2436] "Starting kubelet main sync loop" Apr 13 23:54:59.939699 kubelet[2848]: E0413 23:54:59.902522 2848 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 13 23:55:00.295755 kubelet[2848]: E0413 23:55:00.221483 2848 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 13 23:55:00.505134 kubelet[2848]: E0413 23:55:00.498478 2848 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:55:00.518858 kubelet[2848]: I0413 23:55:00.518560 2848 factory.go:223] Registration of the containerd container factory successfully Apr 13 23:55:00.916028 kubelet[2848]: E0413 23:55:00.909726 2848 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:55:01.781190 kubelet[2848]: E0413 23:55:01.770868 2848 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:55:03.439867 kubelet[2848]: E0413 23:55:03.412900 2848 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:55:06.653305 kubelet[2848]: E0413 23:55:06.649878 2848 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 13 23:55:07.301181 kubelet[2848]: I0413 23:55:07.298653 2848 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 13 23:55:07.303523 kubelet[2848]: I0413 23:55:07.302868 2848 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 13 23:55:07.310681 kubelet[2848]: I0413 23:55:07.310284 2848 state_mem.go:36] "Initialized new in-memory state store" Apr 13 23:55:07.420142 kubelet[2848]: I0413 23:55:07.419713 2848 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 13 23:55:07.442143 kubelet[2848]: I0413 23:55:07.425348 2848 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 13 23:55:07.480169 kubelet[2848]: I0413 23:55:07.473678 2848 policy_none.go:49] "None policy: Start" Apr 13 23:55:07.497147 kubelet[2848]: I0413 23:55:07.489622 2848 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 13 23:55:07.506075 kubelet[2848]: I0413 23:55:07.504188 2848 state_mem.go:35] "Initializing new in-memory state store" Apr 13 23:55:07.698324 update_engine[1462]: I20260413 23:55:07.697067 1462 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 13 23:55:07.731760 kubelet[2848]: I0413 23:55:07.725734 2848 state_mem.go:75] "Updated machine memory state" Apr 13 23:55:07.792522 update_engine[1462]: I20260413 23:55:07.792222 1462 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 13 23:55:07.797120 update_engine[1462]: I20260413 23:55:07.794941 1462 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 13 23:55:07.817019 update_engine[1462]: E20260413 23:55:07.816654 1462 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 13 23:55:07.826299 update_engine[1462]: I20260413 23:55:07.823105 1462 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 13 23:55:07.943230 update_engine[1462]: I20260413 23:55:07.842548 1462 omaha_request_action.cc:617] Omaha request response: Apr 13 23:55:07.949563 update_engine[1462]: E20260413 23:55:07.945951 1462 omaha_request_action.cc:636] Omaha request network transfer failed. Apr 13 23:55:07.955113 update_engine[1462]: I20260413 23:55:07.952349 1462 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Apr 13 23:55:07.956746 update_engine[1462]: I20260413 23:55:07.956606 1462 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 13 23:55:07.958128 update_engine[1462]: I20260413 23:55:07.958057 1462 update_attempter.cc:306] Processing Done. Apr 13 23:55:07.963912 update_engine[1462]: E20260413 23:55:07.963714 1462 update_attempter.cc:619] Update failed. Apr 13 23:55:07.965376 update_engine[1462]: I20260413 23:55:07.965157 1462 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Apr 13 23:55:07.966566 update_engine[1462]: I20260413 23:55:07.966208 1462 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Apr 13 23:55:07.971134 update_engine[1462]: I20260413 23:55:07.970821 1462 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Apr 13 23:55:07.977168 update_engine[1462]: I20260413 23:55:07.972939 1462 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 13 23:55:07.981638 update_engine[1462]: I20260413 23:55:07.978587 1462 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 13 23:55:07.981638 update_engine[1462]: I20260413 23:55:07.978636 1462 omaha_request_action.cc:272] Request: Apr 13 23:55:07.981638 update_engine[1462]: Apr 13 23:55:07.981638 update_engine[1462]: Apr 13 23:55:07.981638 update_engine[1462]: Apr 13 23:55:07.981638 update_engine[1462]: Apr 13 23:55:07.981638 update_engine[1462]: Apr 13 23:55:07.981638 update_engine[1462]: Apr 13 23:55:07.981638 update_engine[1462]: I20260413 23:55:07.978644 1462 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 13 23:55:07.994465 update_engine[1462]: I20260413 23:55:07.989456 1462 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 13 23:55:08.001217 update_engine[1462]: I20260413 23:55:07.997948 1462 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 13 23:55:08.029370 update_engine[1462]: E20260413 23:55:08.027763 1462 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.028891 1462 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.028959 1462 omaha_request_action.cc:617] Omaha request response: Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.029014 1462 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.029021 1462 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.029026 1462 update_attempter.cc:306] Processing Done. Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.029054 1462 update_attempter.cc:310] Error event sent. Apr 13 23:55:08.029370 update_engine[1462]: I20260413 23:55:08.029131 1462 update_check_scheduler.cc:74] Next update check in 48m28s Apr 13 23:55:08.127460 locksmithd[1500]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Apr 13 23:55:08.127460 locksmithd[1500]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Apr 13 23:55:09.804410 kubelet[2848]: E0413 23:55:09.795990 2848 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 13 23:55:09.868231 kubelet[2848]: I0413 23:55:09.855841 2848 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 13 23:55:09.900209 kubelet[2848]: I0413 23:55:09.888821 2848 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 13 23:55:09.998217 kubelet[2848]: I0413 23:55:09.988175 2848 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 13 23:55:10.497125 kubelet[2848]: E0413 23:55:10.493293 2848 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 13 23:55:11.895086 kubelet[2848]: I0413 23:55:11.893956 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3ffefc5deebb452093137d8633d90466-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3ffefc5deebb452093137d8633d90466\") " pod="kube-system/kube-apiserver-localhost" Apr 13 23:55:12.051136 kubelet[2848]: I0413 23:55:12.044820 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3ffefc5deebb452093137d8633d90466-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3ffefc5deebb452093137d8633d90466\") " pod="kube-system/kube-apiserver-localhost" Apr 13 23:55:12.103216 kubelet[2848]: I0413 23:55:12.098729 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3ffefc5deebb452093137d8633d90466-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3ffefc5deebb452093137d8633d90466\") " pod="kube-system/kube-apiserver-localhost" Apr 13 23:55:12.228857 kubelet[2848]: I0413 23:55:12.222426 2848 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Apr 13 23:55:12.585482 kubelet[2848]: I0413 23:55:12.485959 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:12.622196 kubelet[2848]: I0413 23:55:12.617604 2848 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:12.741904 kubelet[2848]: I0413 23:55:12.741661 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:13.398517 kubelet[2848]: I0413 23:55:13.383246 2848 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 13 23:55:13.437548 kubelet[2848]: I0413 23:55:13.338910 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/39798d73a6894e44ae801eb773bf9a39-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"39798d73a6894e44ae801eb773bf9a39\") " pod="kube-system/kube-scheduler-localhost" Apr 13 23:55:13.497168 kubelet[2848]: I0413 23:55:13.495957 2848 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Apr 13 23:55:13.538135 kubelet[2848]: I0413 23:55:13.512608 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:13.555598 kubelet[2848]: I0413 23:55:13.554468 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:13.732686 kubelet[2848]: I0413 23:55:13.696672 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ebf8e820819e4b80bc03d078b9ba80f5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"ebf8e820819e4b80bc03d078b9ba80f5\") " pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:13.954758 kubelet[2848]: I0413 23:55:13.954412 2848 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 13 23:55:15.480351 kubelet[2848]: E0413 23:55:15.478304 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.517s" Apr 13 23:55:16.675325 kubelet[2848]: E0413 23:55:16.661781 2848 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Apr 13 23:55:16.941202 kubelet[2848]: E0413 23:55:16.940714 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:17.032071 kubelet[2848]: E0413 23:55:17.031770 2848 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Apr 13 23:55:17.397535 kubelet[2848]: E0413 23:55:17.392719 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.478s" Apr 13 23:55:17.397535 kubelet[2848]: E0413 23:55:17.393765 2848 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Apr 13 23:55:17.413900 kubelet[2848]: E0413 23:55:17.404754 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:17.421169 kubelet[2848]: E0413 23:55:17.415656 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:18.406217 kubelet[2848]: E0413 23:55:18.400842 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:18.662781 kubelet[2848]: E0413 23:55:18.653293 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:18.914819 kubelet[2848]: E0413 23:55:18.913600 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:19.069356 kubelet[2848]: I0413 23:55:19.067784 2848 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Apr 13 23:55:19.104132 kubelet[2848]: I0413 23:55:19.090962 2848 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Apr 13 23:55:19.482340 kubelet[2848]: E0413 23:55:19.480658 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:19.499713 kubelet[2848]: E0413 23:55:19.498032 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:20.968108 kubelet[2848]: E0413 23:55:20.943604 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.022s" Apr 13 23:55:25.394142 kubelet[2848]: E0413 23:55:25.388814 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.406s" Apr 13 23:55:28.009084 kubelet[2848]: E0413 23:55:27.996207 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.949s" Apr 13 23:55:30.079209 kubelet[2848]: E0413 23:55:30.076074 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.949s" Apr 13 23:55:31.387951 kubelet[2848]: E0413 23:55:31.385791 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:31.399188 kubelet[2848]: E0413 23:55:31.392400 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:31.543140 kubelet[2848]: E0413 23:55:31.538302 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:33.028965 kubelet[2848]: E0413 23:55:33.024847 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.716s" Apr 13 23:55:35.214364 kubelet[2848]: E0413 23:55:35.137346 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.993s" Apr 13 23:55:38.935314 kubelet[2848]: E0413 23:55:38.753722 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.491s" Apr 13 23:55:39.959740 kubelet[2848]: E0413 23:55:39.959540 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:55:41.025098 kubelet[2848]: E0413 23:55:41.022055 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.924s" Apr 13 23:55:43.101118 kubelet[2848]: E0413 23:55:42.967641 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.6s" Apr 13 23:55:45.737159 kubelet[2848]: E0413 23:55:45.736354 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.608s" Apr 13 23:55:47.837254 kubelet[2848]: E0413 23:55:47.824494 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.815s" Apr 13 23:55:51.080109 kubelet[2848]: E0413 23:55:51.076617 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.053s" Apr 13 23:55:54.795848 kubelet[2848]: E0413 23:55:54.794851 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.86s" Apr 13 23:55:56.514838 kubelet[2848]: E0413 23:55:56.513858 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.716s" Apr 13 23:55:57.797937 systemd[1]: cri-containerd-460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2.scope: Deactivated successfully. Apr 13 23:55:57.801214 systemd[1]: cri-containerd-460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2.scope: Consumed 43.112s CPU time, 18.2M memory peak, 0B memory swap peak. Apr 13 23:55:59.702406 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2-rootfs.mount: Deactivated successfully. Apr 13 23:55:59.798893 containerd[1468]: time="2026-04-13T23:55:59.790141778Z" level=info msg="shim disconnected" id=460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2 namespace=k8s.io Apr 13 23:55:59.798893 containerd[1468]: time="2026-04-13T23:55:59.794321393Z" level=warning msg="cleaning up after shim disconnected" id=460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2 namespace=k8s.io Apr 13 23:55:59.798893 containerd[1468]: time="2026-04-13T23:55:59.796339728Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 13 23:55:59.800515 kubelet[2848]: E0413 23:55:59.791730 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.967s" Apr 13 23:56:00.169255 containerd[1468]: time="2026-04-13T23:56:00.167549985Z" level=warning msg="cleanup warnings time=\"2026-04-13T23:56:00Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 13 23:56:03.325153 kubelet[2848]: E0413 23:56:03.319565 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.386s" Apr 13 23:56:03.892359 kubelet[2848]: I0413 23:56:03.889891 2848 scope.go:117] "RemoveContainer" containerID="a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b" Apr 13 23:56:03.997903 kubelet[2848]: I0413 23:56:03.997534 2848 scope.go:117] "RemoveContainer" containerID="460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2" Apr 13 23:56:04.068067 kubelet[2848]: E0413 23:56:04.067494 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:56:04.920806 containerd[1468]: time="2026-04-13T23:56:04.920576295Z" level=info msg="RemoveContainer for \"a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b\"" Apr 13 23:56:05.416206 containerd[1468]: time="2026-04-13T23:56:05.411813375Z" level=info msg="RemoveContainer for \"a359485c36667c500f6c92ffc12e523838a1127c10137e5ffa32ec5aa0c4163b\" returns successfully" Apr 13 23:56:05.836585 containerd[1468]: time="2026-04-13T23:56:05.822942824Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:3,}" Apr 13 23:56:05.927265 kubelet[2848]: E0413 23:56:05.923912 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.013s" Apr 13 23:56:06.506035 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3009711066.mount: Deactivated successfully. Apr 13 23:56:06.635909 containerd[1468]: time="2026-04-13T23:56:06.632114362Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:3,} returns container id \"5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2\"" Apr 13 23:56:07.014389 containerd[1468]: time="2026-04-13T23:56:07.003343007Z" level=info msg="StartContainer for \"5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2\"" Apr 13 23:56:07.167944 kubelet[2848]: E0413 23:56:07.131933 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.183s" Apr 13 23:56:08.694078 systemd[1]: Started cri-containerd-5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2.scope - libcontainer container 5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2. Apr 13 23:56:11.908195 containerd[1468]: time="2026-04-13T23:56:11.894592218Z" level=info msg="StartContainer for \"5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2\" returns successfully" Apr 13 23:56:12.094950 kubelet[2848]: E0413 23:56:11.987809 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.053s" Apr 13 23:56:21.142363 kubelet[2848]: E0413 23:56:21.141522 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="9.037s" Apr 13 23:56:23.251102 kubelet[2848]: E0413 23:56:23.246347 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.93s" Apr 13 23:56:24.456116 kubelet[2848]: E0413 23:56:24.433514 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:56:25.906860 kubelet[2848]: E0413 23:56:25.900744 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:56:27.843227 kubelet[2848]: E0413 23:56:27.778961 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:56:29.096465 kubelet[2848]: E0413 23:56:29.096127 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.06s" Apr 13 23:56:31.091266 kubelet[2848]: E0413 23:56:31.084523 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:56:35.575157 kubelet[2848]: E0413 23:56:35.574513 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.571s" Apr 13 23:56:42.162341 kubelet[2848]: E0413 23:56:42.134484 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.066s" Apr 13 23:56:43.329928 kubelet[2848]: E0413 23:56:43.311599 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.076s" Apr 13 23:56:45.801104 kubelet[2848]: E0413 23:56:45.800351 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.852s" Apr 13 23:56:49.991725 kubelet[2848]: E0413 23:56:49.899300 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.952s" Apr 13 23:56:51.570693 kubelet[2848]: E0413 23:56:51.552535 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.471s" Apr 13 23:56:53.617397 kubelet[2848]: E0413 23:56:53.611590 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:56:55.465164 kubelet[2848]: E0413 23:56:55.458723 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.447s" Apr 13 23:56:58.058150 kubelet[2848]: E0413 23:56:58.040870 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.46s" Apr 13 23:57:00.777101 kubelet[2848]: E0413 23:57:00.774217 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.684s" Apr 13 23:57:00.870170 kubelet[2848]: E0413 23:57:00.868761 2848 kubelet_node_status.go:460] "Node not becoming ready in time after startup" Apr 13 23:57:02.901237 kubelet[2848]: E0413 23:57:02.843339 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.94s" Apr 13 23:57:02.971876 kubelet[2848]: E0413 23:57:02.955172 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:03.618669 kubelet[2848]: E0413 23:57:03.618212 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:57:04.712282 kubelet[2848]: E0413 23:57:04.711950 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.776s" Apr 13 23:57:08.308928 kubelet[2848]: E0413 23:57:08.304677 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:10.238302 kubelet[2848]: E0413 23:57:10.228234 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.321s" Apr 13 23:57:13.116753 kubelet[2848]: E0413 23:57:13.113762 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.17s" Apr 13 23:57:13.386910 kubelet[2848]: E0413 23:57:13.384280 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:18.481091 kubelet[2848]: E0413 23:57:18.477357 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:23.547739 kubelet[2848]: E0413 23:57:23.547427 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:28.618127 kubelet[2848]: E0413 23:57:28.610207 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:31.133237 kubelet[2848]: I0413 23:57:31.120223 2848 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 13 23:57:31.319166 containerd[1468]: time="2026-04-13T23:57:31.317950005Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 13 23:57:31.554476 kubelet[2848]: I0413 23:57:31.554170 2848 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 13 23:57:37.589146 kubelet[2848]: E0413 23:57:37.577618 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:38.546055 kubelet[2848]: E0413 23:57:38.545504 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="8.635s" Apr 13 23:57:40.604441 systemd[1]: cri-containerd-5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2.scope: Deactivated successfully. Apr 13 23:57:40.611556 systemd[1]: cri-containerd-5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2.scope: Consumed 27.447s CPU time. Apr 13 23:57:41.827067 kubelet[2848]: E0413 23:57:41.806900 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.225s" Apr 13 23:57:44.108163 kubelet[2848]: E0413 23:57:43.936037 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:45.765629 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2-rootfs.mount: Deactivated successfully. Apr 13 23:57:46.060596 containerd[1468]: time="2026-04-13T23:57:45.945294051Z" level=info msg="shim disconnected" id=5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2 namespace=k8s.io Apr 13 23:57:46.091221 containerd[1468]: time="2026-04-13T23:57:46.071618251Z" level=warning msg="cleaning up after shim disconnected" id=5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2 namespace=k8s.io Apr 13 23:57:46.105290 containerd[1468]: time="2026-04-13T23:57:46.082469069Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 13 23:57:46.926207 kubelet[2848]: E0413 23:57:46.904837 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.914s" Apr 13 23:57:47.110352 kubelet[2848]: E0413 23:57:47.023084 2848 kubelet_node_status.go:548] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-04-13T23:57:33Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-13T23:57:33Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-13T23:57:33Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-13T23:57:33Z\\\",\\\"type\\\":\\\"Ready\\\"}]}}\" for node \"localhost\": Patch \"https://10.0.0.25:6443/api/v1/nodes/localhost/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 13 23:57:47.570573 containerd[1468]: time="2026-04-13T23:57:47.545885937Z" level=warning msg="cleanup warnings time=\"2026-04-13T23:57:47Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 13 23:57:49.959006 kubelet[2848]: E0413 23:57:49.833824 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:51.138586 kubelet[2848]: E0413 23:57:51.129025 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.212s" Apr 13 23:57:52.185733 kubelet[2848]: E0413 23:57:52.185335 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.009s" Apr 13 23:57:52.195075 kubelet[2848]: I0413 23:57:52.193559 2848 scope.go:117] "RemoveContainer" containerID="460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2" Apr 13 23:57:52.294035 kubelet[2848]: I0413 23:57:52.293528 2848 scope.go:117] "RemoveContainer" containerID="5d3efca04524fdc618c7fa73bbe3c8b38b902dedc751da011437bd09c5e531a2" Apr 13 23:57:52.375071 kubelet[2848]: E0413 23:57:52.366750 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:57:52.681187 containerd[1468]: time="2026-04-13T23:57:52.680704554Z" level=info msg="RemoveContainer for \"460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2\"" Apr 13 23:57:52.963107 containerd[1468]: time="2026-04-13T23:57:52.962451384Z" level=info msg="RemoveContainer for \"460a9a00de4e0f94e4386c64e8e44db3435291defec5dd04fcc4f813f92944d2\" returns successfully" Apr 13 23:57:53.713112 kubelet[2848]: E0413 23:57:53.688915 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.494s" Apr 13 23:57:54.996123 containerd[1468]: time="2026-04-13T23:57:54.941049651Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:4,}" Apr 13 23:57:55.505928 kubelet[2848]: E0413 23:57:55.502776 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:57:55.920901 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1586588097.mount: Deactivated successfully. Apr 13 23:57:56.317428 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2579949571.mount: Deactivated successfully. Apr 13 23:57:56.500224 kubelet[2848]: E0413 23:57:56.463447 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.418s" Apr 13 23:57:56.630454 containerd[1468]: time="2026-04-13T23:57:56.597241885Z" level=info msg="CreateContainer within sandbox \"14b5020c0bef5cacaca1876bfb6a97013bb6fd7aaaa90df6ba9cd63e42543210\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:4,} returns container id \"2d8e7d0e2448a14abad068803d8aa43de085abb634f44a70c1225d38548f725e\"" Apr 13 23:57:57.221593 containerd[1468]: time="2026-04-13T23:57:57.214930920Z" level=info msg="StartContainer for \"2d8e7d0e2448a14abad068803d8aa43de085abb634f44a70c1225d38548f725e\"" Apr 13 23:57:58.831080 systemd[1]: Started cri-containerd-2d8e7d0e2448a14abad068803d8aa43de085abb634f44a70c1225d38548f725e.scope - libcontainer container 2d8e7d0e2448a14abad068803d8aa43de085abb634f44a70c1225d38548f725e. Apr 13 23:57:59.179367 kubelet[2848]: E0413 23:57:59.172955 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.371s" Apr 13 23:58:00.696862 containerd[1468]: time="2026-04-13T23:58:00.696178733Z" level=info msg="StartContainer for \"2d8e7d0e2448a14abad068803d8aa43de085abb634f44a70c1225d38548f725e\" returns successfully" Apr 13 23:58:03.487210 kubelet[2848]: E0413 23:58:03.475384 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.055s" Apr 13 23:58:05.289899 kubelet[2848]: E0413 23:58:05.287920 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:58:08.682165 kubelet[2848]: E0413 23:58:08.679536 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="5.098s" Apr 13 23:58:10.995225 kubelet[2848]: E0413 23:58:10.994013 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:58:11.814855 kubelet[2848]: E0413 23:58:11.813859 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 13 23:58:12.568754 kubelet[2848]: E0413 23:58:12.526558 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.726s" Apr 13 23:58:15.764259 sudo[1646]: pam_unix(sudo:session): session closed for user root Apr 13 23:58:15.786472 sshd[1643]: pam_unix(sshd:session): session closed for user core Apr 13 23:58:15.938913 systemd[1]: sshd@6-10.0.0.25:22-10.0.0.1:54586.service: Deactivated successfully. Apr 13 23:58:16.063839 systemd[1]: session-7.scope: Deactivated successfully. Apr 13 23:58:16.067384 systemd[1]: session-7.scope: Consumed 3min 33.569s CPU time, 162.2M memory peak, 0B memory swap peak. Apr 13 23:58:16.080915 systemd-logind[1457]: Session 7 logged out. Waiting for processes to exit. Apr 13 23:58:16.103196 systemd-logind[1457]: Removed session 7. Apr 13 23:58:16.591543 kubelet[2848]: E0413 23:58:16.590844 2848 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 13 23:58:16.629214 kubelet[2848]: E0413 23:58:16.624536 2848 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.839s" Apr 13 23:58:17.417286 kubelet[2848]: E0413 23:58:17.416569 2848 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"