Apr 20 19:06:38.819958 kernel: Linux version 6.12.81-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 15.2.1_p20260214 p5) 15.2.1 20260214, GNU ld (Gentoo 2.46.0 p1) 2.46.0) #1 SMP PREEMPT_DYNAMIC Tue Apr 14 02:21:25 -00 2026 Apr 20 19:06:38.820069 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=70de22794192cd52e167b5a4b1ae0509811ded61dbe4152dfc02378f843ae81a Apr 20 19:06:38.820085 kernel: BIOS-provided physical RAM map: Apr 20 19:06:38.820114 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000002ffff] usable Apr 20 19:06:38.820121 kernel: BIOS-e820: [mem 0x0000000000030000-0x000000000004ffff] reserved Apr 20 19:06:38.820130 kernel: BIOS-e820: [mem 0x0000000000050000-0x000000000009efff] usable Apr 20 19:06:38.820140 kernel: BIOS-e820: [mem 0x000000000009f000-0x000000000009ffff] reserved Apr 20 19:06:38.820169 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009b8ecfff] usable Apr 20 19:06:38.820178 kernel: BIOS-e820: [mem 0x000000009b8ed000-0x000000009bb6cfff] reserved Apr 20 19:06:38.820186 kernel: BIOS-e820: [mem 0x000000009bb6d000-0x000000009bb7efff] ACPI data Apr 20 19:06:38.820195 kernel: BIOS-e820: [mem 0x000000009bb7f000-0x000000009bbfefff] ACPI NVS Apr 20 19:06:38.820219 kernel: BIOS-e820: [mem 0x000000009bbff000-0x000000009bfb0fff] usable Apr 20 19:06:38.820229 kernel: BIOS-e820: [mem 0x000000009bfb1000-0x000000009bfb4fff] reserved Apr 20 19:06:38.820236 kernel: BIOS-e820: [mem 0x000000009bfb5000-0x000000009bfb6fff] ACPI NVS Apr 20 19:06:38.820246 kernel: BIOS-e820: [mem 0x000000009bfb7000-0x000000009bffffff] usable Apr 20 19:06:38.820254 kernel: BIOS-e820: [mem 0x000000009c000000-0x000000009cffffff] reserved Apr 20 19:06:38.820261 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Apr 20 19:06:38.820268 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Apr 20 19:06:38.820294 kernel: NX (Execute Disable) protection: active Apr 20 19:06:38.820319 kernel: APIC: Static calls initialized Apr 20 19:06:38.820329 kernel: e820: update [mem 0x9a142018-0x9a14bc57] usable ==> usable Apr 20 19:06:38.820339 kernel: e820: update [mem 0x9a105018-0x9a141e57] usable ==> usable Apr 20 19:06:38.820346 kernel: extended physical RAM map: Apr 20 19:06:38.820357 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000002ffff] usable Apr 20 19:06:38.820366 kernel: reserve setup_data: [mem 0x0000000000030000-0x000000000004ffff] reserved Apr 20 19:06:38.820414 kernel: reserve setup_data: [mem 0x0000000000050000-0x000000000009efff] usable Apr 20 19:06:38.820425 kernel: reserve setup_data: [mem 0x000000000009f000-0x000000000009ffff] reserved Apr 20 19:06:38.820435 kernel: reserve setup_data: [mem 0x0000000000100000-0x000000009a105017] usable Apr 20 19:06:38.820442 kernel: reserve setup_data: [mem 0x000000009a105018-0x000000009a141e57] usable Apr 20 19:06:38.820450 kernel: reserve setup_data: [mem 0x000000009a141e58-0x000000009a142017] usable Apr 20 19:06:38.820457 kernel: reserve setup_data: [mem 0x000000009a142018-0x000000009a14bc57] usable Apr 20 19:06:38.820464 kernel: reserve setup_data: [mem 0x000000009a14bc58-0x000000009b8ecfff] usable Apr 20 19:06:38.820474 kernel: reserve setup_data: [mem 0x000000009b8ed000-0x000000009bb6cfff] reserved Apr 20 19:06:38.820481 kernel: reserve setup_data: [mem 0x000000009bb6d000-0x000000009bb7efff] ACPI data Apr 20 19:06:38.820488 kernel: reserve setup_data: [mem 0x000000009bb7f000-0x000000009bbfefff] ACPI NVS Apr 20 19:06:38.820496 kernel: reserve setup_data: [mem 0x000000009bbff000-0x000000009bfb0fff] usable Apr 20 19:06:38.820503 kernel: reserve setup_data: [mem 0x000000009bfb1000-0x000000009bfb4fff] reserved Apr 20 19:06:38.820513 kernel: reserve setup_data: [mem 0x000000009bfb5000-0x000000009bfb6fff] ACPI NVS Apr 20 19:06:38.820520 kernel: reserve setup_data: [mem 0x000000009bfb7000-0x000000009bffffff] usable Apr 20 19:06:38.820528 kernel: reserve setup_data: [mem 0x000000009c000000-0x000000009cffffff] reserved Apr 20 19:06:38.820540 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Apr 20 19:06:38.820565 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Apr 20 19:06:38.820577 kernel: efi: EFI v2.7 by EDK II Apr 20 19:06:38.820585 kernel: efi: SMBIOS=0x9b9d5000 ACPI=0x9bb7e000 ACPI 2.0=0x9bb7e014 MEMATTR=0x9a1b4018 RNG=0x9bb73018 Apr 20 19:06:38.820592 kernel: random: crng init done Apr 20 19:06:38.820600 kernel: Kernel is locked down from EFI Secure Boot; see man kernel_lockdown.7 Apr 20 19:06:38.820610 kernel: secureboot: Secure boot enabled Apr 20 19:06:38.820617 kernel: SMBIOS 2.8 present. Apr 20 19:06:38.820625 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Apr 20 19:06:38.820633 kernel: DMI: Memory slots populated: 1/1 Apr 20 19:06:38.820640 kernel: Hypervisor detected: KVM Apr 20 19:06:38.820648 kernel: last_pfn = 0x9c000 max_arch_pfn = 0x10000000000 Apr 20 19:06:38.820659 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Apr 20 19:06:38.820667 kernel: kvm-clock: using sched offset of 19760385138 cycles Apr 20 19:06:38.820675 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 20 19:06:38.820684 kernel: tsc: Detected 2793.438 MHz processor Apr 20 19:06:38.820694 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Apr 20 19:06:38.820717 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Apr 20 19:06:38.820728 kernel: last_pfn = 0x9c000 max_arch_pfn = 0x10000000000 Apr 20 19:06:38.820737 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Apr 20 19:06:38.820745 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Apr 20 19:06:38.820759 kernel: Using GB pages for direct mapping Apr 20 19:06:38.820782 kernel: ACPI: Early table checksum verification disabled Apr 20 19:06:38.820792 kernel: ACPI: RSDP 0x000000009BB7E014 000024 (v02 BOCHS ) Apr 20 19:06:38.820804 kernel: ACPI: XSDT 0x000000009BB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Apr 20 19:06:38.820812 kernel: ACPI: FACP 0x000000009BB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 20 19:06:38.820820 kernel: ACPI: DSDT 0x000000009BB7A000 002237 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 20 19:06:38.820831 kernel: ACPI: FACS 0x000000009BBDD000 000040 Apr 20 19:06:38.820842 kernel: ACPI: APIC 0x000000009BB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 20 19:06:38.820850 kernel: ACPI: HPET 0x000000009BB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 20 19:06:38.820880 kernel: ACPI: MCFG 0x000000009BB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 20 19:06:38.820888 kernel: ACPI: WAET 0x000000009BB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 20 19:06:38.820897 kernel: ACPI: BGRT 0x000000009BB74000 000038 (v01 INTEL EDK2 00000002 01000013) Apr 20 19:06:38.820905 kernel: ACPI: Reserving FACP table memory at [mem 0x9bb79000-0x9bb790f3] Apr 20 19:06:38.820913 kernel: ACPI: Reserving DSDT table memory at [mem 0x9bb7a000-0x9bb7c236] Apr 20 19:06:38.820930 kernel: ACPI: Reserving FACS table memory at [mem 0x9bbdd000-0x9bbdd03f] Apr 20 19:06:38.820940 kernel: ACPI: Reserving APIC table memory at [mem 0x9bb78000-0x9bb7808f] Apr 20 19:06:38.820982 kernel: ACPI: Reserving HPET table memory at [mem 0x9bb77000-0x9bb77037] Apr 20 19:06:38.820995 kernel: ACPI: Reserving MCFG table memory at [mem 0x9bb76000-0x9bb7603b] Apr 20 19:06:38.821052 kernel: ACPI: Reserving WAET table memory at [mem 0x9bb75000-0x9bb75027] Apr 20 19:06:38.821060 kernel: ACPI: Reserving BGRT table memory at [mem 0x9bb74000-0x9bb74037] Apr 20 19:06:38.821069 kernel: No NUMA configuration found Apr 20 19:06:38.821077 kernel: Faking a node at [mem 0x0000000000000000-0x000000009bffffff] Apr 20 19:06:38.821091 kernel: NODE_DATA(0) allocated [mem 0x9bf57dc0-0x9bf5efff] Apr 20 19:06:38.821139 kernel: Zone ranges: Apr 20 19:06:38.821151 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Apr 20 19:06:38.821175 kernel: DMA32 [mem 0x0000000001000000-0x000000009bffffff] Apr 20 19:06:38.821186 kernel: Normal empty Apr 20 19:06:38.821194 kernel: Device empty Apr 20 19:06:38.821203 kernel: Movable zone start for each node Apr 20 19:06:38.821212 kernel: Early memory node ranges Apr 20 19:06:38.821221 kernel: node 0: [mem 0x0000000000001000-0x000000000002ffff] Apr 20 19:06:38.821257 kernel: node 0: [mem 0x0000000000050000-0x000000000009efff] Apr 20 19:06:38.821265 kernel: node 0: [mem 0x0000000000100000-0x000000009b8ecfff] Apr 20 19:06:38.821274 kernel: node 0: [mem 0x000000009bbff000-0x000000009bfb0fff] Apr 20 19:06:38.821282 kernel: node 0: [mem 0x000000009bfb7000-0x000000009bffffff] Apr 20 19:06:38.821291 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009bffffff] Apr 20 19:06:38.821299 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 20 19:06:38.821308 kernel: On node 0, zone DMA: 32 pages in unavailable ranges Apr 20 19:06:38.822352 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Apr 20 19:06:38.822368 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Apr 20 19:06:38.822424 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Apr 20 19:06:38.822436 kernel: On node 0, zone DMA32: 16384 pages in unavailable ranges Apr 20 19:06:38.822448 kernel: ACPI: PM-Timer IO Port: 0x608 Apr 20 19:06:38.822483 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Apr 20 19:06:38.822493 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Apr 20 19:06:38.822532 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Apr 20 19:06:38.822559 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Apr 20 19:06:38.822570 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Apr 20 19:06:38.822581 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Apr 20 19:06:38.822592 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Apr 20 19:06:38.822603 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Apr 20 19:06:38.822614 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Apr 20 19:06:38.822647 kernel: TSC deadline timer available Apr 20 19:06:38.822657 kernel: CPU topo: Max. logical packages: 1 Apr 20 19:06:38.822667 kernel: CPU topo: Max. logical dies: 1 Apr 20 19:06:38.822676 kernel: CPU topo: Max. dies per package: 1 Apr 20 19:06:38.822685 kernel: CPU topo: Max. threads per core: 1 Apr 20 19:06:38.822760 kernel: CPU topo: Num. cores per package: 4 Apr 20 19:06:38.822792 kernel: CPU topo: Num. threads per package: 4 Apr 20 19:06:38.822803 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Apr 20 19:06:38.822832 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Apr 20 19:06:38.822844 kernel: kvm-guest: KVM setup pv remote TLB flush Apr 20 19:06:38.822876 kernel: kvm-guest: setup PV sched yield Apr 20 19:06:38.822887 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Apr 20 19:06:38.822899 kernel: Booting paravirtualized kernel on KVM Apr 20 19:06:38.822911 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Apr 20 19:06:38.822943 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Apr 20 19:06:38.822955 kernel: percpu: Embedded 60 pages/cpu s207960 r8192 d29608 u524288 Apr 20 19:06:38.822966 kernel: pcpu-alloc: s207960 r8192 d29608 u524288 alloc=1*2097152 Apr 20 19:06:38.822978 kernel: pcpu-alloc: [0] 0 1 2 3 Apr 20 19:06:38.822989 kernel: kvm-guest: PV spinlocks enabled Apr 20 19:06:38.823001 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Apr 20 19:06:38.823040 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=70de22794192cd52e167b5a4b1ae0509811ded61dbe4152dfc02378f843ae81a Apr 20 19:06:38.823075 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 20 19:06:38.823087 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 20 19:06:38.823099 kernel: Fallback order for Node 0: 0 Apr 20 19:06:38.823110 kernel: Built 1 zonelists, mobility grouping on. Total pages: 638054 Apr 20 19:06:38.823122 kernel: Policy zone: DMA32 Apr 20 19:06:38.823133 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 20 19:06:38.823145 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Apr 20 19:06:38.823177 kernel: ftrace: allocating 40346 entries in 158 pages Apr 20 19:06:38.823189 kernel: ftrace: allocated 158 pages with 5 groups Apr 20 19:06:38.823200 kernel: Dynamic Preempt: voluntary Apr 20 19:06:38.823212 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 20 19:06:38.823225 kernel: rcu: RCU event tracing is enabled. Apr 20 19:06:38.823237 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Apr 20 19:06:38.823249 kernel: Trampoline variant of Tasks RCU enabled. Apr 20 19:06:38.823282 kernel: Rude variant of Tasks RCU enabled. Apr 20 19:06:38.823293 kernel: Tracing variant of Tasks RCU enabled. Apr 20 19:06:38.823321 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 20 19:06:38.823333 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Apr 20 19:06:38.823343 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 20 19:06:38.823354 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 20 19:06:38.823364 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Apr 20 19:06:38.823439 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Apr 20 19:06:38.823449 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 20 19:06:38.823459 kernel: Console: colour dummy device 80x25 Apr 20 19:06:38.823469 kernel: printk: legacy console [ttyS0] enabled Apr 20 19:06:38.824417 kernel: ACPI: Core revision 20240827 Apr 20 19:06:38.824437 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Apr 20 19:06:38.824450 kernel: APIC: Switch to symmetric I/O mode setup Apr 20 19:06:38.824495 kernel: x2apic enabled Apr 20 19:06:38.824506 kernel: APIC: Switched APIC routing to: physical x2apic Apr 20 19:06:38.824534 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Apr 20 19:06:38.824545 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Apr 20 19:06:38.824556 kernel: kvm-guest: setup PV IPIs Apr 20 19:06:38.824566 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Apr 20 19:06:38.824576 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x284409db922, max_idle_ns: 440795228871 ns Apr 20 19:06:38.824615 kernel: Calibrating delay loop (skipped) preset value.. 5586.87 BogoMIPS (lpj=2793438) Apr 20 19:06:38.824625 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Apr 20 19:06:38.824634 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Apr 20 19:06:38.824645 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Apr 20 19:06:38.824654 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Apr 20 19:06:38.824664 kernel: Spectre V2 : Mitigation: Retpolines Apr 20 19:06:38.824675 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Apr 20 19:06:38.824733 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Apr 20 19:06:38.824745 kernel: RETBleed: Vulnerable Apr 20 19:06:38.824755 kernel: Speculative Store Bypass: Vulnerable Apr 20 19:06:38.824765 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Apr 20 19:06:38.824776 kernel: GDS: Unknown: Dependent on hypervisor status Apr 20 19:06:38.824786 kernel: active return thunk: its_return_thunk Apr 20 19:06:38.824796 kernel: ITS: Mitigation: Aligned branch/return thunks Apr 20 19:06:38.824835 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Apr 20 19:06:38.824845 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Apr 20 19:06:38.824852 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Apr 20 19:06:38.824860 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Apr 20 19:06:38.824870 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Apr 20 19:06:38.824880 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Apr 20 19:06:38.824890 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Apr 20 19:06:38.824928 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Apr 20 19:06:38.824936 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Apr 20 19:06:38.824942 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Apr 20 19:06:38.824949 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Apr 20 19:06:38.824955 kernel: Freeing SMP alternatives memory: 32K Apr 20 19:06:38.824962 kernel: pid_max: default: 32768 minimum: 301 Apr 20 19:06:38.824968 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Apr 20 19:06:38.825036 kernel: landlock: Up and running. Apr 20 19:06:38.825047 kernel: SELinux: Initializing. Apr 20 19:06:38.825056 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 20 19:06:38.825065 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 20 19:06:38.825075 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8370C CPU @ 2.80GHz (family: 0x6, model: 0x6a, stepping: 0x6) Apr 20 19:06:38.825084 kernel: Performance Events: unsupported p6 CPU model 106 no PMU driver, software events only. Apr 20 19:06:38.825093 kernel: signal: max sigframe size: 3632 Apr 20 19:06:38.825125 kernel: rcu: Hierarchical SRCU implementation. Apr 20 19:06:38.825138 kernel: rcu: Max phase no-delay instances is 400. Apr 20 19:06:38.825149 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Apr 20 19:06:38.825159 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Apr 20 19:06:38.825170 kernel: smp: Bringing up secondary CPUs ... Apr 20 19:06:38.825180 kernel: smpboot: x86: Booting SMP configuration: Apr 20 19:06:38.825191 kernel: .... node #0, CPUs: #1 #2 #3 Apr 20 19:06:38.825225 kernel: smp: Brought up 1 node, 4 CPUs Apr 20 19:06:38.825232 kernel: smpboot: Total of 4 processors activated (22347.50 BogoMIPS) Apr 20 19:06:38.825240 kernel: Memory: 2381824K/2552216K available (14336K kernel code, 2458K rwdata, 31736K rodata, 15944K init, 2284K bss, 164496K reserved, 0K cma-reserved) Apr 20 19:06:38.825247 kernel: devtmpfs: initialized Apr 20 19:06:38.825253 kernel: x86/mm: Memory block size: 128MB Apr 20 19:06:38.825260 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9bb7f000-0x9bbfefff] (524288 bytes) Apr 20 19:06:38.825266 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9bfb5000-0x9bfb6fff] (8192 bytes) Apr 20 19:06:38.825291 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 20 19:06:38.825298 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Apr 20 19:06:38.825304 kernel: pinctrl core: initialized pinctrl subsystem Apr 20 19:06:38.825311 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 20 19:06:38.825317 kernel: audit: initializing netlink subsys (disabled) Apr 20 19:06:38.825323 kernel: audit: type=2000 audit(1776711968.210:1): state=initialized audit_enabled=0 res=1 Apr 20 19:06:38.825330 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 20 19:06:38.825353 kernel: thermal_sys: Registered thermal governor 'user_space' Apr 20 19:06:38.825360 kernel: cpuidle: using governor menu Apr 20 19:06:38.825366 kernel: efi: Freeing EFI boot services memory: 42800K Apr 20 19:06:38.825423 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 20 19:06:38.825432 kernel: dca service started, version 1.12.1 Apr 20 19:06:38.825443 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Apr 20 19:06:38.825453 kernel: PCI: Using configuration type 1 for base access Apr 20 19:06:38.825463 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Apr 20 19:06:38.825502 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 20 19:06:38.825511 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Apr 20 19:06:38.825521 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 20 19:06:38.825531 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Apr 20 19:06:38.825541 kernel: ACPI: Added _OSI(Module Device) Apr 20 19:06:38.825553 kernel: ACPI: Added _OSI(Processor Device) Apr 20 19:06:38.825566 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 20 19:06:38.827268 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 20 19:06:38.827282 kernel: ACPI: Interpreter enabled Apr 20 19:06:38.827295 kernel: ACPI: PM: (supports S0 S5) Apr 20 19:06:38.827308 kernel: ACPI: Using IOAPIC for interrupt routing Apr 20 19:06:38.827321 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Apr 20 19:06:38.827334 kernel: PCI: Using E820 reservations for host bridge windows Apr 20 19:06:38.827347 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Apr 20 19:06:38.827438 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 20 19:06:38.827845 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 20 19:06:38.828079 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Apr 20 19:06:38.828234 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Apr 20 19:06:38.828248 kernel: PCI host bridge to bus 0000:00 Apr 20 19:06:38.844947 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Apr 20 19:06:38.846626 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Apr 20 19:06:38.846782 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Apr 20 19:06:38.846927 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Apr 20 19:06:38.865183 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Apr 20 19:06:38.867098 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Apr 20 19:06:38.870444 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 20 19:06:38.870716 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Apr 20 19:06:38.870883 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Apr 20 19:06:38.870992 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Apr 20 19:06:38.871183 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Apr 20 19:06:38.871437 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Apr 20 19:06:38.871590 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Apr 20 19:06:38.871746 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Apr 20 19:06:38.871892 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Apr 20 19:06:38.873232 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Apr 20 19:06:38.874984 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Apr 20 19:06:38.876488 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Apr 20 19:06:38.877347 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Apr 20 19:06:38.878881 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Apr 20 19:06:38.879116 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Apr 20 19:06:38.879272 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Apr 20 19:06:38.880979 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Apr 20 19:06:38.884420 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Apr 20 19:06:38.884664 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Apr 20 19:06:38.884998 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Apr 20 19:06:38.885180 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Apr 20 19:06:38.885285 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Apr 20 19:06:38.885583 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Apr 20 19:06:38.885744 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Apr 20 19:06:38.885901 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Apr 20 19:06:38.886043 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Apr 20 19:06:38.886163 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Apr 20 19:06:38.886178 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Apr 20 19:06:38.886230 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Apr 20 19:06:38.886241 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Apr 20 19:06:38.886249 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Apr 20 19:06:38.886256 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Apr 20 19:06:38.886262 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Apr 20 19:06:38.886269 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Apr 20 19:06:38.886276 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Apr 20 19:06:38.886306 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Apr 20 19:06:38.886313 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Apr 20 19:06:38.886319 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Apr 20 19:06:38.886325 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Apr 20 19:06:38.886332 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Apr 20 19:06:38.886339 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Apr 20 19:06:38.886345 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Apr 20 19:06:38.887176 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Apr 20 19:06:38.887192 kernel: iommu: Default domain type: Translated Apr 20 19:06:38.887203 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Apr 20 19:06:38.887215 kernel: efivars: Registered efivars operations Apr 20 19:06:38.887227 kernel: PCI: Using ACPI for IRQ routing Apr 20 19:06:38.887239 kernel: PCI: pci_cache_line_size set to 64 bytes Apr 20 19:06:38.887250 kernel: e820: reserve RAM buffer [mem 0x0009f000-0x0009ffff] Apr 20 19:06:38.888347 kernel: e820: reserve RAM buffer [mem 0x9a105018-0x9bffffff] Apr 20 19:06:38.888364 kernel: e820: reserve RAM buffer [mem 0x9a142018-0x9bffffff] Apr 20 19:06:38.888425 kernel: e820: reserve RAM buffer [mem 0x9b8ed000-0x9bffffff] Apr 20 19:06:38.888437 kernel: e820: reserve RAM buffer [mem 0x9bfb1000-0x9bffffff] Apr 20 19:06:38.888752 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Apr 20 19:06:38.888898 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Apr 20 19:06:38.890680 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Apr 20 19:06:38.890782 kernel: vgaarb: loaded Apr 20 19:06:38.890793 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Apr 20 19:06:38.890804 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Apr 20 19:06:38.890815 kernel: clocksource: Switched to clocksource kvm-clock Apr 20 19:06:38.890826 kernel: VFS: Disk quotas dquot_6.6.0 Apr 20 19:06:38.890838 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 20 19:06:38.890848 kernel: pnp: PnP ACPI init Apr 20 19:06:38.892687 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Apr 20 19:06:38.892756 kernel: pnp: PnP ACPI: found 6 devices Apr 20 19:06:38.892768 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Apr 20 19:06:38.892778 kernel: NET: Registered PF_INET protocol family Apr 20 19:06:38.892789 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 20 19:06:38.892801 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 20 19:06:38.892812 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 20 19:06:38.894211 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 20 19:06:38.894224 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 20 19:06:38.894236 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 20 19:06:38.894248 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 20 19:06:38.894261 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 20 19:06:38.894273 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 20 19:06:38.898258 kernel: NET: Registered PF_XDP protocol family Apr 20 19:06:38.939105 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Apr 20 19:06:38.943948 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Apr 20 19:06:38.944201 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Apr 20 19:06:38.944332 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Apr 20 19:06:38.949501 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Apr 20 19:06:38.949641 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Apr 20 19:06:38.949807 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Apr 20 19:06:38.949925 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Apr 20 19:06:38.949938 kernel: PCI: CLS 0 bytes, default 64 Apr 20 19:06:38.949947 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Apr 20 19:06:38.949957 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x284409db922, max_idle_ns: 440795228871 ns Apr 20 19:06:38.949967 kernel: Initialise system trusted keyrings Apr 20 19:06:38.949977 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 20 19:06:38.950662 kernel: Key type asymmetric registered Apr 20 19:06:38.950670 kernel: Asymmetric key parser 'x509' registered Apr 20 19:06:38.950677 kernel: hrtimer: interrupt took 30858336 ns Apr 20 19:06:38.950685 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 20 19:06:38.950821 kernel: io scheduler mq-deadline registered Apr 20 19:06:38.950843 kernel: io scheduler kyber registered Apr 20 19:06:38.950850 kernel: io scheduler bfq registered Apr 20 19:06:38.950879 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Apr 20 19:06:38.950890 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Apr 20 19:06:38.950899 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Apr 20 19:06:38.950910 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Apr 20 19:06:38.950919 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 20 19:06:38.950929 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Apr 20 19:06:38.950940 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Apr 20 19:06:38.950979 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Apr 20 19:06:38.950988 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Apr 20 19:06:38.951312 kernel: rtc_cmos 00:04: RTC can wake from S4 Apr 20 19:06:38.951333 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Apr 20 19:06:38.952090 kernel: rtc_cmos 00:04: registered as rtc0 Apr 20 19:06:38.952223 kernel: rtc_cmos 00:04: setting system clock to 2026-04-20T19:06:19 UTC (1776711979) Apr 20 19:06:38.952364 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Apr 20 19:06:38.952435 kernel: intel_pstate: CPU model not supported Apr 20 19:06:38.952447 kernel: efifb: probing for efifb Apr 20 19:06:38.952458 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Apr 20 19:06:38.952469 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Apr 20 19:06:38.952480 kernel: efifb: scrolling: redraw Apr 20 19:06:38.952491 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 20 19:06:38.953838 kernel: Console: switching to colour frame buffer device 160x50 Apr 20 19:06:38.953904 kernel: fb0: EFI VGA frame buffer device Apr 20 19:06:38.953913 kernel: pstore: Using crash dump compression: deflate Apr 20 19:06:38.953921 kernel: pstore: Registered efi_pstore as persistent store backend Apr 20 19:06:38.953929 kernel: NET: Registered PF_INET6 protocol family Apr 20 19:06:38.953956 kernel: Segment Routing with IPv6 Apr 20 19:06:38.953964 kernel: In-situ OAM (IOAM) with IPv6 Apr 20 19:06:38.953972 kernel: NET: Registered PF_PACKET protocol family Apr 20 19:06:38.953980 kernel: Key type dns_resolver registered Apr 20 19:06:38.953989 kernel: IPI shorthand broadcast: enabled Apr 20 19:06:38.953997 kernel: sched_clock: Marking stable (10636016398, 2206839198)->(14513605794, -1670750198) Apr 20 19:06:38.954030 kernel: registered taskstats version 1 Apr 20 19:06:38.954057 kernel: Loading compiled-in X.509 certificates Apr 20 19:06:38.954065 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.81-flatcar: 7cf14208c08026297bea8a5678f7340932b35e4b' Apr 20 19:06:38.954074 kernel: Demotion targets for Node 0: null Apr 20 19:06:38.954082 kernel: Key type .fscrypt registered Apr 20 19:06:38.954090 kernel: Key type fscrypt-provisioning registered Apr 20 19:06:38.954098 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 20 19:06:38.954106 kernel: ima: Allocated hash algorithm: sha1 Apr 20 19:06:38.954722 kernel: ima: No architecture policies found Apr 20 19:06:38.954731 kernel: clk: Disabling unused clocks Apr 20 19:06:38.954739 kernel: Freeing unused kernel image (initmem) memory: 15944K Apr 20 19:06:38.954748 kernel: Write protecting the kernel read-only data: 47104k Apr 20 19:06:38.954756 kernel: Freeing unused kernel image (rodata/data gap) memory: 1032K Apr 20 19:06:38.954764 kernel: Run /init as init process Apr 20 19:06:38.954772 kernel: with arguments: Apr 20 19:06:38.956042 kernel: /init Apr 20 19:06:38.956985 kernel: with environment: Apr 20 19:06:38.956994 kernel: HOME=/ Apr 20 19:06:38.957002 kernel: TERM=linux Apr 20 19:06:38.957071 kernel: SCSI subsystem initialized Apr 20 19:06:38.957078 kernel: libata version 3.00 loaded. Apr 20 19:06:38.957276 kernel: ahci 0000:00:1f.2: version 3.0 Apr 20 19:06:38.957331 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Apr 20 19:06:38.957525 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Apr 20 19:06:38.957627 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Apr 20 19:06:38.957723 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Apr 20 19:06:38.957965 kernel: scsi host0: ahci Apr 20 19:06:38.958133 kernel: scsi host1: ahci Apr 20 19:06:38.958284 kernel: scsi host2: ahci Apr 20 19:06:38.958946 kernel: scsi host3: ahci Apr 20 19:06:38.959906 kernel: scsi host4: ahci Apr 20 19:06:38.963208 kernel: scsi host5: ahci Apr 20 19:06:38.963275 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 26 lpm-pol 1 Apr 20 19:06:38.964220 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 26 lpm-pol 1 Apr 20 19:06:38.964237 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 26 lpm-pol 1 Apr 20 19:06:38.964248 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 26 lpm-pol 1 Apr 20 19:06:38.964261 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 26 lpm-pol 1 Apr 20 19:06:38.964272 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 26 lpm-pol 1 Apr 20 19:06:38.964283 kernel: ata4: SATA link down (SStatus 0 SControl 300) Apr 20 19:06:38.964296 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Apr 20 19:06:38.964347 kernel: ata2: SATA link down (SStatus 0 SControl 300) Apr 20 19:06:38.964358 kernel: ata6: SATA link down (SStatus 0 SControl 300) Apr 20 19:06:38.965086 kernel: ata5: SATA link down (SStatus 0 SControl 300) Apr 20 19:06:38.965149 kernel: ata3.00: LPM support broken, forcing max_power Apr 20 19:06:38.965162 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Apr 20 19:06:38.966130 kernel: ata3.00: applying bridge limits Apr 20 19:06:38.966146 kernel: ata1: SATA link down (SStatus 0 SControl 300) Apr 20 19:06:38.967085 kernel: ata3.00: LPM support broken, forcing max_power Apr 20 19:06:38.967101 kernel: ata3.00: configured for UDMA/100 Apr 20 19:06:38.967474 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 20 19:06:38.967673 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Apr 20 19:06:38.967803 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Apr 20 19:06:38.967923 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Apr 20 19:06:38.967962 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 20 19:06:38.967969 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 20 19:06:38.967976 kernel: GPT:16515071 != 27000831 Apr 20 19:06:38.967983 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 20 19:06:38.967990 kernel: GPT:16515071 != 27000831 Apr 20 19:06:38.967997 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 20 19:06:38.968003 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 20 19:06:38.969985 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Apr 20 19:06:38.970037 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 20 19:06:38.970049 kernel: device-mapper: uevent: version 1.0.3 Apr 20 19:06:38.970060 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Apr 20 19:06:38.970070 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Apr 20 19:06:38.970081 kernel: raid6: avx512x4 gen() 28286 MB/s Apr 20 19:06:38.970091 kernel: raid6: avx512x2 gen() 29663 MB/s Apr 20 19:06:38.974209 kernel: raid6: avx512x1 gen() 7480 MB/s Apr 20 19:06:38.974268 kernel: raid6: avx2x4 gen() 14116 MB/s Apr 20 19:06:38.974280 kernel: raid6: avx2x2 gen() 24134 MB/s Apr 20 19:06:38.974290 kernel: raid6: avx2x1 gen() 7187 MB/s Apr 20 19:06:38.974300 kernel: raid6: using algorithm avx512x2 gen() 29663 MB/s Apr 20 19:06:38.974309 kernel: raid6: .... xor() 11746 MB/s, rmw enabled Apr 20 19:06:38.974321 kernel: raid6: using avx512x2 recovery algorithm Apr 20 19:06:38.974427 kernel: xor: automatically using best checksumming function avx Apr 20 19:06:38.974439 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 20 19:06:38.974449 kernel: BTRFS: device fsid 2b1891e6-d4d2-4c02-a1ed-3a6feccae86f devid 1 transid 45 /dev/mapper/usr (253:0) scanned by mount (181) Apr 20 19:06:38.974459 kernel: BTRFS info (device dm-0): first mount of filesystem 2b1891e6-d4d2-4c02-a1ed-3a6feccae86f Apr 20 19:06:38.974470 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Apr 20 19:06:38.974481 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Apr 20 19:06:38.974490 kernel: BTRFS info (device dm-0 state E): enabling free space tree Apr 20 19:06:38.974528 kernel: loop: module loaded Apr 20 19:06:38.974538 kernel: loop0: detected capacity change from 0 to 106960 Apr 20 19:06:38.974549 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 20 19:06:38.974561 systemd[1]: /etc/systemd/system.conf.d/nocgroup.conf:2: Support for option DefaultCPUAccounting= has been removed and it is ignored Apr 20 19:06:38.974574 systemd[1]: /etc/systemd/system.conf.d/nocgroup.conf:5: Support for option DefaultBlockIOAccounting= has been removed and it is ignored Apr 20 19:06:38.974586 systemd[1]: Successfully made /usr/ read-only. Apr 20 19:06:38.974625 systemd[1]: systemd 258.3 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 20 19:06:38.974636 systemd[1]: Detected virtualization kvm. Apr 20 19:06:38.974647 systemd[1]: Detected architecture x86-64. Apr 20 19:06:38.974659 systemd[1]: Running in initrd. Apr 20 19:06:38.974669 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Apr 20 19:06:38.974678 systemd[1]: No hostname configured, using default hostname. Apr 20 19:06:38.974711 systemd[1]: Hostname set to . Apr 20 19:06:38.974722 systemd[1]: Queued start job for default target initrd.target. Apr 20 19:06:38.974732 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Apr 20 19:06:38.974743 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 20 19:06:38.974755 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 20 19:06:38.974768 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 20 19:06:38.974806 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 20 19:06:38.974817 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 20 19:06:38.974828 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 20 19:06:38.974838 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 20 19:06:38.974850 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 20 19:06:38.974860 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Apr 20 19:06:38.974870 systemd[1]: Reached target paths.target - Path Units. Apr 20 19:06:38.974907 systemd[1]: Reached target slices.target - Slice Units. Apr 20 19:06:38.974918 systemd[1]: Reached target swap.target - Swaps. Apr 20 19:06:38.974929 systemd[1]: Reached target timers.target - Timer Units. Apr 20 19:06:38.974939 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 20 19:06:38.974949 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 20 19:06:38.974960 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Apr 20 19:06:38.974998 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 20 19:06:38.975035 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Apr 20 19:06:38.975046 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 20 19:06:38.975056 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 20 19:06:38.975067 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 20 19:06:38.975077 systemd[1]: Reached target sockets.target - Socket Units. Apr 20 19:06:38.975088 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 20 19:06:38.979657 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 20 19:06:38.980266 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 20 19:06:38.980289 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 20 19:06:38.980311 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Apr 20 19:06:38.980343 systemd[1]: Starting systemd-fsck-usr.service... Apr 20 19:06:38.980494 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 20 19:06:38.980516 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 20 19:06:38.980536 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 20 19:06:38.980556 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 20 19:06:38.980566 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 20 19:06:38.980603 systemd[1]: Finished systemd-fsck-usr.service. Apr 20 19:06:38.983469 systemd-journald[320]: Collecting audit messages is enabled. Apr 20 19:06:38.983664 kernel: audit: type=1130 audit(1776711998.803:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:38.983756 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 20 19:06:38.983789 systemd-journald[320]: Journal started Apr 20 19:06:38.983885 systemd-journald[320]: Runtime Journal (/run/log/journal/8118aa160470456397a4c2dcfc0629e4) is 5.9M, max 47.8M, 41.8M free. Apr 20 19:06:38.803000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:38.996968 systemd[1]: Started systemd-journald.service - Journal Service. Apr 20 19:06:38.997239 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 20 19:06:39.003000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.010033 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 20 19:06:39.016998 kernel: audit: type=1130 audit(1776711999.003:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.021175 kernel: Bridge firewalling registered Apr 20 19:06:39.023194 systemd-modules-load[323]: Inserted module 'br_netfilter' Apr 20 19:06:39.038000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.062273 kernel: audit: type=1130 audit(1776711999.038:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.174504 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 20 19:06:39.216000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.247748 kernel: audit: type=1130 audit(1776711999.216:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.275411 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 20 19:06:39.299000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.322477 kernel: audit: type=1130 audit(1776711999.299:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.467934 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 20 19:06:39.485984 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 20 19:06:39.497869 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 20 19:06:39.508563 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 20 19:06:39.661515 systemd-tmpfiles[342]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Apr 20 19:06:39.670657 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 20 19:06:39.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.687083 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 20 19:06:39.715000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.721083 kernel: audit: type=1130 audit(1776711999.670:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.723873 kernel: audit: type=1130 audit(1776711999.715:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.734270 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 20 19:06:39.735000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.755474 kernel: audit: type=1130 audit(1776711999.735:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.804000 audit: BPF prog-id=5 op=LOAD Apr 20 19:06:39.809138 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 20 19:06:39.818178 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 20 19:06:39.833000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.857900 kernel: audit: type=1334 audit(1776711999.804:10): prog-id=5 op=LOAD Apr 20 19:06:39.857944 kernel: audit: type=1130 audit(1776711999.833:11): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:39.890132 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 20 19:06:40.325654 dracut-cmdline[360]: dracut-109 Apr 20 19:06:40.432213 dracut-cmdline[360]: Using kernel command line parameters: SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=70de22794192cd52e167b5a4b1ae0509811ded61dbe4152dfc02378f843ae81a Apr 20 19:06:40.821261 systemd-resolved[359]: Positive Trust Anchors: Apr 20 19:06:40.822208 systemd-resolved[359]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 20 19:06:40.822214 systemd-resolved[359]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Apr 20 19:06:40.822261 systemd-resolved[359]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 20 19:06:41.056134 systemd-resolved[359]: Defaulting to hostname 'linux'. Apr 20 19:06:41.158204 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 20 19:06:41.189000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:41.190344 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 20 19:06:45.970296 kernel: Loading iSCSI transport class v2.0-870. Apr 20 19:06:46.210750 kernel: iscsi: registered transport (tcp) Apr 20 19:06:46.553735 kernel: iscsi: registered transport (qla4xxx) Apr 20 19:06:46.555720 kernel: QLogic iSCSI HBA Driver Apr 20 19:06:47.835712 systemd[1]: Starting systemd-network-generator.service - Generate Network Units from Kernel Command Line... Apr 20 19:06:48.352012 systemd[1]: Finished systemd-network-generator.service - Generate Network Units from Kernel Command Line. Apr 20 19:06:48.390000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:48.479099 kernel: kauditd_printk_skb: 1 callbacks suppressed Apr 20 19:06:48.479136 kernel: audit: type=1130 audit(1776712008.390:13): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:48.497308 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 20 19:06:50.255560 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 20 19:06:50.267000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:50.277758 kernel: audit: type=1130 audit(1776712010.267:14): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:50.315305 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 20 19:06:50.358965 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 20 19:06:50.732668 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 20 19:06:50.739000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:50.756525 kernel: audit: type=1130 audit(1776712010.739:15): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:50.758000 audit: BPF prog-id=6 op=LOAD Apr 20 19:06:50.759000 audit: BPF prog-id=7 op=LOAD Apr 20 19:06:50.760028 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 20 19:06:50.771209 kernel: audit: type=1334 audit(1776712010.758:16): prog-id=6 op=LOAD Apr 20 19:06:50.771242 kernel: audit: type=1334 audit(1776712010.759:17): prog-id=7 op=LOAD Apr 20 19:06:51.006226 systemd-udevd[591]: Using default interface naming scheme 'v258'. Apr 20 19:06:51.379204 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 20 19:06:51.381000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:51.391049 kernel: audit: type=1130 audit(1776712011.381:18): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:51.385641 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 20 19:06:51.731718 dracut-pre-trigger[655]: rd.md=0: removing MD RAID activation Apr 20 19:06:51.927992 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 20 19:06:51.927000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:51.939000 audit: BPF prog-id=8 op=LOAD Apr 20 19:06:51.943588 kernel: audit: type=1130 audit(1776712011.927:19): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:51.943653 kernel: audit: type=1334 audit(1776712011.939:20): prog-id=8 op=LOAD Apr 20 19:06:51.946800 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 20 19:06:52.134110 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 20 19:06:52.140000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:52.148970 kernel: audit: type=1130 audit(1776712012.140:21): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:52.147792 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 20 19:06:52.362256 systemd-networkd[726]: lo: Link UP Apr 20 19:06:52.380190 systemd-networkd[726]: lo: Gained carrier Apr 20 19:06:52.393967 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 20 19:06:52.397000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:52.401840 systemd[1]: Reached target network.target - Network. Apr 20 19:06:52.409338 kernel: audit: type=1130 audit(1776712012.397:22): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:55.260312 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 20 19:06:55.281000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:55.330729 kernel: audit: type=1130 audit(1776712015.281:23): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:55.358279 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 20 19:06:56.750710 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Apr 20 19:06:56.887980 kernel: cryptd: max_cpu_qlen set to 1000 Apr 20 19:06:56.893926 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Apr 20 19:06:57.011714 kernel: AES CTR mode by8 optimization enabled Apr 20 19:06:57.028486 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Apr 20 19:06:57.051250 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Apr 20 19:06:57.167602 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Apr 20 19:06:57.206073 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 20 19:06:57.246213 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 20 19:06:57.250137 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 20 19:06:57.257000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.278341 kernel: audit: type=1131 audit(1776712017.257:24): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.257725 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 20 19:06:57.290357 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 20 19:06:57.341835 systemd-networkd[726]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Apr 20 19:06:57.341860 systemd-networkd[726]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 20 19:06:57.348257 systemd-networkd[726]: eth0: Link UP Apr 20 19:06:57.349735 systemd-networkd[726]: eth0: Gained carrier Apr 20 19:06:57.349754 systemd-networkd[726]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Apr 20 19:06:57.393916 disk-uuid[867]: Primary Header is updated. Apr 20 19:06:57.393916 disk-uuid[867]: Secondary Entries is updated. Apr 20 19:06:57.393916 disk-uuid[867]: Secondary Header is updated. Apr 20 19:06:57.420470 systemd-networkd[726]: eth0: DHCPv4 address 10.0.0.18/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 20 19:06:57.444447 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 20 19:06:57.451000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.487749 kernel: audit: type=1130 audit(1776712017.451:25): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.506303 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 20 19:06:57.510513 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 20 19:06:57.515111 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 20 19:06:57.545919 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 20 19:06:57.633728 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 20 19:06:57.643000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.651657 kernel: audit: type=1130 audit(1776712017.643:26): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.853619 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 20 19:06:57.866000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:57.886123 kernel: audit: type=1130 audit(1776712017.866:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:58.662269 disk-uuid[870]: Warning: The kernel is still using the old partition table. Apr 20 19:06:58.662269 disk-uuid[870]: The new table will be used at the next reboot or after you Apr 20 19:06:58.662269 disk-uuid[870]: run partprobe(8) or kpartx(8) Apr 20 19:06:58.662269 disk-uuid[870]: The operation has completed successfully. Apr 20 19:06:58.918703 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 20 19:06:58.918891 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 20 19:06:58.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:58.931152 systemd-networkd[726]: eth0: Gained IPv6LL Apr 20 19:06:58.940000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:58.954531 kernel: audit: type=1130 audit(1776712018.939:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:58.945269 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 20 19:06:58.972658 kernel: audit: type=1131 audit(1776712018.940:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:59.339720 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (898) Apr 20 19:06:59.353357 kernel: BTRFS info (device vda6): first mount of filesystem 17906e87-85d1-46f5-980e-3e85555360cf Apr 20 19:06:59.358404 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 20 19:06:59.385743 kernel: BTRFS info (device vda6): turning on async discard Apr 20 19:06:59.387083 kernel: BTRFS info (device vda6): enabling free space tree Apr 20 19:06:59.618441 kernel: BTRFS info (device vda6): last unmount of filesystem 17906e87-85d1-46f5-980e-3e85555360cf Apr 20 19:06:59.622891 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 20 19:06:59.635000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:59.645125 kernel: audit: type=1130 audit(1776712019.635:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:06:59.713970 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 20 19:07:01.705963 ignition[917]: Ignition 2.24.0 Apr 20 19:07:01.705996 ignition[917]: Stage: fetch-offline Apr 20 19:07:01.710034 ignition[917]: no configs at "/usr/lib/ignition/base.d" Apr 20 19:07:01.710321 ignition[917]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 20 19:07:01.710724 ignition[917]: parsed url from cmdline: "" Apr 20 19:07:01.710728 ignition[917]: no config URL provided Apr 20 19:07:01.710948 ignition[917]: reading system config file "/usr/lib/ignition/user.ign" Apr 20 19:07:01.710960 ignition[917]: no config at "/usr/lib/ignition/user.ign" Apr 20 19:07:01.711035 ignition[917]: op(1): [started] loading QEMU firmware config module Apr 20 19:07:01.711039 ignition[917]: op(1): executing: "modprobe" "qemu_fw_cfg" Apr 20 19:07:01.834082 ignition[917]: op(1): [finished] loading QEMU firmware config module Apr 20 19:07:02.194966 ignition[917]: parsing config with SHA512: e8a2d1bc98d12939045a642e86ba3068f14e0f0d0b7a0fa4655125345d40ac6004c1643e401a937918ec9ee094c2f6dc51fbde3fb7aa722a1a6cd9411dd3ed91 Apr 20 19:07:02.328648 unknown[917]: fetched base config from "system" Apr 20 19:07:02.328662 unknown[917]: fetched user config from "qemu" Apr 20 19:07:02.333883 ignition[917]: fetch-offline: fetch-offline passed Apr 20 19:07:02.334043 ignition[917]: Ignition finished successfully Apr 20 19:07:02.351752 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 20 19:07:02.362000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:02.487234 kernel: audit: type=1130 audit(1776712022.362:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:02.488301 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Apr 20 19:07:02.511778 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 20 19:07:03.534917 ignition[927]: Ignition 2.24.0 Apr 20 19:07:03.545269 ignition[927]: Stage: kargs Apr 20 19:07:03.552661 ignition[927]: no configs at "/usr/lib/ignition/base.d" Apr 20 19:07:03.552676 ignition[927]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 20 19:07:03.622505 ignition[927]: kargs: kargs passed Apr 20 19:07:03.623845 ignition[927]: Ignition finished successfully Apr 20 19:07:03.633989 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 20 19:07:03.643000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:03.654557 kernel: audit: type=1130 audit(1776712023.643:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:03.776768 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 20 19:07:04.530798 ignition[935]: Ignition 2.24.0 Apr 20 19:07:04.530882 ignition[935]: Stage: disks Apr 20 19:07:04.531863 ignition[935]: no configs at "/usr/lib/ignition/base.d" Apr 20 19:07:04.531875 ignition[935]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 20 19:07:04.585946 ignition[935]: disks: disks passed Apr 20 19:07:04.629843 ignition[935]: Ignition finished successfully Apr 20 19:07:04.657447 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 20 19:07:04.663000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:04.677764 kernel: audit: type=1130 audit(1776712024.663:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:04.678274 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 20 19:07:04.686553 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 20 19:07:04.702871 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 20 19:07:04.715617 systemd[1]: Reached target sysinit.target - System Initialization. Apr 20 19:07:04.804733 systemd[1]: Reached target basic.target - Basic System. Apr 20 19:07:04.967153 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 20 19:07:05.960420 systemd-fsck[946]: ROOT: clean, 15/456736 files, 38230/456704 blocks Apr 20 19:07:06.042512 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 20 19:07:06.050000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:06.064014 kernel: audit: type=1130 audit(1776712026.050:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:06.076413 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 20 19:07:07.232928 kernel: EXT4-fs (vda9): mounted filesystem 2bdffc2e-451a-418b-b04b-9e3cd9229e7e r/w with ordered data mode. Quota mode: none. Apr 20 19:07:07.279177 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 20 19:07:07.288455 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 20 19:07:07.408486 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 20 19:07:07.460770 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 20 19:07:07.470284 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Apr 20 19:07:07.472096 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 20 19:07:07.472191 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 20 19:07:07.585976 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 20 19:07:07.636948 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (957) Apr 20 19:07:07.644633 kernel: BTRFS info (device vda6): first mount of filesystem 17906e87-85d1-46f5-980e-3e85555360cf Apr 20 19:07:07.650997 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 20 19:07:07.651258 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 20 19:07:07.768664 kernel: BTRFS info (device vda6): turning on async discard Apr 20 19:07:07.768931 kernel: BTRFS info (device vda6): enabling free space tree Apr 20 19:07:07.848916 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 20 19:07:16.379290 kernel: loop1: detected capacity change from 0 to 43472 Apr 20 19:07:16.448761 kernel: loop1: p1 p2 p3 Apr 20 19:07:17.142863 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:07:17.143264 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:07:17.143298 kernel: device-mapper: table: 253:1: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:07:17.147259 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:07:17.152759 systemd-confext[1048]: device-mapper: reload ioctl on 036bab43330e5a16b58b2997d79b59667c299046b83a0d438261a470d6586a8f-verity (253:1) failed: Invalid argument Apr 20 19:07:17.265941 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:07:23.083882 kernel: erofs: (device dm-1): mounted with root inode @ nid 40. Apr 20 19:07:24.926605 kernel: loop2: detected capacity change from 0 to 43472 Apr 20 19:07:24.948880 kernel: loop2: p1 p2 p3 Apr 20 19:07:25.452566 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:07:25.464935 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:07:25.461452 (sd-merge)[1062]: device-mapper: reload ioctl on 036bab43330e5a16b58b2997d79b59667c299046b83a0d438261a470d6586a8f-verity (253:1) failed: Invalid argument Apr 20 19:07:25.486684 kernel: device-mapper: table: 253:1: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:07:25.486973 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:07:25.543719 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:07:38.265292 kernel: erofs: (device dm-1): mounted with root inode @ nid 40. Apr 20 19:07:38.597009 (sd-merge)[1062]: Using extensions '00-flatcar-default.raw'. Apr 20 19:07:39.375414 (sd-merge)[1062]: Merged extensions into '/sysroot/etc'. Apr 20 19:07:40.356857 initrd-setup-root[1069]: /etc 00-flatcar-default Mon 2026-04-20 19:06:39 UTC Apr 20 19:07:40.484245 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 20 19:07:40.513000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:40.544297 kernel: audit: type=1130 audit(1776712060.513:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:40.723660 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 20 19:07:40.771255 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 20 19:07:41.212896 kernel: BTRFS info (device vda6): last unmount of filesystem 17906e87-85d1-46f5-980e-3e85555360cf Apr 20 19:07:41.249077 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 20 19:07:41.417241 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 20 19:07:41.439000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:41.451141 kernel: audit: type=1130 audit(1776712061.439:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:45.204179 ignition[1079]: INFO : Ignition 2.24.0 Apr 20 19:07:45.204179 ignition[1079]: INFO : Stage: mount Apr 20 19:07:45.219016 ignition[1079]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 20 19:07:45.219016 ignition[1079]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 20 19:07:45.233744 ignition[1079]: INFO : mount: mount passed Apr 20 19:07:45.233744 ignition[1079]: INFO : Ignition finished successfully Apr 20 19:07:45.344234 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 20 19:07:45.388000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:45.420937 kernel: audit: type=1130 audit(1776712065.388:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:07:45.598503 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 20 19:07:47.237344 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 20 19:07:47.647728 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1094) Apr 20 19:07:47.654447 kernel: BTRFS info (device vda6): first mount of filesystem 17906e87-85d1-46f5-980e-3e85555360cf Apr 20 19:07:47.655939 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Apr 20 19:07:47.713855 kernel: BTRFS info (device vda6): turning on async discard Apr 20 19:07:47.717980 kernel: BTRFS info (device vda6): enabling free space tree Apr 20 19:07:47.735804 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 20 19:07:49.098143 ignition[1112]: INFO : Ignition 2.24.0 Apr 20 19:07:49.117278 ignition[1112]: INFO : Stage: files Apr 20 19:07:49.117278 ignition[1112]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 20 19:07:49.117278 ignition[1112]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 20 19:07:49.153853 ignition[1112]: DEBUG : files: compiled without relabeling support, skipping Apr 20 19:07:49.316211 ignition[1112]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 20 19:07:49.316211 ignition[1112]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 20 19:07:49.359113 ignition[1112]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 20 19:07:49.384643 ignition[1112]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 20 19:07:49.463237 unknown[1112]: wrote ssh authorized keys file for user: core Apr 20 19:07:49.505789 ignition[1112]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 20 19:07:49.517443 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 20 19:07:49.529264 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Apr 20 19:07:50.330922 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 20 19:07:55.078849 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 20 19:07:55.129707 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 20 19:07:55.151625 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Apr 20 19:07:56.312501 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Apr 20 19:08:05.625963 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 20 19:08:05.633505 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Apr 20 19:08:05.641082 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Apr 20 19:08:05.641082 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 20 19:08:05.654250 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 20 19:08:05.787299 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 20 19:08:05.787299 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 20 19:08:05.787299 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-x86-64.raw: attempt #1 Apr 20 19:08:06.218868 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Apr 20 19:08:26.144298 ignition[1112]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Apr 20 19:08:26.144298 ignition[1112]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Apr 20 19:08:26.167225 ignition[1112]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 20 19:08:26.177288 ignition[1112]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 20 19:08:26.177288 ignition[1112]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Apr 20 19:08:26.177288 ignition[1112]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Apr 20 19:08:26.177288 ignition[1112]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 20 19:08:26.230286 ignition[1112]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 20 19:08:26.230286 ignition[1112]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Apr 20 19:08:26.230286 ignition[1112]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Apr 20 19:08:27.751472 ignition[1112]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Apr 20 19:08:28.056955 ignition[1112]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Apr 20 19:08:28.077335 ignition[1112]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Apr 20 19:08:28.077335 ignition[1112]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Apr 20 19:08:28.077335 ignition[1112]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Apr 20 19:08:28.253254 ignition[1112]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 20 19:08:28.355851 ignition[1112]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 20 19:08:28.355851 ignition[1112]: INFO : files: files passed Apr 20 19:08:28.355851 ignition[1112]: INFO : Ignition finished successfully Apr 20 19:08:28.564484 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 20 19:08:28.660000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:28.675671 kernel: audit: type=1130 audit(1776712108.660:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:28.763804 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 20 19:08:28.882109 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 20 19:08:29.063192 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 20 19:08:29.063514 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 20 19:08:29.073000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:29.073000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:29.100313 kernel: audit: type=1130 audit(1776712109.073:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:29.100452 kernel: audit: type=1131 audit(1776712109.073:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:29.274167 initrd-setup-root-after-ignition[1143]: grep: /sysroot/oem/oem-release: No such file or directory Apr 20 19:08:29.347955 initrd-setup-root-after-ignition[1149]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 20 19:08:29.373363 initrd-setup-root-after-ignition[1145]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 20 19:08:29.373363 initrd-setup-root-after-ignition[1145]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 20 19:08:30.006166 kernel: loop3: detected capacity change from 0 to 43472 Apr 20 19:08:30.029625 kernel: loop3: p1 p2 p3 Apr 20 19:08:30.728553 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:30.735699 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:08:30.737221 kernel: device-mapper: table: 253:2: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:08:30.739545 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:08:30.739902 systemd-confext[1151]: device-mapper: reload ioctl on loop3p1-verity (253:2) failed: Invalid argument Apr 20 19:08:30.908313 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:33.350598 kernel: erofs: (device dm-2): mounted with root inode @ nid 40. Apr 20 19:08:33.858542 kernel: loop4: detected capacity change from 0 to 43472 Apr 20 19:08:33.886261 kernel: loop4: p1 p2 p3 Apr 20 19:08:34.353130 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:34.353604 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:08:34.353626 kernel: device-mapper: table: 253:2: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:08:34.358955 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:08:34.361295 (sd-merge)[1163]: device-mapper: reload ioctl on loop4p1-verity (253:2) failed: Invalid argument Apr 20 19:08:34.392869 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:36.351695 kernel: erofs: (device dm-2): mounted with root inode @ nid 40. Apr 20 19:08:36.374102 (sd-merge)[1163]: Skipping extension refresh because no change was found, use --always-refresh=yes to always do a refresh. Apr 20 19:08:36.480834 kernel: device-mapper: ioctl: remove_all left 2 open device(s) Apr 20 19:08:36.684168 kernel: loop4: detected capacity change from 0 to 178200 Apr 20 19:08:36.700161 kernel: loop4: p1 p2 p3 Apr 20 19:08:37.464648 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:37.465186 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:08:37.469056 kernel: device-mapper: table: 253:2: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:08:37.475142 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:08:37.479031 systemd-sysext[1171]: device-mapper: reload ioctl on 47e3a0d62726bde98fcb471f946aa0f0e9f97280e4f7267ec40f142aba643eb6-verity (253:2) failed: Invalid argument Apr 20 19:08:37.560603 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:38.964318 kernel: erofs: (device dm-2): mounted with root inode @ nid 39. Apr 20 19:08:39.364446 kernel: loop5: detected capacity change from 0 to 378016 Apr 20 19:08:39.383267 kernel: loop5: p1 p2 p3 Apr 20 19:08:39.768262 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:39.769822 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:08:39.769910 kernel: device-mapper: table: 253:2: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:08:39.770169 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:08:39.774298 systemd-sysext[1171]: device-mapper: reload ioctl on 5f63b01eb609e19b7df6b1f3554b098a8644903507171258f91f339ee69140b0-verity (253:2) failed: Invalid argument Apr 20 19:08:39.806962 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:41.362281 kernel: erofs: (device dm-2): mounted with root inode @ nid 39. Apr 20 19:08:41.684921 kernel: loop6: detected capacity change from 0 to 228704 Apr 20 19:08:42.141474 kernel: loop7: detected capacity change from 0 to 178200 Apr 20 19:08:42.154602 kernel: loop7: p1 p2 p3 Apr 20 19:08:42.631996 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:42.638402 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:08:42.638554 kernel: device-mapper: table: 253:2: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:08:42.638568 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:08:42.641655 (sd-merge)[1188]: device-mapper: reload ioctl on 47e3a0d62726bde98fcb471f946aa0f0e9f97280e4f7267ec40f142aba643eb6-verity (253:2) failed: Invalid argument Apr 20 19:08:42.831878 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:43.850079 kernel: erofs: (device dm-2): mounted with root inode @ nid 39. Apr 20 19:08:44.028167 kernel: loop1: detected capacity change from 0 to 378016 Apr 20 19:08:44.038058 kernel: loop1: p1 p2 p3 Apr 20 19:08:44.418858 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:44.419279 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:08:44.421990 kernel: device-mapper: table: 253:3: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:08:44.424120 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:08:44.426260 (sd-merge)[1188]: device-mapper: reload ioctl on 5f63b01eb609e19b7df6b1f3554b098a8644903507171258f91f339ee69140b0-verity (253:3) failed: Invalid argument Apr 20 19:08:44.445667 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:08:45.876813 kernel: erofs: (device dm-3): mounted with root inode @ nid 39. Apr 20 19:08:46.016165 kernel: loop3: detected capacity change from 0 to 228704 Apr 20 19:08:46.292314 (sd-merge)[1188]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes-v1.33.8-x86-64.raw'. Apr 20 19:08:46.380519 (sd-merge)[1188]: Merged extensions into '/sysroot/usr'. Apr 20 19:08:46.389730 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 20 19:08:46.469000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:46.479245 kernel: audit: type=1130 audit(1776712126.469:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:46.481251 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 20 19:08:46.501009 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 20 19:08:47.364103 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 20 19:08:47.364287 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 20 19:08:47.413000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:47.420000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:47.430198 kernel: audit: type=1130 audit(1776712127.413:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:47.422347 systemd[1]: initrd-parse-etc.service: Triggering OnSuccess= dependencies. Apr 20 19:08:47.433815 kernel: audit: type=1131 audit(1776712127.420:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:47.434331 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 20 19:08:47.434661 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 20 19:08:47.441613 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 20 19:08:47.465521 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 20 19:08:49.090741 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 20 19:08:49.145000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:49.164452 kernel: audit: type=1130 audit(1776712129.145:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:49.273615 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 20 19:08:50.423346 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 20 19:08:50.446672 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 20 19:08:50.579280 systemd[1]: Stopped target timers.target - Timer Units. Apr 20 19:08:50.615069 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 20 19:08:50.617016 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 20 19:08:50.634000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:50.638175 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 20 19:08:50.644488 kernel: audit: type=1131 audit(1776712130.634:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:50.660222 systemd[1]: Stopped target basic.target - Basic System. Apr 20 19:08:50.695757 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 20 19:08:50.789773 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 20 19:08:50.813255 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 20 19:08:50.947197 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Apr 20 19:08:50.959592 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 20 19:08:50.988076 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 20 19:08:50.999781 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 20 19:08:51.009712 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 20 19:08:51.034278 systemd[1]: Stopped target swap.target - Swaps. Apr 20 19:08:51.074207 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 20 19:08:51.156947 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 20 19:08:51.173000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.177969 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 20 19:08:51.181823 kernel: audit: type=1131 audit(1776712131.173:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.203340 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 20 19:08:51.230238 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 20 19:08:51.248191 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 20 19:08:51.270053 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 20 19:08:51.347867 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 20 19:08:51.364000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.371962 kernel: audit: type=1131 audit(1776712131.364:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.376171 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 20 19:08:51.397532 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 20 19:08:51.415000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.425157 systemd[1]: ignition-fetch-offline.service: Consumed 1.455s CPU time. Apr 20 19:08:51.440347 systemd[1]: Stopped target paths.target - Path Units. Apr 20 19:08:51.455719 kernel: audit: type=1131 audit(1776712131.415:48): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.455635 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 20 19:08:51.459013 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 20 19:08:51.484292 systemd[1]: Stopped target slices.target - Slice Units. Apr 20 19:08:51.604009 systemd[1]: Stopped target sockets.target - Socket Units. Apr 20 19:08:51.620015 systemd[1]: iscsid.socket: Deactivated successfully. Apr 20 19:08:51.642876 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 20 19:08:51.687324 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 20 19:08:51.748202 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 20 19:08:51.795202 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Apr 20 19:08:51.801104 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Apr 20 19:08:51.814912 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 20 19:08:51.843944 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 20 19:08:51.868000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.871919 systemd[1]: initrd-setup-root-after-ignition.service: Consumed 3.443s CPU time. Apr 20 19:08:51.933815 kernel: audit: type=1131 audit(1776712131.868:49): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.949163 systemd[1]: ignition-files.service: Deactivated successfully. Apr 20 19:08:51.964415 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 20 19:08:51.964000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.974169 kernel: audit: type=1131 audit(1776712131.964:50): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:51.965087 systemd[1]: ignition-files.service: Consumed 32.594s CPU time. Apr 20 19:08:51.987288 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 20 19:08:52.009275 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 20 19:08:52.012741 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 20 19:08:52.019000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.026012 kernel: audit: type=1131 audit(1776712132.019:51): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.013065 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 20 19:08:52.026000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.048931 kernel: audit: type=1131 audit(1776712132.026:52): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.019513 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 20 19:08:52.019629 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 20 19:08:52.058000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.073082 kernel: audit: type=1131 audit(1776712132.058:53): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.026490 systemd[1]: systemd-udev-trigger.service: Consumed 1.633s CPU time. Apr 20 19:08:52.026571 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 20 19:08:52.026712 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 20 19:08:52.235028 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 20 19:08:52.235250 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 20 19:08:52.454000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.454000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.475146 kernel: audit: type=1130 audit(1776712132.454:54): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.475180 kernel: audit: type=1131 audit(1776712132.454:55): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.487477 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 20 19:08:52.508251 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 20 19:08:52.508674 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 20 19:08:52.512000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.521489 kernel: audit: type=1131 audit(1776712132.512:56): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:52.885211 ignition[1217]: INFO : Ignition 2.24.0 Apr 20 19:08:52.885211 ignition[1217]: INFO : Stage: umount Apr 20 19:08:52.918341 ignition[1217]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 20 19:08:52.918341 ignition[1217]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 20 19:08:52.918341 ignition[1217]: INFO : umount: umount passed Apr 20 19:08:52.942614 ignition[1217]: INFO : Ignition finished successfully Apr 20 19:08:53.045826 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 20 19:08:53.095228 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 20 19:08:53.129000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.136493 kernel: audit: type=1131 audit(1776712133.129:57): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.144222 systemd[1]: Stopped target network.target - Network. Apr 20 19:08:53.181935 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 20 19:08:53.183242 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 20 19:08:53.207000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.208631 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 20 19:08:53.208854 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 20 19:08:53.223000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.246923 kernel: audit: type=1131 audit(1776712133.207:58): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.298990 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 20 19:08:53.301682 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 20 19:08:53.308000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.312494 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 20 19:08:53.319000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.312892 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 20 19:08:53.323602 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 20 19:08:53.333000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.323839 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 20 19:08:53.333978 systemd[1]: initrd-setup-root.service: Consumed 6.354s CPU time. Apr 20 19:08:53.334779 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 20 19:08:53.372712 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 20 19:08:53.459254 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 20 19:08:53.478522 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 20 19:08:53.478000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.551763 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 20 19:08:53.552457 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 20 19:08:53.566000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.606000 audit: BPF prog-id=5 op=UNLOAD Apr 20 19:08:53.615000 audit: BPF prog-id=8 op=UNLOAD Apr 20 19:08:53.620318 systemd[1]: Stopped target network-pre.target - Preparation for Network. Apr 20 19:08:53.620783 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 20 19:08:53.620841 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 20 19:08:53.725815 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 20 19:08:53.726452 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 20 19:08:53.725000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.726510 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 20 19:08:53.747000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.726975 systemd[1]: parse-ip-for-networkd.service: Consumed 1.064s CPU time. Apr 20 19:08:53.755000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.727013 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 20 19:08:53.727036 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 20 19:08:53.749778 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 20 19:08:53.749911 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 20 19:08:53.760318 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 20 19:08:53.863003 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 20 19:08:53.868786 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 20 19:08:53.880000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.882281 systemd[1]: systemd-udevd.service: Consumed 21.362s CPU time. Apr 20 19:08:53.888781 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 20 19:08:53.891830 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 20 19:08:53.911927 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 20 19:08:53.912623 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 20 19:08:53.917000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.919833 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 20 19:08:53.921000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.925000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.920037 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 20 19:08:53.935000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.954000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.922168 systemd[1]: dracut-cmdline.service: Consumed 4.559s CPU time. Apr 20 19:08:53.960000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.923495 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 20 19:08:53.982000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.924153 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 20 19:08:53.987000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:53.931845 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 20 19:08:53.934570 systemd[1]: systemd-network-generator.service: Deactivated successfully. Apr 20 19:08:53.934901 systemd[1]: Stopped systemd-network-generator.service - Generate Network Units from Kernel Command Line. Apr 20 19:08:53.935680 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 20 19:08:53.935739 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 20 19:08:53.957559 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Apr 20 19:08:53.959356 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 20 19:08:53.965984 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 20 19:08:53.968082 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 20 19:08:53.984098 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 20 19:08:53.984181 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 20 19:08:54.112225 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 20 19:08:54.112431 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 20 19:08:54.122000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:54.123000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:54.128051 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 20 19:08:54.131116 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 20 19:08:54.141000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:08:54.144235 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 20 19:08:54.262517 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 20 19:08:54.545313 systemd[1]: Switching root. Apr 20 19:08:54.869431 systemd-journald[320]: Received SIGTERM from PID 1 (systemd). Apr 20 19:08:54.869699 systemd-journald[320]: Journal stopped Apr 20 19:10:16.973513 kernel: SELinux: policy capability network_peer_controls=1 Apr 20 19:10:16.973789 kernel: SELinux: policy capability open_perms=1 Apr 20 19:10:16.973810 kernel: SELinux: policy capability extended_socket_class=1 Apr 20 19:10:16.973824 kernel: SELinux: policy capability always_check_network=0 Apr 20 19:10:16.973842 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 20 19:10:16.973915 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 20 19:10:16.973929 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 20 19:10:16.973942 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 20 19:10:16.973960 kernel: SELinux: policy capability userspace_initial_context=0 Apr 20 19:10:16.973977 systemd[1]: Successfully loaded SELinux policy in 389.975ms. Apr 20 19:10:16.974017 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 315.640ms. Apr 20 19:10:16.974038 systemd[1]: systemd 258.3 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 20 19:10:16.974073 systemd[1]: Detected virtualization kvm. Apr 20 19:10:16.974089 systemd[1]: Detected architecture x86-64. Apr 20 19:10:16.974104 systemd[1]: Detected first boot. Apr 20 19:10:16.974126 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Apr 20 19:10:16.974140 kernel: kauditd_printk_skb: 25 callbacks suppressed Apr 20 19:10:16.974154 kernel: audit: type=1334 audit(1776712137.633:84): prog-id=9 op=LOAD Apr 20 19:10:16.974186 kernel: audit: type=1334 audit(1776712137.634:85): prog-id=9 op=UNLOAD Apr 20 19:10:16.974201 zram_generator::config[1265]: No configuration found. Apr 20 19:10:16.974234 kernel: Guest personality initialized and is inactive Apr 20 19:10:16.974262 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Apr 20 19:10:16.974275 kernel: Initialized host personality Apr 20 19:10:16.974288 kernel: NET: Registered PF_VSOCK protocol family Apr 20 19:10:16.974300 systemd-ssh-generator[1261]: Failed to query local AF_VSOCK CID: Cannot assign requested address Apr 20 19:10:16.974340 (sd-exec-[1246]: /usr/lib/systemd/system-generators/systemd-ssh-generator failed with exit status 1. Apr 20 19:10:16.974357 systemd[1]: Applying preset policy. Apr 20 19:10:16.998201 systemd[1]: Created symlink '/etc/systemd/system/multi-user.target.wants/prepare-helm.service' → '/etc/systemd/system/prepare-helm.service'. Apr 20 19:10:17.002578 systemd[1]: Created symlink '/etc/systemd/system/timers.target.wants/google-oslogin-cache.timer' → '/usr/lib/systemd/system/google-oslogin-cache.timer'. Apr 20 19:10:17.016644 systemd[1]: Populated /etc with preset unit settings. Apr 20 19:10:17.022939 systemd[1]: /usr/lib/systemd/system/update-engine.service:10: Support for option BlockIOWeight= has been removed and it is ignored Apr 20 19:10:17.023063 kernel: audit: type=1334 audit(1776712208.713:86): prog-id=10 op=LOAD Apr 20 19:10:17.023085 kernel: audit: type=1334 audit(1776712208.821:87): prog-id=2 op=UNLOAD Apr 20 19:10:17.023101 kernel: audit: type=1334 audit(1776712208.887:88): prog-id=11 op=LOAD Apr 20 19:10:17.023121 kernel: audit: type=1334 audit(1776712208.966:89): prog-id=12 op=LOAD Apr 20 19:10:17.023133 kernel: audit: type=1334 audit(1776712208.972:90): prog-id=3 op=UNLOAD Apr 20 19:10:17.023146 kernel: audit: type=1334 audit(1776712208.979:91): prog-id=4 op=UNLOAD Apr 20 19:10:17.023159 kernel: audit: type=1334 audit(1776712209.320:92): prog-id=13 op=LOAD Apr 20 19:10:17.023193 kernel: audit: type=1334 audit(1776712209.324:93): prog-id=10 op=UNLOAD Apr 20 19:10:17.023210 kernel: audit: type=1334 audit(1776712209.333:94): prog-id=14 op=LOAD Apr 20 19:10:17.023223 kernel: audit: type=1334 audit(1776712209.352:95): prog-id=15 op=LOAD Apr 20 19:10:17.023236 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 20 19:10:17.023250 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 20 19:10:17.023264 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 20 19:10:17.023298 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 20 19:10:17.023314 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 20 19:10:17.023330 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 20 19:10:17.023344 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 20 19:10:17.023358 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 20 19:10:17.023415 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 20 19:10:17.023450 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 20 19:10:17.023466 systemd[1]: Created slice user.slice - User and Session Slice. Apr 20 19:10:17.023482 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 20 19:10:17.023497 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 20 19:10:17.023513 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 20 19:10:17.023539 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 20 19:10:17.023553 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 20 19:10:17.023581 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 20 19:10:17.023596 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Apr 20 19:10:17.023611 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 20 19:10:17.023627 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 20 19:10:17.023641 systemd[1]: Reached target imports.target - Image Downloads. Apr 20 19:10:17.023667 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 20 19:10:17.023681 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 20 19:10:17.023695 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 20 19:10:17.023726 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 20 19:10:17.023741 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 20 19:10:17.023757 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 20 19:10:17.023772 systemd[1]: Reached target remote-integritysetup.target - Remote Integrity Protected Volumes. Apr 20 19:10:17.023788 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Apr 20 19:10:17.023803 systemd[1]: Reached target slices.target - Slice Units. Apr 20 19:10:17.023817 systemd[1]: Reached target swap.target - Swaps. Apr 20 19:10:17.023850 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 20 19:10:17.023866 systemd[1]: Listening on systemd-ask-password.socket - Query the User Interactively for a Password. Apr 20 19:10:17.023881 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 20 19:10:17.023913 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Apr 20 19:10:17.023928 systemd[1]: Listening on systemd-factory-reset.socket - Factory Reset Management. Apr 20 19:10:17.023941 kernel: kauditd_printk_skb: 12 callbacks suppressed Apr 20 19:10:17.023956 kernel: audit: type=1335 audit(1776712214.320:108): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Apr 20 19:10:17.023991 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Apr 20 19:10:17.024007 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Apr 20 19:10:17.024020 systemd[1]: Listening on systemd-networkd-varlink.socket - Network Service Varlink Socket. Apr 20 19:10:17.024936 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 20 19:10:17.024976 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Apr 20 19:10:17.024991 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Apr 20 19:10:17.025005 systemd[1]: Listening on systemd-resolved-monitor.socket - Resolve Monitor Varlink Socket. Apr 20 19:10:17.025043 systemd[1]: Listening on systemd-resolved-varlink.socket - Resolve Service Varlink Socket. Apr 20 19:10:17.025058 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 20 19:10:17.025074 systemd[1]: Listening on systemd-udevd-varlink.socket - udev Varlink Socket. Apr 20 19:10:17.025089 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 20 19:10:17.025103 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 20 19:10:17.025117 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 20 19:10:17.025148 systemd[1]: Mounting media.mount - External Media Directory... Apr 20 19:10:17.025165 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 20 19:10:17.025180 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 20 19:10:17.025195 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 20 19:10:17.025209 systemd[1]: tmp.mount: x-systemd.graceful-option=usrquota specified, but option is not available, suppressing. Apr 20 19:10:17.025222 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 20 19:10:17.025237 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 20 19:10:17.025270 systemd[1]: Reached target machines.target - Virtual Machines and Containers. Apr 20 19:10:17.025285 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 20 19:10:17.025302 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 20 19:10:17.025321 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 20 19:10:17.025354 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 20 19:10:17.025400 systemd[1]: modprobe@dm_mod.service - Load Kernel Module dm_mod was skipped because of an unmet condition check (ConditionKernelModuleLoaded=!dm_mod). Apr 20 19:10:17.025419 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 20 19:10:17.025434 systemd[1]: modprobe@efi_pstore.service - Load Kernel Module efi_pstore was skipped because of an unmet condition check (ConditionKernelModuleLoaded=!efi_pstore). Apr 20 19:10:17.025449 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 20 19:10:17.025463 systemd[1]: modprobe@loop.service - Load Kernel Module loop was skipped because of an unmet condition check (ConditionKernelModuleLoaded=!loop). Apr 20 19:10:17.025499 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 20 19:10:17.025515 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 20 19:10:17.025543 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 20 19:10:17.025572 kernel: audit: type=1131 audit(1776712216.510:109): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.025603 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 20 19:10:17.025619 systemd[1]: Stopped systemd-fsck-usr.service. Apr 20 19:10:17.025635 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 20 19:10:17.025649 kernel: audit: type=1131 audit(1776712216.527:110): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.025663 kernel: audit: type=1334 audit(1776712216.552:111): prog-id=18 op=UNLOAD Apr 20 19:10:17.025676 kernel: audit: type=1334 audit(1776712216.553:112): prog-id=17 op=UNLOAD Apr 20 19:10:17.025708 kernel: audit: type=1334 audit(1776712216.560:113): prog-id=19 op=LOAD Apr 20 19:10:17.025723 kernel: audit: type=1334 audit(1776712216.560:114): prog-id=20 op=LOAD Apr 20 19:10:17.025736 kernel: audit: type=1334 audit(1776712216.564:115): prog-id=21 op=LOAD Apr 20 19:10:17.025751 kernel: fuse: init (API version 7.41) Apr 20 19:10:17.025784 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 20 19:10:17.025797 kernel: ACPI: bus type drm_connector registered Apr 20 19:10:17.025811 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 20 19:10:17.025825 systemd[1]: Starting systemd-network-generator.service - Generate Network Units from Kernel Command Line... Apr 20 19:10:17.025841 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 20 19:10:17.025869 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 20 19:10:17.025885 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Apr 20 19:10:17.025936 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 20 19:10:17.025952 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 20 19:10:17.025968 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 20 19:10:17.025984 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 20 19:10:17.025998 systemd[1]: Mounted media.mount - External Media Directory. Apr 20 19:10:17.026012 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 20 19:10:17.026026 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 20 19:10:17.026060 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 20 19:10:17.026118 systemd-journald[1331]: Collecting audit messages is enabled. Apr 20 19:10:17.026175 kernel: audit: type=1305 audit(1776712216.941:116): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Apr 20 19:10:17.026191 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 20 19:10:17.026207 kernel: audit: type=1300 audit(1776712216.941:116): arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7ffc2d3a6460 a2=4000 a3=0 items=0 ppid=1 pid=1331 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:10:17.026241 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 20 19:10:17.026255 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 20 19:10:17.026270 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 20 19:10:17.026284 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 20 19:10:17.026301 systemd-journald[1331]: Journal started Apr 20 19:10:17.026328 systemd-journald[1331]: Runtime Journal (/run/log/journal/8118aa160470456397a4c2dcfc0629e4) is 5.9M, max 47.8M, 41.8M free. Apr 20 19:10:14.320000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Apr 20 19:10:16.510000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:16.527000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:16.552000 audit: BPF prog-id=18 op=UNLOAD Apr 20 19:10:16.553000 audit: BPF prog-id=17 op=UNLOAD Apr 20 19:10:16.560000 audit: BPF prog-id=19 op=LOAD Apr 20 19:10:16.560000 audit: BPF prog-id=20 op=LOAD Apr 20 19:10:16.564000 audit: BPF prog-id=21 op=LOAD Apr 20 19:10:16.941000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Apr 20 19:10:16.941000 audit[1331]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7ffc2d3a6460 a2=4000 a3=0 items=0 ppid=1 pid=1331 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:10:16.941000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Apr 20 19:10:16.966000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.011000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.011000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:08.143580 systemd[1]: Queued start job for default target multi-user.target. Apr 20 19:10:10.105134 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Apr 20 19:10:17.039000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.051000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:10.121316 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 20 19:10:10.160277 systemd[1]: systemd-journald.service: Consumed 3.445s CPU time. Apr 20 19:10:17.086000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.087279 systemd[1]: Started systemd-journald.service - Journal Service. Apr 20 19:10:17.087733 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 20 19:10:17.117577 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 20 19:10:17.126000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.129000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.159647 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 20 19:10:17.169000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.184786 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 20 19:10:17.201000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.274250 systemd[1]: Finished systemd-network-generator.service - Generate Network Units from Kernel Command Line. Apr 20 19:10:17.288000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.309360 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 20 19:10:17.314000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.315959 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Apr 20 19:10:17.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:17.587474 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 20 19:10:17.670094 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Apr 20 19:10:17.733256 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 20 19:10:17.888785 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 20 19:10:17.927043 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 20 19:10:17.927225 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 20 19:10:17.944862 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Apr 20 19:10:17.961050 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 20 19:10:18.019052 systemd[1]: Starting systemd-confext.service - Merge System Configuration Images into /etc/... Apr 20 19:10:18.084474 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 20 19:10:18.104679 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 20 19:10:18.109780 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 20 19:10:18.162424 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 20 19:10:18.242629 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 20 19:10:18.275294 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 20 19:10:18.376222 systemd[1]: Starting systemd-userdb-load-credentials.service - Load JSON user/group Records from Credentials... Apr 20 19:10:18.377030 systemd-journald[1331]: Time spent on flushing to /var/log/journal/8118aa160470456397a4c2dcfc0629e4 is 353.974ms for 1317 entries. Apr 20 19:10:18.377030 systemd-journald[1331]: System Journal (/var/log/journal/8118aa160470456397a4c2dcfc0629e4) is 8M, max 163.5M, 155.5M free. Apr 20 19:10:18.562000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:18.584000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:18.604000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:18.426467 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 20 19:10:18.946226 systemd-journald[1331]: Received client request to flush runtime journal. Apr 20 19:10:18.448756 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 20 19:10:18.990953 kernel: loop4: detected capacity change from 0 to 43472 Apr 20 19:10:18.482722 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 20 19:10:18.991258 kernel: loop4: p1 p2 p3 Apr 20 19:10:18.578275 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 20 19:10:18.599038 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 20 19:10:18.878232 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 20 19:10:18.958768 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Apr 20 19:10:19.012000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:18.992824 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 20 19:10:18.992967 systemd-tmpfiles[1381]: ACLs are not supported, ignoring. Apr 20 19:10:18.992983 systemd-tmpfiles[1381]: ACLs are not supported, ignoring. Apr 20 19:10:19.074827 systemd[1]: Finished systemd-userdb-load-credentials.service - Load JSON user/group Records from Credentials. Apr 20 19:10:19.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdb-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:19.134071 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 20 19:10:19.158000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:19.363983 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 20 19:10:19.437000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:19.378675 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Apr 20 19:10:19.453943 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:10:19.439632 systemd-confext[1385]: device-mapper: reload ioctl on loop4p1-verity (253:4) failed: Invalid argument Apr 20 19:10:19.454353 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:10:19.462190 kernel: device-mapper: table: 253:4: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:10:19.462722 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:10:19.462751 kernel: kauditd_printk_skb: 20 callbacks suppressed Apr 20 19:10:19.462775 kernel: audit: type=1130 audit(1776712219.437:136): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:19.462800 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:10:19.470852 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 20 19:10:20.483628 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 20 19:10:20.488000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:20.493407 kernel: audit: type=1130 audit(1776712220.488:137): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:20.501000 audit: BPF prog-id=22 op=LOAD Apr 20 19:10:20.504565 kernel: audit: type=1334 audit(1776712220.501:138): prog-id=22 op=LOAD Apr 20 19:10:20.504000 audit: BPF prog-id=23 op=LOAD Apr 20 19:10:20.505915 kernel: audit: type=1334 audit(1776712220.504:139): prog-id=23 op=LOAD Apr 20 19:10:20.505000 audit: BPF prog-id=24 op=LOAD Apr 20 19:10:20.512787 kernel: audit: type=1334 audit(1776712220.505:140): prog-id=24 op=LOAD Apr 20 19:10:20.518314 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Apr 20 19:10:20.525000 audit: BPF prog-id=25 op=LOAD Apr 20 19:10:20.527697 kernel: audit: type=1334 audit(1776712220.525:141): prog-id=25 op=LOAD Apr 20 19:10:20.529892 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 20 19:10:20.551000 audit: BPF prog-id=26 op=LOAD Apr 20 19:10:20.556021 kernel: audit: type=1334 audit(1776712220.551:142): prog-id=26 op=LOAD Apr 20 19:10:20.556711 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 20 19:10:20.581095 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 20 19:10:20.687448 systemd[1]: Starting modprobe@tun.service - Load Kernel Module tun... Apr 20 19:10:20.725000 audit: BPF prog-id=27 op=LOAD Apr 20 19:10:20.730317 kernel: audit: type=1334 audit(1776712220.725:143): prog-id=27 op=LOAD Apr 20 19:10:20.725000 audit: BPF prog-id=28 op=LOAD Apr 20 19:10:20.725000 audit: BPF prog-id=29 op=LOAD Apr 20 19:10:20.731041 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 20 19:10:20.741177 kernel: audit: type=1334 audit(1776712220.725:144): prog-id=28 op=LOAD Apr 20 19:10:20.741206 kernel: audit: type=1334 audit(1776712220.725:145): prog-id=29 op=LOAD Apr 20 19:10:20.770028 systemd-tmpfiles[1409]: ACLs are not supported, ignoring. Apr 20 19:10:20.770045 systemd-tmpfiles[1409]: ACLs are not supported, ignoring. Apr 20 19:10:20.871243 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 20 19:10:20.877000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:20.914560 kernel: tun: Universal TUN/TAP device driver, 1.6 Apr 20 19:10:20.930789 systemd[1]: modprobe@tun.service: Deactivated successfully. Apr 20 19:10:20.932959 systemd[1]: Finished modprobe@tun.service - Load Kernel Module tun. Apr 20 19:10:20.934000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@tun comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:20.939000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@tun comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:20.942000 audit: BPF prog-id=30 op=LOAD Apr 20 19:10:20.942000 audit: BPF prog-id=31 op=LOAD Apr 20 19:10:20.942000 audit: BPF prog-id=32 op=LOAD Apr 20 19:10:20.954901 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Apr 20 19:10:21.168354 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 20 19:10:21.176000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:21.663206 systemd-nsresourced[1416]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Apr 20 19:10:21.744602 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Apr 20 19:10:21.748000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:22.210228 systemd-oomd[1406]: No swap; memory pressure usage will be degraded Apr 20 19:10:22.265720 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 20 19:10:22.280279 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Apr 20 19:10:22.279000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:22.339000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:22.407698 systemd[1]: Reached target time-set.target - System Time Set. Apr 20 19:10:23.229930 systemd-resolved[1407]: Positive Trust Anchors: Apr 20 19:10:23.235492 systemd-resolved[1407]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 20 19:10:23.235513 systemd-resolved[1407]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Apr 20 19:10:23.235554 systemd-resolved[1407]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 20 19:10:23.813786 systemd-resolved[1407]: Defaulting to hostname 'linux'. Apr 20 19:10:24.162137 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 20 19:10:24.225000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:24.237519 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 20 19:10:45.189359 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 20 19:10:45.309000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:45.330117 kernel: kauditd_printk_skb: 11 callbacks suppressed Apr 20 19:10:45.331467 kernel: audit: type=1130 audit(1776712245.309:157): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:45.379000 audit: BPF prog-id=7 op=UNLOAD Apr 20 19:10:45.380000 audit: BPF prog-id=6 op=UNLOAD Apr 20 19:10:45.389355 kernel: audit: type=1334 audit(1776712245.379:158): prog-id=7 op=UNLOAD Apr 20 19:10:45.394316 kernel: audit: type=1334 audit(1776712245.380:159): prog-id=6 op=UNLOAD Apr 20 19:10:45.448000 audit: BPF prog-id=33 op=LOAD Apr 20 19:10:45.451000 audit: BPF prog-id=34 op=LOAD Apr 20 19:10:45.454444 kernel: audit: type=1334 audit(1776712245.448:160): prog-id=33 op=LOAD Apr 20 19:10:45.454586 kernel: audit: type=1334 audit(1776712245.451:161): prog-id=34 op=LOAD Apr 20 19:10:45.559161 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 20 19:10:50.968522 systemd-udevd[1437]: Using default interface naming scheme 'v258'. Apr 20 19:10:59.756343 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 20 19:10:59.792000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:59.826448 kernel: audit: type=1130 audit(1776712259.792:162): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:10:59.859000 audit: BPF prog-id=35 op=LOAD Apr 20 19:10:59.868671 kernel: audit: type=1334 audit(1776712259.859:163): prog-id=35 op=LOAD Apr 20 19:11:00.004669 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 20 19:11:00.683193 systemd-networkd[1439]: lo: Link UP Apr 20 19:11:00.683262 systemd-networkd[1439]: lo: Gained carrier Apr 20 19:11:00.746125 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 20 19:11:00.753000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:00.761794 kernel: audit: type=1130 audit(1776712260.753:164): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:00.761801 systemd[1]: Reached target network.target - Network. Apr 20 19:11:00.782497 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Apr 20 19:11:00.791865 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 20 19:11:00.856000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-persistent-storage comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:00.850825 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Apr 20 19:11:00.864227 kernel: audit: type=1130 audit(1776712260.856:165): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-persistent-storage comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:00.955833 systemd-networkd[1439]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Apr 20 19:11:00.955841 systemd-networkd[1439]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 20 19:11:00.958974 systemd-networkd[1439]: eth0: Link UP Apr 20 19:11:00.963009 systemd-networkd[1439]: eth0: Gained carrier Apr 20 19:11:00.963876 systemd-networkd[1439]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Apr 20 19:11:00.968441 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Apr 20 19:11:00.993922 systemd-networkd[1439]: eth0: DHCPv4 address 10.0.0.18/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 20 19:11:00.997324 systemd-timesyncd[1408]: Network configuration changed, trying to establish connection. Apr 20 19:11:01.689182 systemd-resolved[1407]: Clock change detected. Flushing caches. Apr 20 19:11:01.689218 systemd-timesyncd[1408]: Contacted time server 10.0.0.1:123 (10.0.0.1). Apr 20 19:11:01.689341 systemd-timesyncd[1408]: Initial clock synchronization to Mon 2026-04-20 19:11:01.689067 UTC. Apr 20 19:11:01.731297 kernel: mousedev: PS/2 mouse device common for all mice Apr 20 19:11:01.851277 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Apr 20 19:11:01.861369 kernel: ACPI: button: Power Button [PWRF] Apr 20 19:11:01.962468 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Apr 20 19:11:02.043381 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 20 19:11:02.326487 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Apr 20 19:11:02.377724 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Apr 20 19:11:02.428751 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Apr 20 19:11:02.497953 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 20 19:11:02.506000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:02.516423 kernel: audit: type=1130 audit(1776712262.506:166): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:02.947026 systemd-networkd[1439]: eth0: Gained IPv6LL Apr 20 19:11:03.036763 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 20 19:11:03.040000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:03.049334 kernel: audit: type=1130 audit(1776712263.040:167): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:03.049756 systemd[1]: Reached target network-online.target - Network is Online. Apr 20 19:11:03.318723 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 20 19:11:03.544628 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 20 19:11:03.545160 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 20 19:11:03.551000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:03.572956 kernel: audit: type=1130 audit(1776712263.551:168): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:03.567000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:03.616159 kernel: audit: type=1131 audit(1776712263.567:169): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:03.753549 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 20 19:11:03.909288 kernel: erofs: (device dm-4): mounted with root inode @ nid 40. Apr 20 19:11:04.458142 kernel: loop4: detected capacity change from 0 to 43472 Apr 20 19:11:04.498721 kernel: loop4: p1 p2 p3 Apr 20 19:11:04.713200 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:04.713546 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:11:04.713705 kernel: device-mapper: table: 253:4: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:11:04.718220 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:11:04.718271 (sd-merge)[1504]: device-mapper: reload ioctl on loop4p1-verity (253:4) failed: Invalid argument Apr 20 19:11:04.718654 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 20 19:11:04.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:04.733685 kernel: audit: type=1130 audit(1776712264.724:170): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:04.735928 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:05.344634 kernel: erofs: (device dm-4): mounted with root inode @ nid 40. Apr 20 19:11:05.928092 (sd-merge)[1504]: Skipping extension refresh because no change was found, use --always-refresh=yes to always do a refresh. Apr 20 19:11:06.001357 kernel: device-mapper: ioctl: remove_all left 4 open device(s) Apr 20 19:11:06.151748 systemd[1]: Finished systemd-confext.service - Merge System Configuration Images into /etc/. Apr 20 19:11:06.160000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-confext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:06.170827 kernel: audit: type=1130 audit(1776712266.160:171): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-confext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:06.542645 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 20 19:11:08.106644 kernel: loop4: detected capacity change from 0 to 378016 Apr 20 19:11:08.157818 kernel: loop4: p1 p2 p3 Apr 20 19:11:08.745452 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:08.751186 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:11:08.751442 kernel: device-mapper: table: 253:4: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:11:08.751457 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:11:08.755191 systemd-sysext[1514]: device-mapper: reload ioctl on loop4p1-verity (253:4) failed: Invalid argument Apr 20 19:11:08.768290 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:10.543989 kernel: erofs: (device dm-4): mounted with root inode @ nid 39. Apr 20 19:11:12.740963 kernel: loop4: detected capacity change from 0 to 178200 Apr 20 19:11:12.760190 kernel: loop4: p1 p2 p3 Apr 20 19:11:13.080806 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:13.081325 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:11:13.081405 kernel: device-mapper: table: 253:4: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:11:13.081424 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:11:13.082583 systemd-sysext[1514]: device-mapper: reload ioctl on loop4p1-verity (253:4) failed: Invalid argument Apr 20 19:11:13.113518 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:13.579627 kernel: erofs: (device dm-4): mounted with root inode @ nid 39. Apr 20 19:11:13.905450 kernel: loop4: detected capacity change from 0 to 228704 Apr 20 19:11:14.697440 kernel: loop4: detected capacity change from 0 to 378016 Apr 20 19:11:14.761968 kernel: loop4: p1 p2 p3 Apr 20 19:11:16.178540 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:16.179527 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:11:16.179618 kernel: device-mapper: table: 253:4: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:11:16.185336 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:11:16.190076 (sd-merge)[1534]: device-mapper: reload ioctl on loop4p1-verity (253:4) failed: Invalid argument Apr 20 19:11:16.202317 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:17.363077 kernel: erofs: (device dm-4): mounted with root inode @ nid 39. Apr 20 19:11:17.431579 kernel: loop5: detected capacity change from 0 to 178200 Apr 20 19:11:17.437592 kernel: loop5: p1 p2 p3 Apr 20 19:11:18.173017 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:18.202489 kernel: device-mapper: verity: Unrecognized verity feature request: root_hash_sig_key_desc Apr 20 19:11:18.198451 (sd-merge)[1534]: device-mapper: reload ioctl on loop5p1-verity (253:5) failed: Invalid argument Apr 20 19:11:18.205878 kernel: device-mapper: table: 253:5: verity: Unrecognized verity feature request (-EINVAL) Apr 20 19:11:18.207763 kernel: device-mapper: ioctl: error adding target to table Apr 20 19:11:18.219284 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 20 19:11:19.845285 kernel: erofs: (device dm-5): mounted with root inode @ nid 39. Apr 20 19:11:20.084001 kernel: loop6: detected capacity change from 0 to 228704 Apr 20 19:11:20.549846 (sd-merge)[1534]: Skipping extension refresh because no change was found, use --always-refresh=yes to always do a refresh. Apr 20 19:11:20.695506 kernel: device-mapper: ioctl: remove_all left 4 open device(s) Apr 20 19:11:20.699662 kernel: device-mapper: ioctl: remove_all left 4 open device(s) Apr 20 19:11:20.706871 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 20 19:11:20.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:20.745907 kernel: audit: type=1130 audit(1776712280.728:172): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:21.038797 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 20 19:11:28.241876 systemd-tmpfiles[1551]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Apr 20 19:11:28.263927 systemd-tmpfiles[1551]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Apr 20 19:11:28.625768 systemd-tmpfiles[1551]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 20 19:11:30.351393 systemd-tmpfiles[1551]: ACLs are not supported, ignoring. Apr 20 19:11:30.351484 systemd-tmpfiles[1551]: ACLs are not supported, ignoring. Apr 20 19:11:31.664964 systemd-tmpfiles[1551]: Detected autofs mount point /boot during canonicalization of boot. Apr 20 19:11:31.717216 systemd-tmpfiles[1551]: Skipping /boot Apr 20 19:11:35.350842 systemd-tmpfiles[1551]: Detected autofs mount point /boot during canonicalization of boot. Apr 20 19:11:35.350876 systemd-tmpfiles[1551]: Skipping /boot Apr 20 19:11:39.291048 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 20 19:11:39.297000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:39.306634 kernel: audit: type=1130 audit(1776712299.297:173): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:40.142104 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 20 19:11:40.540904 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 20 19:11:40.695857 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 20 19:11:40.735618 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 20 19:11:40.780914 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 20 19:11:40.924000 audit[1561]: AUDIT1127 pid=1561 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Apr 20 19:11:40.935323 kernel: audit: type=1127 audit(1776712300.924:174): pid=1561 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Apr 20 19:11:41.007758 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 20 19:11:41.021000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:41.056690 kernel: audit: type=1130 audit(1776712301.021:175): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:41.203111 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 20 19:11:41.206000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:41.215672 kernel: audit: type=1130 audit(1776712301.206:176): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:11:41.650000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Apr 20 19:11:41.650000 audit[1583]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe951dae80 a2=420 a3=0 items=0 ppid=1557 pid=1583 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:11:41.650000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Apr 20 19:11:41.703631 augenrules[1583]: No rules Apr 20 19:11:41.708441 kernel: audit: type=1305 audit(1776712301.650:177): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Apr 20 19:11:41.706095 systemd[1]: audit-rules.service: Deactivated successfully. Apr 20 19:11:41.709162 kernel: audit: type=1300 audit(1776712301.650:177): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe951dae80 a2=420 a3=0 items=0 ppid=1557 pid=1583 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:11:41.709286 kernel: audit: type=1327 audit(1776712301.650:177): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Apr 20 19:11:41.717862 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 20 19:11:41.842098 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 20 19:11:41.851730 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 20 19:12:10.772300 ldconfig[1559]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 20 19:12:11.420961 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 20 19:12:12.019370 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 20 19:12:16.852279 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 20 19:12:17.044323 systemd[1]: Reached target sysinit.target - System Initialization. Apr 20 19:12:17.192197 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 20 19:12:17.203923 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 20 19:12:17.226479 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Apr 20 19:12:17.635537 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 20 19:12:17.715723 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 20 19:12:17.906111 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Apr 20 19:12:18.047417 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Apr 20 19:12:18.081880 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 20 19:12:18.178558 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 20 19:12:18.179177 systemd[1]: Reached target paths.target - Path Units. Apr 20 19:12:18.258900 systemd[1]: Reached target timers.target - Timer Units. Apr 20 19:12:18.306263 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 20 19:12:19.484569 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 20 19:12:20.080509 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Apr 20 19:12:21.185774 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 20 19:12:21.206425 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Apr 20 19:12:21.352931 systemd[1]: Listening on systemd-logind-varlink.socket - User Login Management Varlink Socket. Apr 20 19:12:21.450932 systemd[1]: Listening on systemd-machined.socket - Virtual Machine and Container Registration Service Socket. Apr 20 19:12:21.994816 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 20 19:12:25.040494 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 20 19:12:25.136691 systemd[1]: Reached target sockets.target - Socket Units. Apr 20 19:12:25.165183 systemd[1]: Reached target basic.target - Basic System. Apr 20 19:12:25.243249 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 20 19:12:25.243409 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 20 19:12:25.361035 systemd[1]: Starting containerd.service - containerd container runtime... Apr 20 19:12:25.700374 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Apr 20 19:12:25.706452 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 20 19:12:25.724287 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 20 19:12:26.043344 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 20 19:12:26.069331 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 20 19:12:26.079065 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 20 19:12:26.094791 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Apr 20 19:12:26.178026 jq[1603]: false Apr 20 19:12:26.177945 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:12:26.217865 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 20 19:12:26.325065 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 20 19:12:26.374793 google_oslogin_nss_cache[1605]: oslogin_cache_refresh[1605]: Refreshing passwd entry cache Apr 20 19:12:26.374784 oslogin_cache_refresh[1605]: Refreshing passwd entry cache Apr 20 19:12:26.453732 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 20 19:12:26.474422 google_oslogin_nss_cache[1605]: oslogin_cache_refresh[1605]: Failure getting users, quitting Apr 20 19:12:26.474422 google_oslogin_nss_cache[1605]: oslogin_cache_refresh[1605]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Apr 20 19:12:26.474422 google_oslogin_nss_cache[1605]: oslogin_cache_refresh[1605]: Refreshing group entry cache Apr 20 19:12:26.474798 extend-filesystems[1604]: Found /dev/vda6 Apr 20 19:12:26.468786 oslogin_cache_refresh[1605]: Failure getting users, quitting Apr 20 19:12:26.484384 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 20 19:12:26.468808 oslogin_cache_refresh[1605]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Apr 20 19:12:26.470825 oslogin_cache_refresh[1605]: Refreshing group entry cache Apr 20 19:12:26.526871 google_oslogin_nss_cache[1605]: oslogin_cache_refresh[1605]: Failure getting groups, quitting Apr 20 19:12:26.526871 google_oslogin_nss_cache[1605]: oslogin_cache_refresh[1605]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Apr 20 19:12:26.520716 oslogin_cache_refresh[1605]: Failure getting groups, quitting Apr 20 19:12:26.522007 oslogin_cache_refresh[1605]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Apr 20 19:12:26.565941 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 20 19:12:26.625439 extend-filesystems[1604]: Found /dev/vda9 Apr 20 19:12:26.642494 extend-filesystems[1604]: Checking size of /dev/vda9 Apr 20 19:12:26.667378 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 20 19:12:26.690746 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 20 19:12:26.726656 systemd[1]: Starting update-engine.service - Update Engine... Apr 20 19:12:26.850848 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 20 19:12:27.006053 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 20 19:12:27.011815 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 20 19:12:27.012190 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 20 19:12:27.014403 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Apr 20 19:12:27.014712 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Apr 20 19:12:27.034691 extend-filesystems[1604]: Resized partition /dev/vda9 Apr 20 19:12:27.041746 systemd[1]: motdgen.service: Deactivated successfully. Apr 20 19:12:27.062203 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 20 19:12:27.174638 extend-filesystems[1639]: resize2fs 1.47.3 (8-Jul-2025) Apr 20 19:12:27.212017 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Apr 20 19:12:27.215840 jq[1627]: true Apr 20 19:12:27.219716 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 20 19:12:27.225340 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 20 19:12:27.450281 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 20 19:12:27.963150 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Apr 20 19:12:28.081372 extend-filesystems[1639]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Apr 20 19:12:28.081372 extend-filesystems[1639]: old_desc_blocks = 1, new_desc_blocks = 1 Apr 20 19:12:28.081372 extend-filesystems[1639]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Apr 20 19:12:28.527818 extend-filesystems[1604]: Resized filesystem in /dev/vda9 Apr 20 19:12:28.432107 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 20 19:12:28.726212 update_engine[1624]: I20260420 19:12:28.701936 1624 main.cc:92] Flatcar Update Engine starting Apr 20 19:12:28.475493 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 20 19:12:28.765706 jq[1647]: true Apr 20 19:12:28.766089 sshd_keygen[1642]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 20 19:12:29.034312 systemd[1]: coreos-metadata.service: Deactivated successfully. Apr 20 19:12:29.037121 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Apr 20 19:12:29.167782 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 20 19:12:29.183278 tar[1641]: linux-amd64/LICENSE Apr 20 19:12:29.183278 tar[1641]: linux-amd64/helm Apr 20 19:12:30.256315 systemd-logind[1618]: Watching system buttons on /dev/input/event2 (Power Button) Apr 20 19:12:30.277630 systemd-logind[1618]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 20 19:12:30.415598 systemd-logind[1618]: New seat seat0. Apr 20 19:12:30.575450 bash[1708]: Updated "/home/core/.ssh/authorized_keys" Apr 20 19:12:30.575538 dbus-daemon[1601]: [system] SELinux support is enabled Apr 20 19:12:30.825427 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 20 19:12:31.782099 dbus-daemon[1601]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 20 19:12:31.850181 update_engine[1624]: I20260420 19:12:31.782804 1624 update_check_scheduler.cc:74] Next update check in 10m17s Apr 20 19:12:31.916429 systemd[1]: Started systemd-logind.service - User Login Management. Apr 20 19:12:31.947860 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 20 19:12:31.963666 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 20 19:12:33.436932 systemd[1]: Started update-engine.service - Update Engine. Apr 20 19:12:34.249346 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 20 19:12:34.481790 systemd[1]: Started sshd@0-1-10.0.0.18:22-10.0.0.1:60662.service - OpenSSH per-connection server daemon (10.0.0.1:60662). Apr 20 19:12:34.797847 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Apr 20 19:12:34.922754 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 20 19:12:34.951557 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 20 19:12:35.036504 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 20 19:12:35.037019 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 20 19:12:35.106499 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 20 19:12:36.420773 systemd[1]: issuegen.service: Deactivated successfully. Apr 20 19:12:36.421521 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 20 19:12:36.678823 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 20 19:12:38.435944 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 20 19:12:39.007571 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 20 19:12:39.673374 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Apr 20 19:12:39.770138 systemd[1]: Reached target getty.target - Login Prompts. Apr 20 19:12:41.037867 locksmithd[1723]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 20 19:12:41.582410 sshd[1722]: Accepted publickey for core from 10.0.0.1 port 60662 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:12:41.788897 sshd-session[1722]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:12:44.139893 containerd[1648]: time="2026-04-20T19:12:44Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Apr 20 19:12:44.656769 containerd[1648]: time="2026-04-20T19:12:44.653983123Z" level=info msg="starting containerd" revision=dea7da592f5d1d2b7755e3a161be07f43fad8f75 version=v2.2.1 Apr 20 19:12:44.770403 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 20 19:12:45.283902 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 20 19:12:46.099169 systemd-logind[1618]: New session '1' of user 'core' with class 'user' and type 'tty'. Apr 20 19:12:46.314733 containerd[1648]: time="2026-04-20T19:12:46.306145211Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="547.233µs" Apr 20 19:12:46.314733 containerd[1648]: time="2026-04-20T19:12:46.306512938Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Apr 20 19:12:46.314733 containerd[1648]: time="2026-04-20T19:12:46.306839709Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Apr 20 19:12:46.314733 containerd[1648]: time="2026-04-20T19:12:46.306887182Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Apr 20 19:12:46.359716 containerd[1648]: time="2026-04-20T19:12:46.358883988Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Apr 20 19:12:47.013579 containerd[1648]: time="2026-04-20T19:12:46.370125838Z" level=info msg="loading plugin" id=io.containerd.mount-handler.v1.erofs type=io.containerd.mount-handler.v1 Apr 20 19:12:47.052069 containerd[1648]: time="2026-04-20T19:12:47.047033499Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Apr 20 19:12:47.121657 containerd[1648]: time="2026-04-20T19:12:47.085557645Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Apr 20 19:12:47.121657 containerd[1648]: time="2026-04-20T19:12:47.085809437Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Apr 20 19:12:47.217698 containerd[1648]: time="2026-04-20T19:12:47.204411698Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Apr 20 19:12:47.232656 containerd[1648]: time="2026-04-20T19:12:47.217689722Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Apr 20 19:12:47.232656 containerd[1648]: time="2026-04-20T19:12:47.222536583Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Apr 20 19:12:47.274567 containerd[1648]: time="2026-04-20T19:12:47.232485604Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Apr 20 19:12:47.353817 containerd[1648]: time="2026-04-20T19:12:47.334574059Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Apr 20 19:12:47.353817 containerd[1648]: time="2026-04-20T19:12:47.352409541Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Apr 20 19:12:47.596431 containerd[1648]: time="2026-04-20T19:12:47.564176846Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Apr 20 19:12:47.660423 containerd[1648]: time="2026-04-20T19:12:47.660120932Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Apr 20 19:12:47.675081 containerd[1648]: time="2026-04-20T19:12:47.660489625Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Apr 20 19:12:47.675081 containerd[1648]: time="2026-04-20T19:12:47.661452454Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Apr 20 19:12:47.706577 containerd[1648]: time="2026-04-20T19:12:47.702975689Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Apr 20 19:12:47.710841 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 20 19:12:47.815573 containerd[1648]: time="2026-04-20T19:12:47.735821092Z" level=info msg="metadata content store policy set" policy=shared Apr 20 19:12:48.432620 containerd[1648]: time="2026-04-20T19:12:48.368599476Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Apr 20 19:12:48.542205 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 20 19:12:48.558564 containerd[1648]: time="2026-04-20T19:12:48.543728940Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Apr 20 19:12:48.564781 containerd[1648]: time="2026-04-20T19:12:48.564522311Z" level=info msg="built-in NRI default validator is disabled" Apr 20 19:12:48.572516 containerd[1648]: time="2026-04-20T19:12:48.568804735Z" level=info msg="runtime interface created" Apr 20 19:12:48.673119 containerd[1648]: time="2026-04-20T19:12:48.647299912Z" level=info msg="created NRI interface" Apr 20 19:12:48.781536 containerd[1648]: time="2026-04-20T19:12:48.777398699Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Apr 20 19:12:48.847622 containerd[1648]: time="2026-04-20T19:12:48.844608180Z" level=info msg="skip loading plugin" error="failed to check mkfs.erofs availability: failed to run mkfs.erofs --help: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Apr 20 19:12:48.961408 containerd[1648]: time="2026-04-20T19:12:48.847346810Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Apr 20 19:12:48.961408 containerd[1648]: time="2026-04-20T19:12:48.932913504Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Apr 20 19:12:48.961408 containerd[1648]: time="2026-04-20T19:12:48.946800877Z" level=info msg="loading plugin" id=io.containerd.mount-manager.v1.bolt type=io.containerd.mount-manager.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.961686776Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.962143936Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.962167895Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.962322075Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.964362739Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.968360685Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.989421235Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:48.997858452Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Apr 20 19:12:49.046681 containerd[1648]: time="2026-04-20T19:12:49.013348379Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Apr 20 19:12:49.062511 containerd[1648]: time="2026-04-20T19:12:49.060457004Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Apr 20 19:12:49.075654 containerd[1648]: time="2026-04-20T19:12:49.073027595Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Apr 20 19:12:49.148070 containerd[1648]: time="2026-04-20T19:12:49.081617552Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Apr 20 19:12:49.180343 containerd[1648]: time="2026-04-20T19:12:49.176199975Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Apr 20 19:12:49.196376 containerd[1648]: time="2026-04-20T19:12:49.189840628Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Apr 20 19:12:49.205713 containerd[1648]: time="2026-04-20T19:12:49.203841853Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Apr 20 19:12:49.291681 containerd[1648]: time="2026-04-20T19:12:49.275936937Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Apr 20 19:12:49.321194 containerd[1648]: time="2026-04-20T19:12:49.293462280Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Apr 20 19:12:49.321194 containerd[1648]: time="2026-04-20T19:12:49.310181970Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.mounts type=io.containerd.grpc.v1 Apr 20 19:12:49.321194 containerd[1648]: time="2026-04-20T19:12:49.311205734Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Apr 20 19:12:49.321194 containerd[1648]: time="2026-04-20T19:12:49.311305646Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Apr 20 19:12:49.321194 containerd[1648]: time="2026-04-20T19:12:49.311382906Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Apr 20 19:12:49.335303 containerd[1648]: time="2026-04-20T19:12:49.334345347Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Apr 20 19:12:49.380508 containerd[1648]: time="2026-04-20T19:12:49.378375198Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Apr 20 19:12:49.488099 containerd[1648]: time="2026-04-20T19:12:49.448210625Z" level=info msg="Start snapshots syncer" Apr 20 19:12:49.518944 containerd[1648]: time="2026-04-20T19:12:49.473946282Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Apr 20 19:12:49.536036 tar[1641]: linux-amd64/README.md Apr 20 19:12:49.541263 containerd[1648]: time="2026-04-20T19:12:49.540757874Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Apr 20 19:12:49.541263 containerd[1648]: time="2026-04-20T19:12:49.541054999Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Apr 20 19:12:49.541649 containerd[1648]: time="2026-04-20T19:12:49.541626990Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Apr 20 19:12:49.541982 containerd[1648]: time="2026-04-20T19:12:49.541964427Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Apr 20 19:12:49.542039 containerd[1648]: time="2026-04-20T19:12:49.542030623Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Apr 20 19:12:49.542076 containerd[1648]: time="2026-04-20T19:12:49.542069240Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Apr 20 19:12:49.542107 containerd[1648]: time="2026-04-20T19:12:49.542101080Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Apr 20 19:12:49.542188 containerd[1648]: time="2026-04-20T19:12:49.542181233Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Apr 20 19:12:49.542299 containerd[1648]: time="2026-04-20T19:12:49.542289396Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542328727Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542366822Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542397983Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542492597Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542528289Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542536857Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542667372Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542674634Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542797873Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542822850Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542879930Z" level=info msg="Connect containerd service" Apr 20 19:12:49.549593 containerd[1648]: time="2026-04-20T19:12:49.542976324Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 20 19:12:49.674760 (systemd)[1750]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:12:50.021456 containerd[1648]: time="2026-04-20T19:12:50.012327726Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 20 19:12:50.572744 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 20 19:12:51.422410 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:12:51.821970 systemd-logind[1618]: New session '2' of user 'core' with class 'manager-early' and type 'unspecified'. Apr 20 19:12:52.156061 (kubelet)[1762]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:12:55.581842 containerd[1648]: time="2026-04-20T19:12:55.550927739Z" level=info msg="Start subscribing containerd event" Apr 20 19:12:55.676092 containerd[1648]: time="2026-04-20T19:12:55.584908811Z" level=info msg="Start recovering state" Apr 20 19:12:55.741997 containerd[1648]: time="2026-04-20T19:12:55.740213101Z" level=info msg="Start event monitor" Apr 20 19:12:55.773855 containerd[1648]: time="2026-04-20T19:12:55.740125805Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 20 19:12:55.773855 containerd[1648]: time="2026-04-20T19:12:55.746850891Z" level=info msg="Start cni network conf syncer for default" Apr 20 19:12:55.773855 containerd[1648]: time="2026-04-20T19:12:55.771367208Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 20 19:12:55.782324 containerd[1648]: time="2026-04-20T19:12:55.780857998Z" level=info msg="Start streaming server" Apr 20 19:12:55.838524 containerd[1648]: time="2026-04-20T19:12:55.821651618Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Apr 20 19:12:55.838524 containerd[1648]: time="2026-04-20T19:12:55.832161627Z" level=info msg="runtime interface starting up..." Apr 20 19:12:55.846781 containerd[1648]: time="2026-04-20T19:12:55.838768394Z" level=info msg="starting plugins..." Apr 20 19:12:55.848863 containerd[1648]: time="2026-04-20T19:12:55.848402603Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Apr 20 19:12:55.903443 containerd[1648]: time="2026-04-20T19:12:55.901302874Z" level=info msg="containerd successfully booted in 11.824184s" Apr 20 19:12:55.912082 systemd[1]: Started containerd.service - containerd container runtime. Apr 20 19:12:55.912844 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 20 19:13:06.902912 kubelet[1762]: E0420 19:13:06.901952 1762 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:13:06.931354 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:13:06.931496 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:13:06.932572 systemd[1]: kubelet.service: Consumed 15.030s CPU time, 270.9M memory peak. Apr 20 19:13:16.798710 update_engine[1624]: I20260420 19:13:16.778908 1624 update_attempter.cc:509] Updating boot flags... Apr 20 19:13:17.234577 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 20 19:13:17.622700 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:13:22.708460 systemd[1750]: Queued start job for default target default.target. Apr 20 19:13:24.542833 systemd[1750]: Created slice app.slice - User Application Slice. Apr 20 19:13:24.611067 systemd[1750]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Apr 20 19:13:24.747119 systemd[1750]: Reached target machines.target - Virtual Machines and Containers. Apr 20 19:13:24.808098 systemd[1750]: Reached target paths.target - Paths. Apr 20 19:13:24.898078 systemd[1750]: Reached target timers.target - Timers. Apr 20 19:13:25.681193 systemd[1750]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 20 19:13:26.463967 systemd[1750]: Listening on systemd-ask-password.socket - Query the User Interactively for a Password. Apr 20 19:13:26.932089 systemd[1750]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Apr 20 19:13:27.769133 systemd[1750]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Apr 20 19:13:28.769309 systemd[1750]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 20 19:13:29.354471 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:13:30.063708 (kubelet)[1822]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:13:31.711042 systemd[1750]: Reached target sockets.target - Sockets. Apr 20 19:13:31.765841 systemd[1750]: Reached target basic.target - Basic System. Apr 20 19:13:31.818962 systemd[1750]: Reached target default.target - Main User Target. Apr 20 19:13:31.824421 systemd[1750]: Startup finished in 38.844s. Apr 20 19:13:31.964720 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 20 19:13:32.251777 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 20 19:13:32.336003 systemd[1]: Startup finished in 20.931s (kernel) + 2min 25.165s (initrd) + 4min 36.463s (userspace) = 7min 22.560s. Apr 20 19:13:38.905111 systemd[1]: Started sshd@1-4097-10.0.0.18:22-10.0.0.1:52640.service - OpenSSH per-connection server daemon (10.0.0.1:52640). Apr 20 19:13:43.295841 kubelet[1822]: E0420 19:13:43.283206 1822 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:13:43.465076 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:13:43.502007 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:13:43.558892 sshd[1836]: Accepted publickey for core from 10.0.0.1 port 52640 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:13:43.626187 systemd[1]: kubelet.service: Consumed 10.246s CPU time, 112.7M memory peak. Apr 20 19:13:43.786047 sshd-session[1836]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:13:50.771070 systemd-logind[1618]: New session '3' of user 'core' with class 'user' and type 'tty'. Apr 20 19:14:00.937521 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 20 19:14:05.985069 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 20 19:14:07.948432 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:14:11.287533 sshd[1844]: Connection closed by 10.0.0.1 port 52640 Apr 20 19:14:11.290134 sshd-session[1836]: pam_unix(sshd:session): session closed for user core Apr 20 19:14:11.466447 systemd[1]: sshd@1-4097-10.0.0.18:22-10.0.0.1:52640.service: Deactivated successfully. Apr 20 19:14:11.468859 systemd[1]: sshd@1-4097-10.0.0.18:22-10.0.0.1:52640.service: Consumed 1.412s CPU time, 4.2M memory peak. Apr 20 19:14:11.801570 systemd[1]: session-3.scope: Deactivated successfully. Apr 20 19:14:11.883729 systemd[1]: session-3.scope: Consumed 1.270s CPU time, 2.6M memory peak. Apr 20 19:14:12.713962 systemd-logind[1618]: Session 3 logged out. Waiting for processes to exit. Apr 20 19:14:13.442902 systemd[1]: Started sshd@2-8193-10.0.0.18:22-10.0.0.1:44352.service - OpenSSH per-connection server daemon (10.0.0.1:44352). Apr 20 19:14:13.760986 systemd-logind[1618]: Removed session 3. Apr 20 19:14:17.246447 sshd[1852]: Accepted publickey for core from 10.0.0.1 port 44352 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:14:17.596561 sshd-session[1852]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:14:21.672751 systemd-logind[1618]: New session '4' of user 'core' with class 'user' and type 'tty'. Apr 20 19:14:23.224809 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 20 19:14:25.637737 sshd[1858]: Connection closed by 10.0.0.1 port 44352 Apr 20 19:14:25.654089 sshd-session[1852]: pam_unix(sshd:session): session closed for user core Apr 20 19:14:26.201290 systemd[1]: sshd@2-8193-10.0.0.18:22-10.0.0.1:44352.service: Deactivated successfully. Apr 20 19:14:26.360468 systemd[1]: sshd@2-8193-10.0.0.18:22-10.0.0.1:44352.service: Consumed 1.902s CPU time, 4.4M memory peak. Apr 20 19:14:26.797936 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:14:26.842146 systemd[1]: session-4.scope: Deactivated successfully. Apr 20 19:14:27.144993 systemd-logind[1618]: Session 4 logged out. Waiting for processes to exit. Apr 20 19:14:27.173577 (kubelet)[1864]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:14:28.313770 systemd[1]: Started sshd@3-2-10.0.0.18:22-10.0.0.1:49044.service - OpenSSH per-connection server daemon (10.0.0.1:49044). Apr 20 19:14:28.356116 systemd-logind[1618]: Removed session 4. Apr 20 19:14:32.401675 sshd[1873]: Accepted publickey for core from 10.0.0.1 port 49044 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:14:32.566130 sshd-session[1873]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:14:32.647449 systemd-logind[1618]: New session '5' of user 'core' with class 'user' and type 'tty'. Apr 20 19:14:32.801607 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 20 19:14:32.815107 kubelet[1864]: E0420 19:14:32.806073 1864 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:14:32.817711 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:14:32.817843 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:14:32.818455 systemd[1]: kubelet.service: Consumed 10.575s CPU time, 110.4M memory peak. Apr 20 19:14:33.235366 sshd[1879]: Connection closed by 10.0.0.1 port 49044 Apr 20 19:14:33.236407 sshd-session[1873]: pam_unix(sshd:session): session closed for user core Apr 20 19:14:33.489003 systemd[1]: sshd@3-2-10.0.0.18:22-10.0.0.1:49044.service: Deactivated successfully. Apr 20 19:14:33.493936 systemd[1]: sshd@3-2-10.0.0.18:22-10.0.0.1:49044.service: Consumed 2.128s CPU time, 4.4M memory peak. Apr 20 19:14:33.634952 systemd[1]: session-5.scope: Deactivated successfully. Apr 20 19:14:33.697033 systemd-logind[1618]: Session 5 logged out. Waiting for processes to exit. Apr 20 19:14:33.899050 systemd[1]: Started sshd@4-8194-10.0.0.18:22-10.0.0.1:49046.service - OpenSSH per-connection server daemon (10.0.0.1:49046). Apr 20 19:14:33.902662 systemd-logind[1618]: Removed session 5. Apr 20 19:14:36.782890 sshd[1885]: Accepted publickey for core from 10.0.0.1 port 49046 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:14:36.861268 sshd-session[1885]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:14:36.952770 systemd-logind[1618]: New session '6' of user 'core' with class 'user' and type 'tty'. Apr 20 19:14:36.986801 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 20 19:14:38.298946 sudo[1890]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 20 19:14:38.301858 sudo[1890]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 20 19:14:38.379659 sudo[1890]: pam_unix(sudo:session): session closed for user root Apr 20 19:14:38.585651 sshd[1889]: Connection closed by 10.0.0.1 port 49046 Apr 20 19:14:38.590100 sshd-session[1885]: pam_unix(sshd:session): session closed for user core Apr 20 19:14:38.834662 systemd[1]: sshd@4-8194-10.0.0.18:22-10.0.0.1:49046.service: Deactivated successfully. Apr 20 19:14:38.859730 systemd[1]: sshd@4-8194-10.0.0.18:22-10.0.0.1:49046.service: Consumed 1.497s CPU time, 4.6M memory peak. Apr 20 19:14:38.957913 systemd[1]: session-6.scope: Deactivated successfully. Apr 20 19:14:39.330006 systemd-logind[1618]: Session 6 logged out. Waiting for processes to exit. Apr 20 19:14:40.189107 systemd[1]: Started sshd@5-3-10.0.0.18:22-10.0.0.1:37416.service - OpenSSH per-connection server daemon (10.0.0.1:37416). Apr 20 19:14:40.194928 systemd-logind[1618]: Removed session 6. Apr 20 19:14:43.291164 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Apr 20 19:14:43.613655 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:14:44.233621 sshd[1897]: Accepted publickey for core from 10.0.0.1 port 37416 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:14:44.241770 sshd-session[1897]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:14:44.576531 systemd-logind[1618]: New session '7' of user 'core' with class 'user' and type 'tty'. Apr 20 19:14:44.769670 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 20 19:14:45.851022 sudo[1906]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 20 19:14:45.981448 sudo[1906]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 20 19:14:46.009432 sudo[1906]: pam_unix(sudo:session): session closed for user root Apr 20 19:14:46.448423 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:14:46.457114 sudo[1905]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Apr 20 19:14:46.506649 sudo[1905]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 20 19:14:46.593536 (kubelet)[1914]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:14:46.997927 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 20 19:14:47.710180 kubelet[1914]: E0420 19:14:47.708649 1914 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:14:47.718177 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:14:47.718454 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:14:47.722462 systemd[1]: kubelet.service: Consumed 2.213s CPU time, 111.9M memory peak. Apr 20 19:14:47.735000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Apr 20 19:14:47.737809 augenrules[1943]: No rules Apr 20 19:14:47.735000 audit[1943]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffee70c2830 a2=420 a3=0 items=0 ppid=1923 pid=1943 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:14:47.750490 kernel: audit: type=1305 audit(1776712487.735:178): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Apr 20 19:14:47.735000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Apr 20 19:14:47.765274 kernel: audit: type=1300 audit(1776712487.735:178): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffee70c2830 a2=420 a3=0 items=0 ppid=1923 pid=1943 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:14:47.765366 kernel: audit: type=1327 audit(1776712487.735:178): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Apr 20 19:14:47.799424 systemd[1]: audit-rules.service: Deactivated successfully. Apr 20 19:14:47.815169 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 20 19:14:47.839000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.843000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.849629 sudo[1905]: pam_unix(sudo:session): session closed for user root Apr 20 19:14:47.848000 audit[1905]: AUDIT1106 pid=1905 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.848000 audit[1905]: AUDIT1104 pid=1905 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.913922 kernel: audit: type=1130 audit(1776712487.839:179): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.913985 kernel: audit: type=1131 audit(1776712487.843:180): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.914000 kernel: audit: type=1106 audit(1776712487.848:181): pid=1905 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.914027 kernel: audit: type=1104 audit(1776712487.848:182): pid=1905 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:47.926987 sshd[1904]: Connection closed by 10.0.0.1 port 37416 Apr 20 19:14:47.932087 sshd-session[1897]: pam_unix(sshd:session): session closed for user core Apr 20 19:14:47.932000 audit[1897]: AUDIT1106 pid=1897 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:47.932000 audit[1897]: AUDIT1104 pid=1897 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:47.946625 kernel: audit: type=1106 audit(1776712487.932:183): pid=1897 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:47.949126 kernel: audit: type=1104 audit(1776712487.932:184): pid=1897 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:48.240987 systemd[1]: sshd@5-3-10.0.0.18:22-10.0.0.1:37416.service: Deactivated successfully. Apr 20 19:14:48.254000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-3-10.0.0.18:22-10.0.0.1:37416 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:48.262981 systemd[1]: sshd@5-3-10.0.0.18:22-10.0.0.1:37416.service: Consumed 1.889s CPU time, 4.4M memory peak. Apr 20 19:14:48.268905 kernel: audit: type=1131 audit(1776712488.254:185): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-3-10.0.0.18:22-10.0.0.1:37416 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:48.404088 systemd[1]: session-7.scope: Deactivated successfully. Apr 20 19:14:48.415978 systemd-logind[1618]: Session 7 logged out. Waiting for processes to exit. Apr 20 19:14:48.735545 systemd[1]: Started sshd@6-12289-10.0.0.18:22-10.0.0.1:52626.service - OpenSSH per-connection server daemon (10.0.0.1:52626). Apr 20 19:14:48.734000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-12289-10.0.0.18:22-10.0.0.1:52626 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:14:48.737175 systemd-logind[1618]: Removed session 7. Apr 20 19:14:51.004000 audit[1952]: AUDIT1101 pid=1952 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:51.010943 sshd[1952]: Accepted publickey for core from 10.0.0.1 port 52626 ssh2: RSA SHA256:ZNIzts6V4KYKlrJxXaosrimCRlmsV/+NkZ5UtjwHrjE Apr 20 19:14:51.011000 audit[1952]: AUDIT1103 pid=1952 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:51.016000 audit[1952]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffb6179560 a2=3 a3=0 items=0 ppid=1 pid=1952 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:14:51.016000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Apr 20 19:14:51.023098 sshd-session[1952]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 20 19:14:51.359895 systemd-logind[1618]: New session '8' of user 'core' with class 'user' and type 'tty'. Apr 20 19:14:51.592159 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 20 19:14:51.617000 audit[1952]: AUDIT1105 pid=1952 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:51.630000 audit[1956]: AUDIT1103 pid=1956 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:14:52.385000 audit[1957]: AUDIT1101 pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:52.390105 sudo[1957]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 20 19:14:52.389000 audit[1957]: AUDIT1110 pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:52.389000 audit[1957]: AUDIT1105 pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:14:52.390755 sudo[1957]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 20 19:14:57.896552 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Apr 20 19:14:58.402383 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:15:04.245804 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:15:04.306000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:04.325900 kernel: kauditd_printk_skb: 11 callbacks suppressed Apr 20 19:15:04.326052 kernel: audit: type=1130 audit(1776712504.306:195): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:04.430011 (kubelet)[1986]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:15:05.043520 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 20 19:15:05.295550 (dockerd)[1993]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 20 19:15:08.256212 kubelet[1986]: E0420 19:15:08.254950 1986 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:15:08.265563 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:15:08.265742 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:15:08.282000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:15:08.285422 systemd[1]: kubelet.service: Consumed 5.684s CPU time, 108.4M memory peak. Apr 20 19:15:08.290287 kernel: audit: type=1131 audit(1776712508.282:196): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:15:16.017646 dockerd[1993]: time="2026-04-20T19:15:16.016883605Z" level=info msg="Starting up" Apr 20 19:15:16.042727 dockerd[1993]: time="2026-04-20T19:15:16.041273081Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Apr 20 19:15:16.622709 dockerd[1993]: time="2026-04-20T19:15:16.621596920Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Apr 20 19:15:17.855320 systemd[1]: var-lib-docker-metacopy\x2dcheck930824022-merged.mount: Deactivated successfully. Apr 20 19:15:17.859132 dockerd[1993]: time="2026-04-20T19:15:17.857474789Z" level=info msg="Loading containers: start." Apr 20 19:15:18.202320 kernel: Initializing XFRM netlink socket Apr 20 19:15:18.545638 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Apr 20 19:15:18.639037 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:15:19.808000 audit[2055]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=2055 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:19.808000 audit[2055]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffe739449c0 a2=0 a3=0 items=0 ppid=1993 pid=2055 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:19.808000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Apr 20 19:15:19.886544 kernel: audit: type=1325 audit(1776712519.808:197): table=nat:2 family=2 entries=2 op=nft_register_chain pid=2055 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:19.886666 kernel: audit: type=1300 audit(1776712519.808:197): arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffe739449c0 a2=0 a3=0 items=0 ppid=1993 pid=2055 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:19.886717 kernel: audit: type=1327 audit(1776712519.808:197): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Apr 20 19:15:19.933000 audit[2057]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=2057 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:19.933000 audit[2057]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7fff5f0f0020 a2=0 a3=0 items=0 ppid=1993 pid=2057 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:19.933000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Apr 20 19:15:20.010861 kernel: audit: type=1325 audit(1776712519.933:198): table=filter:3 family=2 entries=2 op=nft_register_chain pid=2057 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.010951 kernel: audit: type=1300 audit(1776712519.933:198): arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7fff5f0f0020 a2=0 a3=0 items=0 ppid=1993 pid=2057 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.010978 kernel: audit: type=1327 audit(1776712519.933:198): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Apr 20 19:15:20.126000 audit[2059]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=2059 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.126000 audit[2059]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd2aae7960 a2=0 a3=0 items=0 ppid=1993 pid=2059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.126000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Apr 20 19:15:20.158437 kernel: audit: type=1325 audit(1776712520.126:199): table=filter:4 family=2 entries=1 op=nft_register_chain pid=2059 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.158532 kernel: audit: type=1300 audit(1776712520.126:199): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd2aae7960 a2=0 a3=0 items=0 ppid=1993 pid=2059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.158559 kernel: audit: type=1327 audit(1776712520.126:199): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Apr 20 19:15:20.298000 audit[2061]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=2061 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.298000 audit[2061]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcf6097e10 a2=0 a3=0 items=0 ppid=1993 pid=2061 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.298000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Apr 20 19:15:20.313066 kernel: audit: type=1325 audit(1776712520.298:200): table=filter:5 family=2 entries=1 op=nft_register_chain pid=2061 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.332000 audit[2063]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=2063 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.332000 audit[2063]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffecd13b6a0 a2=0 a3=0 items=0 ppid=1993 pid=2063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.332000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Apr 20 19:15:20.345000 audit[2065]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=2065 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.345000 audit[2065]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fff57674920 a2=0 a3=0 items=0 ppid=1993 pid=2065 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.345000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Apr 20 19:15:20.357000 audit[2067]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=2067 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.357000 audit[2067]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffd301cfb00 a2=0 a3=0 items=0 ppid=1993 pid=2067 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.357000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Apr 20 19:15:20.449000 audit[2069]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=2069 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.449000 audit[2069]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffd57065330 a2=0 a3=0 items=0 ppid=1993 pid=2069 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.449000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Apr 20 19:15:20.663000 audit[2074]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=2074 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.663000 audit[2074]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7ffdeb3b84b0 a2=0 a3=0 items=0 ppid=1993 pid=2074 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.663000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Apr 20 19:15:20.802000 audit[2076]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=2076 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.802000 audit[2076]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7fff072c93f0 a2=0 a3=0 items=0 ppid=1993 pid=2076 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.802000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Apr 20 19:15:20.833000 audit[2078]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=2078 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.833000 audit[2078]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffdb5de16f0 a2=0 a3=0 items=0 ppid=1993 pid=2078 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.833000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Apr 20 19:15:20.843000 audit[2082]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=2082 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.843000 audit[2082]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fff47fd9720 a2=0 a3=0 items=0 ppid=1993 pid=2082 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.843000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Apr 20 19:15:20.929000 audit[2084]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=2084 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:20.929000 audit[2084]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffcf8375220 a2=0 a3=0 items=0 ppid=1993 pid=2084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:20.929000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Apr 20 19:15:21.006910 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:15:21.046000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:21.136204 (kubelet)[2088]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:15:22.134000 audit[2122]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=2122 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.134000 audit[2122]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7fff37987710 a2=0 a3=0 items=0 ppid=1993 pid=2122 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.134000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Apr 20 19:15:22.231000 audit[2124]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=2124 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.231000 audit[2124]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffe9aa21910 a2=0 a3=0 items=0 ppid=1993 pid=2124 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.231000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Apr 20 19:15:22.398000 audit[2126]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=2126 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.398000 audit[2126]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc1124e170 a2=0 a3=0 items=0 ppid=1993 pid=2126 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.398000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Apr 20 19:15:22.446000 audit[2128]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=2128 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.446000 audit[2128]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffff4717f50 a2=0 a3=0 items=0 ppid=1993 pid=2128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.446000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Apr 20 19:15:22.633000 audit[2130]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=2130 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.633000 audit[2130]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe75866900 a2=0 a3=0 items=0 ppid=1993 pid=2130 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.633000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Apr 20 19:15:22.670000 audit[2132]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=2132 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.670000 audit[2132]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fff89926720 a2=0 a3=0 items=0 ppid=1993 pid=2132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.670000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Apr 20 19:15:22.687455 kubelet[2088]: E0420 19:15:22.687370 2088 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:15:22.766000 audit[2134]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=2134 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.766000 audit[2134]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffdb588a120 a2=0 a3=0 items=0 ppid=1993 pid=2134 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.766000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Apr 20 19:15:22.774603 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:15:22.774963 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:15:22.794000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:15:22.800802 systemd[1]: kubelet.service: Consumed 2.638s CPU time, 108.5M memory peak. Apr 20 19:15:22.802000 audit[2137]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=2137 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.802000 audit[2137]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7fff91dc3d00 a2=0 a3=0 items=0 ppid=1993 pid=2137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.802000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Apr 20 19:15:22.839000 audit[2139]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=2139 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.839000 audit[2139]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7ffd220d7180 a2=0 a3=0 items=0 ppid=1993 pid=2139 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.839000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Apr 20 19:15:22.989000 audit[2141]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=2141 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:22.989000 audit[2141]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd3cd6e930 a2=0 a3=0 items=0 ppid=1993 pid=2141 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:22.989000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Apr 20 19:15:23.233000 audit[2143]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=2143 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:23.233000 audit[2143]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffd647e5980 a2=0 a3=0 items=0 ppid=1993 pid=2143 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:23.233000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Apr 20 19:15:23.342000 audit[2145]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=2145 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:23.342000 audit[2145]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fff69a89760 a2=0 a3=0 items=0 ppid=1993 pid=2145 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:23.342000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Apr 20 19:15:23.503000 audit[2147]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2147 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:23.503000 audit[2147]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffc882d5d50 a2=0 a3=0 items=0 ppid=1993 pid=2147 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:23.503000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Apr 20 19:15:24.027000 audit[2152]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2152 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:24.027000 audit[2152]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffdfa5484d0 a2=0 a3=0 items=0 ppid=1993 pid=2152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:24.027000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Apr 20 19:15:24.281000 audit[2154]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2154 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:24.281000 audit[2154]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffe08f53e80 a2=0 a3=0 items=0 ppid=1993 pid=2154 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:24.281000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Apr 20 19:15:24.721000 audit[2156]: NETFILTER_CFG table=filter:30 family=10 entries=1 op=nft_register_chain pid=2156 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:24.721000 audit[2156]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd9755a1e0 a2=0 a3=0 items=0 ppid=1993 pid=2156 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:24.721000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Apr 20 19:15:24.801000 audit[2158]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_rule pid=2158 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:15:24.801000 audit[2158]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffc5162ce30 a2=0 a3=0 items=0 ppid=1993 pid=2158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:24.801000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Apr 20 19:15:25.252000 audit[2164]: NETFILTER_CFG table=nat:32 family=2 entries=2 op=nft_register_chain pid=2164 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.261120 kernel: kauditd_printk_skb: 82 callbacks suppressed Apr 20 19:15:25.261383 kernel: audit: type=1325 audit(1776712525.252:229): table=nat:32 family=2 entries=2 op=nft_register_chain pid=2164 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.252000 audit[2164]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7ffdef11f350 a2=0 a3=0 items=0 ppid=1993 pid=2164 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.252000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Apr 20 19:15:25.303080 kernel: audit: type=1300 audit(1776712525.252:229): arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7ffdef11f350 a2=0 a3=0 items=0 ppid=1993 pid=2164 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.303353 kernel: audit: type=1327 audit(1776712525.252:229): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Apr 20 19:15:25.322000 audit[2166]: NETFILTER_CFG table=nat:33 family=2 entries=1 op=nft_register_rule pid=2166 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.322000 audit[2166]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffcbeb6a120 a2=0 a3=0 items=0 ppid=1993 pid=2166 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.322000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Apr 20 19:15:25.349277 kernel: audit: type=1325 audit(1776712525.322:230): table=nat:33 family=2 entries=1 op=nft_register_rule pid=2166 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.349638 kernel: audit: type=1300 audit(1776712525.322:230): arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffcbeb6a120 a2=0 a3=0 items=0 ppid=1993 pid=2166 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.354183 kernel: audit: type=1327 audit(1776712525.322:230): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Apr 20 19:15:25.617000 audit[2174]: NETFILTER_CFG table=filter:34 family=2 entries=1 op=nft_register_rule pid=2174 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.617000 audit[2174]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffd0343e840 a2=0 a3=0 items=0 ppid=1993 pid=2174 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.617000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Apr 20 19:15:25.626555 kernel: audit: type=1325 audit(1776712525.617:231): table=filter:34 family=2 entries=1 op=nft_register_rule pid=2174 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.628946 kernel: audit: type=1300 audit(1776712525.617:231): arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffd0343e840 a2=0 a3=0 items=0 ppid=1993 pid=2174 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.631694 kernel: audit: type=1327 audit(1776712525.617:231): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Apr 20 19:15:25.844000 audit[2180]: NETFILTER_CFG table=filter:35 family=2 entries=1 op=nft_register_rule pid=2180 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:25.844000 audit[2180]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffc7577fa40 a2=0 a3=0 items=0 ppid=1993 pid=2180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:25.844000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Apr 20 19:15:25.884934 kernel: audit: type=1325 audit(1776712525.844:232): table=filter:35 family=2 entries=1 op=nft_register_rule pid=2180 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:26.001000 audit[2182]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2182 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:26.001000 audit[2182]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7fffa9426a40 a2=0 a3=0 items=0 ppid=1993 pid=2182 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:26.001000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Apr 20 19:15:26.222000 audit[2184]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2184 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:26.222000 audit[2184]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffffb4fdf00 a2=0 a3=0 items=0 ppid=1993 pid=2184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:26.222000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Apr 20 19:15:26.409000 audit[2186]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2186 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:26.409000 audit[2186]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffd399f3bc0 a2=0 a3=0 items=0 ppid=1993 pid=2186 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:26.409000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Apr 20 19:15:26.602000 audit[2188]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2188 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:15:26.602000 audit[2188]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffdf174f730 a2=0 a3=0 items=0 ppid=1993 pid=2188 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:15:26.602000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Apr 20 19:15:26.814613 systemd-networkd[1439]: docker0: Link UP Apr 20 19:15:26.974742 dockerd[1993]: time="2026-04-20T19:15:26.957958534Z" level=info msg="Loading containers: done." Apr 20 19:15:27.806401 dockerd[1993]: time="2026-04-20T19:15:27.804398070Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 20 19:15:27.810523 dockerd[1993]: time="2026-04-20T19:15:27.807106519Z" level=info msg="Docker daemon" commit=45873be4ae3f5488c9498b3d9f17deaddaf609f4 containerd-snapshotter=false storage-driver=overlay2 version=28.2.2 Apr 20 19:15:27.810854 dockerd[1993]: time="2026-04-20T19:15:27.810738932Z" level=info msg="Initializing buildkit" Apr 20 19:15:27.820168 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3573727314-merged.mount: Deactivated successfully. Apr 20 19:15:27.862919 dockerd[1993]: time="2026-04-20T19:15:27.859169501Z" level=warning msg="CDI setup error /var/run/cdi: failed to monitor for changes: no such file or directory" Apr 20 19:15:27.862919 dockerd[1993]: time="2026-04-20T19:15:27.862736366Z" level=warning msg="CDI setup error /etc/cdi: failed to monitor for changes: no such file or directory" Apr 20 19:15:29.438916 dockerd[1993]: time="2026-04-20T19:15:29.438440549Z" level=info msg="Completed buildkit initialization" Apr 20 19:15:29.611754 dockerd[1993]: time="2026-04-20T19:15:29.611093135Z" level=info msg="Daemon has completed initialization" Apr 20 19:15:29.611754 dockerd[1993]: time="2026-04-20T19:15:29.611694093Z" level=info msg="API listen on /run/docker.sock" Apr 20 19:15:29.616553 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 20 19:15:29.618000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:33.106617 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Apr 20 19:15:33.186883 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:15:36.201568 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:15:36.200000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:36.206141 kernel: kauditd_printk_skb: 15 callbacks suppressed Apr 20 19:15:36.206219 kernel: audit: type=1130 audit(1776712536.200:238): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:36.274000 (kubelet)[2239]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:15:37.553158 kubelet[2239]: E0420 19:15:37.546400 2239 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:15:37.673470 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:15:37.684589 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:15:37.691000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:15:37.693020 systemd[1]: kubelet.service: Consumed 2.536s CPU time, 108.4M memory peak. Apr 20 19:15:37.700312 kernel: audit: type=1131 audit(1776712537.691:239): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:15:47.752108 containerd[1648]: time="2026-04-20T19:15:47.737102147Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\"" Apr 20 19:15:47.846798 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Apr 20 19:15:48.121130 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:15:54.257974 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:15:54.310000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:54.326734 kernel: audit: type=1130 audit(1776712554.310:240): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:15:54.390024 (kubelet)[2264]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:15:57.820537 kubelet[2264]: E0420 19:15:57.816594 2264 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:15:57.844062 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:15:57.845139 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:15:57.935000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:15:57.945627 systemd[1]: kubelet.service: Consumed 5.363s CPU time, 109.8M memory peak. Apr 20 19:15:58.316053 kernel: audit: type=1131 audit(1776712557.935:241): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:04.243831 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1763483951.mount: Deactivated successfully. Apr 20 19:16:08.124743 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Apr 20 19:16:08.224499 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:16:11.991660 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:16:12.016000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:12.048144 kernel: audit: type=1130 audit(1776712572.016:242): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:12.157698 (kubelet)[2295]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:16:13.342145 kubelet[2295]: E0420 19:16:13.341211 2295 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:16:13.349773 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:16:13.350001 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:16:13.349000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:13.350903 systemd[1]: kubelet.service: Consumed 2.936s CPU time, 110.8M memory peak. Apr 20 19:16:13.357582 kernel: audit: type=1131 audit(1776712573.349:243): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:23.650822 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Apr 20 19:16:23.965837 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:16:25.370786 containerd[1648]: time="2026-04-20T19:16:25.365061223Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:16:25.493450 containerd[1648]: time="2026-04-20T19:16:25.368438655Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.11: active requests=0, bytes read=30182263" Apr 20 19:16:25.727112 containerd[1648]: time="2026-04-20T19:16:25.717490933Z" level=info msg="ImageCreate event name:\"sha256:7ea99c30f23b106a042b6c46e565fddb42b20bbe58ba6852e562eed03477aec2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:16:26.937897 containerd[1648]: time="2026-04-20T19:16:26.932011835Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:16:27.026348 containerd[1648]: time="2026-04-20T19:16:27.024431373Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.11\" with image id \"sha256:7ea99c30f23b106a042b6c46e565fddb42b20bbe58ba6852e562eed03477aec2\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\", size \"30190588\" in 39.255371273s" Apr 20 19:16:27.048181 containerd[1648]: time="2026-04-20T19:16:27.032011808Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\" returns image reference \"sha256:7ea99c30f23b106a042b6c46e565fddb42b20bbe58ba6852e562eed03477aec2\"" Apr 20 19:16:27.387436 containerd[1648]: time="2026-04-20T19:16:27.362799138Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\"" Apr 20 19:16:27.442000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:27.442845 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:16:27.469634 kernel: audit: type=1130 audit(1776712587.442:244): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:27.643956 (kubelet)[2356]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:16:29.838161 kubelet[2356]: E0420 19:16:29.824448 2356 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:16:29.972550 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:16:30.003487 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:16:30.007000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:30.008766 systemd[1]: kubelet.service: Consumed 3.506s CPU time, 110.9M memory peak. Apr 20 19:16:30.185896 kernel: audit: type=1131 audit(1776712590.007:245): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:40.496069 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Apr 20 19:16:40.874214 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:16:43.598762 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:16:43.598000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:43.609733 kernel: audit: type=1130 audit(1776712603.598:246): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:43.652607 (kubelet)[2376]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:16:44.579930 kubelet[2376]: E0420 19:16:44.578447 2376 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:16:44.603535 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:16:44.613000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:44.608025 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:16:44.623363 kernel: audit: type=1131 audit(1776712604.613:247): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:44.624482 systemd[1]: kubelet.service: Consumed 1.958s CPU time, 110.6M memory peak. Apr 20 19:16:46.534062 containerd[1648]: time="2026-04-20T19:16:46.532474099Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:16:46.552632 containerd[1648]: time="2026-04-20T19:16:46.550957447Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.11: active requests=0, bytes read=26165891" Apr 20 19:16:46.563128 containerd[1648]: time="2026-04-20T19:16:46.562982061Z" level=info msg="ImageCreate event name:\"sha256:c75dc8a6c47e2f7491fa2e367879f53c6f46053066e6b7135df4b154ddd94a1f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:16:46.861869 containerd[1648]: time="2026-04-20T19:16:46.859851752Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:16:46.863589 containerd[1648]: time="2026-04-20T19:16:46.863272505Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.11\" with image id \"sha256:c75dc8a6c47e2f7491fa2e367879f53c6f46053066e6b7135df4b154ddd94a1f\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\", size \"27737794\" in 19.458458419s" Apr 20 19:16:46.863589 containerd[1648]: time="2026-04-20T19:16:46.863349473Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\" returns image reference \"sha256:c75dc8a6c47e2f7491fa2e367879f53c6f46053066e6b7135df4b154ddd94a1f\"" Apr 20 19:16:46.891171 containerd[1648]: time="2026-04-20T19:16:46.889559396Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\"" Apr 20 19:16:54.804825 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Apr 20 19:16:55.014685 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:16:57.100570 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:16:57.142000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:57.178023 kernel: audit: type=1130 audit(1776712617.142:248): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:16:57.215100 (kubelet)[2396]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:16:58.331378 kubelet[2396]: E0420 19:16:58.329828 2396 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:16:58.346712 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:16:58.358735 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:16:58.360000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:16:58.365258 systemd[1]: kubelet.service: Consumed 2.085s CPU time, 110.2M memory peak. Apr 20 19:16:58.373930 kernel: audit: type=1131 audit(1776712618.360:249): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:00.200360 containerd[1648]: time="2026-04-20T19:17:00.199725053Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:00.222620 containerd[1648]: time="2026-04-20T19:17:00.222076863Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.11: active requests=0, bytes read=20280985" Apr 20 19:17:00.393059 containerd[1648]: time="2026-04-20T19:17:00.390431746Z" level=info msg="ImageCreate event name:\"sha256:3febad3451e2d599688a8ad13d19d03c48c9054be209342c748fac2bb6c56f97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:00.625345 containerd[1648]: time="2026-04-20T19:17:00.624576308Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:00.820186 containerd[1648]: time="2026-04-20T19:17:00.819445230Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.11\" with image id \"sha256:3febad3451e2d599688a8ad13d19d03c48c9054be209342c748fac2bb6c56f97\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\", size \"21856121\" in 13.923273235s" Apr 20 19:17:00.820186 containerd[1648]: time="2026-04-20T19:17:00.819744988Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\" returns image reference \"sha256:3febad3451e2d599688a8ad13d19d03c48c9054be209342c748fac2bb6c56f97\"" Apr 20 19:17:00.994345 containerd[1648]: time="2026-04-20T19:17:00.992786935Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\"" Apr 20 19:17:08.487817 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Apr 20 19:17:08.501759 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:17:10.047855 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:17:10.047000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:10.054261 kernel: audit: type=1130 audit(1776712630.047:250): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:10.150356 (kubelet)[2417]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:17:11.079605 kubelet[2417]: E0420 19:17:11.060831 2417 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:17:11.143657 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:17:11.143000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:11.143920 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:17:11.144744 systemd[1]: kubelet.service: Consumed 1.781s CPU time, 109.6M memory peak. Apr 20 19:17:11.150271 kernel: audit: type=1131 audit(1776712631.143:251): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:12.852854 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount763173820.mount: Deactivated successfully. Apr 20 19:17:14.739583 containerd[1648]: time="2026-04-20T19:17:14.739018462Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:14.746999 containerd[1648]: time="2026-04-20T19:17:14.741700922Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.11: active requests=0, bytes read=20491882" Apr 20 19:17:14.841006 containerd[1648]: time="2026-04-20T19:17:14.788163221Z" level=info msg="ImageCreate event name:\"sha256:4ce1332df15d2a0b1c2d3b18292afb4ff670070401211daebb00b7293b26f6d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:14.969083 containerd[1648]: time="2026-04-20T19:17:14.968602381Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:14.999029 containerd[1648]: time="2026-04-20T19:17:14.997376874Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.11\" with image id \"sha256:4ce1332df15d2a0b1c2d3b18292afb4ff670070401211daebb00b7293b26f6d0\", repo tag \"registry.k8s.io/kube-proxy:v1.33.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\", size \"32009730\" in 14.002320563s" Apr 20 19:17:14.999029 containerd[1648]: time="2026-04-20T19:17:14.997591274Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\" returns image reference \"sha256:4ce1332df15d2a0b1c2d3b18292afb4ff670070401211daebb00b7293b26f6d0\"" Apr 20 19:17:15.001524 containerd[1648]: time="2026-04-20T19:17:15.001469682Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Apr 20 19:17:19.761896 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3719203242.mount: Deactivated successfully. Apr 20 19:17:21.254002 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. Apr 20 19:17:21.341999 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:17:23.887772 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:17:23.886000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:23.896860 kernel: audit: type=1130 audit(1776712643.886:252): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:23.923430 (kubelet)[2451]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:17:26.411344 kubelet[2451]: E0420 19:17:26.409971 2451 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:17:26.423049 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:17:26.433000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:26.423365 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:17:26.443609 kernel: audit: type=1131 audit(1776712646.433:253): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:26.434942 systemd[1]: kubelet.service: Consumed 3.374s CPU time, 110.5M memory peak. Apr 20 19:17:36.480121 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 14. Apr 20 19:17:36.486597 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:17:37.145106 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:17:37.149000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:37.155348 kernel: audit: type=1130 audit(1776712657.149:254): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:37.169893 (kubelet)[2506]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:17:37.542933 kubelet[2506]: E0420 19:17:37.542113 2506 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:17:37.550000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:37.549420 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:17:37.559153 kernel: audit: type=1131 audit(1776712657.550:255): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:37.549808 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:17:37.551780 systemd[1]: kubelet.service: Consumed 771ms CPU time, 110M memory peak. Apr 20 19:17:38.279770 containerd[1648]: time="2026-04-20T19:17:38.277833168Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:38.287579 containerd[1648]: time="2026-04-20T19:17:38.286181064Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20931441" Apr 20 19:17:38.369895 containerd[1648]: time="2026-04-20T19:17:38.366014105Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:38.581087 containerd[1648]: time="2026-04-20T19:17:38.576827580Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:38.621860 containerd[1648]: time="2026-04-20T19:17:38.619977697Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 23.618248263s" Apr 20 19:17:38.635571 containerd[1648]: time="2026-04-20T19:17:38.622038373Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Apr 20 19:17:38.650499 containerd[1648]: time="2026-04-20T19:17:38.645038589Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 20 19:17:42.385702 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2137109997.mount: Deactivated successfully. Apr 20 19:17:42.626279 containerd[1648]: time="2026-04-20T19:17:42.624196380Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:42.641875 containerd[1648]: time="2026-04-20T19:17:42.638679447Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Apr 20 19:17:42.642569 containerd[1648]: time="2026-04-20T19:17:42.642464490Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:42.762133 containerd[1648]: time="2026-04-20T19:17:42.761562751Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:17:42.875027 containerd[1648]: time="2026-04-20T19:17:42.873974068Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 4.220569611s" Apr 20 19:17:42.875027 containerd[1648]: time="2026-04-20T19:17:42.874357481Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Apr 20 19:17:42.897798 containerd[1648]: time="2026-04-20T19:17:42.896945693Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Apr 20 19:17:47.764524 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 15. Apr 20 19:17:47.811991 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:17:48.887161 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1556047775.mount: Deactivated successfully. Apr 20 19:17:48.925087 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:17:48.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:48.942454 kernel: audit: type=1130 audit(1776712668.929:256): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:17:48.955113 (kubelet)[2531]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:17:49.351511 kubelet[2531]: E0420 19:17:49.350547 2531 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:17:49.358064 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:17:49.359755 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:17:49.368000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:49.374384 systemd[1]: kubelet.service: Consumed 984ms CPU time, 110.6M memory peak. Apr 20 19:17:49.443634 kernel: audit: type=1131 audit(1776712669.368:257): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:17:59.585215 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 16. Apr 20 19:17:59.734759 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:18:02.433111 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:18:02.438000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:02.449975 kernel: audit: type=1130 audit(1776712682.438:258): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:02.561932 (kubelet)[2555]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:18:03.974654 kubelet[2555]: E0420 19:18:03.974125 2555 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:18:04.041672 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:18:04.041943 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:18:04.058000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:04.061808 systemd[1]: kubelet.service: Consumed 2.606s CPU time, 107.9M memory peak. Apr 20 19:18:04.066524 kernel: audit: type=1131 audit(1776712684.058:259): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:14.240205 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 17. Apr 20 19:18:14.369173 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:18:15.345775 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:18:15.344000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:15.362111 kernel: audit: type=1130 audit(1776712695.344:260): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:15.447923 (kubelet)[2611]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:18:16.464632 kubelet[2611]: E0420 19:18:16.463197 2611 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:18:16.488182 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:18:16.488441 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:18:16.489888 systemd[1]: kubelet.service: Consumed 1.441s CPU time, 109.9M memory peak. Apr 20 19:18:16.488000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:16.563196 kernel: audit: type=1131 audit(1776712696.488:261): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:19.226804 containerd[1648]: time="2026-04-20T19:18:19.226171824Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:18:19.247090 containerd[1648]: time="2026-04-20T19:18:19.238579288Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=23711450" Apr 20 19:18:19.278922 containerd[1648]: time="2026-04-20T19:18:19.277174332Z" level=info msg="ImageCreate event name:\"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:18:19.459789 containerd[1648]: time="2026-04-20T19:18:19.459329980Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 20 19:18:19.462508 containerd[1648]: time="2026-04-20T19:18:19.460680082Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"23716032\" in 36.563647156s" Apr 20 19:18:19.462508 containerd[1648]: time="2026-04-20T19:18:19.460712440Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\"" Apr 20 19:18:24.039954 systemd[1750]: Created slice background.slice - User Background Tasks Slice. Apr 20 19:18:24.041625 systemd[1750]: Starting systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories... Apr 20 19:18:24.252572 systemd[1750]: Finished systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories. Apr 20 19:18:26.738826 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 18. Apr 20 19:18:26.742469 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:18:28.398465 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:18:28.397000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:28.408612 kernel: audit: type=1130 audit(1776712708.397:262): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:28.477713 (kubelet)[2667]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 20 19:18:29.387773 kubelet[2667]: E0420 19:18:29.385956 2667 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 20 19:18:29.399095 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 20 19:18:29.399320 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 20 19:18:29.433000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:29.436167 systemd[1]: kubelet.service: Consumed 1.772s CPU time, 110.5M memory peak. Apr 20 19:18:29.441510 kernel: audit: type=1131 audit(1776712709.433:263): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:38.815512 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:18:38.814000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:38.815941 systemd[1]: kubelet.service: Consumed 1.772s CPU time, 110.5M memory peak. Apr 20 19:18:38.814000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:38.846870 kernel: audit: type=1130 audit(1776712718.814:264): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:38.847393 kernel: audit: type=1131 audit(1776712718.814:265): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:38.968903 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:18:39.547097 systemd[1]: Reload requested from client PID 2684 ('systemctl') (unit session-8.scope)... Apr 20 19:18:39.547290 systemd[1]: Reloading... Apr 20 19:18:39.841515 zram_generator::config[2738]: No configuration found. Apr 20 19:18:39.861731 systemd-ssh-generator[2732]: Failed to query local AF_VSOCK CID: Cannot assign requested address Apr 20 19:18:41.812850 (sd-exec-[2715]: /usr/lib/systemd/system-generators/systemd-ssh-generator failed with exit status 1. Apr 20 19:18:45.961108 systemd[1]: /usr/lib/systemd/system/update-engine.service:10: Support for option BlockIOWeight= has been removed and it is ignored Apr 20 19:18:50.852308 systemd[1]: Reloading finished in 11304 ms. Apr 20 19:18:50.989000 audit: BPF prog-id=39 op=LOAD Apr 20 19:18:50.991000 audit: BPF prog-id=30 op=UNLOAD Apr 20 19:18:50.997995 kernel: audit: type=1334 audit(1776712730.989:266): prog-id=39 op=LOAD Apr 20 19:18:50.992000 audit: BPF prog-id=40 op=LOAD Apr 20 19:18:50.998893 kernel: audit: type=1334 audit(1776712730.991:267): prog-id=30 op=UNLOAD Apr 20 19:18:50.998996 kernel: audit: type=1334 audit(1776712730.992:268): prog-id=40 op=LOAD Apr 20 19:18:50.992000 audit: BPF prog-id=41 op=LOAD Apr 20 19:18:51.002646 kernel: audit: type=1334 audit(1776712730.992:269): prog-id=41 op=LOAD Apr 20 19:18:50.993000 audit: BPF prog-id=31 op=UNLOAD Apr 20 19:18:51.002875 kernel: audit: type=1334 audit(1776712730.993:270): prog-id=31 op=UNLOAD Apr 20 19:18:50.993000 audit: BPF prog-id=32 op=UNLOAD Apr 20 19:18:51.000000 audit: BPF prog-id=42 op=LOAD Apr 20 19:18:51.010321 kernel: audit: type=1334 audit(1776712730.993:271): prog-id=32 op=UNLOAD Apr 20 19:18:51.000000 audit: BPF prog-id=35 op=UNLOAD Apr 20 19:18:51.011402 kernel: audit: type=1334 audit(1776712731.000:272): prog-id=42 op=LOAD Apr 20 19:18:51.011688 kernel: audit: type=1334 audit(1776712731.000:273): prog-id=35 op=UNLOAD Apr 20 19:18:51.001000 audit: BPF prog-id=43 op=LOAD Apr 20 19:18:51.014654 kernel: audit: type=1334 audit(1776712731.001:274): prog-id=43 op=LOAD Apr 20 19:18:51.001000 audit: BPF prog-id=26 op=UNLOAD Apr 20 19:18:51.015154 kernel: audit: type=1334 audit(1776712731.001:275): prog-id=26 op=UNLOAD Apr 20 19:18:51.001000 audit: BPF prog-id=44 op=LOAD Apr 20 19:18:51.001000 audit: BPF prog-id=25 op=UNLOAD Apr 20 19:18:51.005000 audit: BPF prog-id=45 op=LOAD Apr 20 19:18:51.005000 audit: BPF prog-id=36 op=UNLOAD Apr 20 19:18:51.005000 audit: BPF prog-id=46 op=LOAD Apr 20 19:18:51.005000 audit: BPF prog-id=47 op=LOAD Apr 20 19:18:51.005000 audit: BPF prog-id=37 op=UNLOAD Apr 20 19:18:51.005000 audit: BPF prog-id=38 op=UNLOAD Apr 20 19:18:51.006000 audit: BPF prog-id=48 op=LOAD Apr 20 19:18:51.006000 audit: BPF prog-id=49 op=LOAD Apr 20 19:18:51.006000 audit: BPF prog-id=33 op=UNLOAD Apr 20 19:18:51.006000 audit: BPF prog-id=34 op=UNLOAD Apr 20 19:18:51.009000 audit: BPF prog-id=50 op=LOAD Apr 20 19:18:51.009000 audit: BPF prog-id=27 op=UNLOAD Apr 20 19:18:51.009000 audit: BPF prog-id=51 op=LOAD Apr 20 19:18:51.009000 audit: BPF prog-id=52 op=LOAD Apr 20 19:18:51.009000 audit: BPF prog-id=28 op=UNLOAD Apr 20 19:18:51.009000 audit: BPF prog-id=29 op=UNLOAD Apr 20 19:18:51.015000 audit: BPF prog-id=53 op=LOAD Apr 20 19:18:51.015000 audit: BPF prog-id=22 op=UNLOAD Apr 20 19:18:51.015000 audit: BPF prog-id=54 op=LOAD Apr 20 19:18:51.015000 audit: BPF prog-id=55 op=LOAD Apr 20 19:18:51.015000 audit: BPF prog-id=23 op=UNLOAD Apr 20 19:18:51.015000 audit: BPF prog-id=24 op=UNLOAD Apr 20 19:18:51.016000 audit: BPF prog-id=56 op=LOAD Apr 20 19:18:51.016000 audit: BPF prog-id=19 op=UNLOAD Apr 20 19:18:51.016000 audit: BPF prog-id=57 op=LOAD Apr 20 19:18:51.016000 audit: BPF prog-id=58 op=LOAD Apr 20 19:18:51.016000 audit: BPF prog-id=20 op=UNLOAD Apr 20 19:18:51.016000 audit: BPF prog-id=21 op=UNLOAD Apr 20 19:18:51.162814 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 20 19:18:51.164925 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 20 19:18:51.166065 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:18:51.166197 systemd[1]: kubelet.service: Consumed 557ms CPU time, 98.6M memory peak. Apr 20 19:18:51.164000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Apr 20 19:18:51.235568 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:18:54.152551 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:18:54.151000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:18:54.179668 (kubelet)[2785]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 20 19:18:55.223781 kubelet[2785]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 20 19:18:55.223781 kubelet[2785]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 20 19:18:55.223781 kubelet[2785]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 20 19:18:55.265518 kubelet[2785]: I0420 19:18:55.226138 2785 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 20 19:18:56.588381 kubelet[2785]: I0420 19:18:56.587754 2785 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 20 19:18:56.588381 kubelet[2785]: I0420 19:18:56.587917 2785 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 20 19:18:56.621215 kubelet[2785]: I0420 19:18:56.603728 2785 server.go:956] "Client rotation is on, will bootstrap in background" Apr 20 19:18:57.046876 kubelet[2785]: I0420 19:18:57.046432 2785 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 20 19:18:57.046876 kubelet[2785]: E0420 19:18:57.046431 2785 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.18:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 20 19:18:57.266416 kubelet[2785]: I0420 19:18:57.265977 2785 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 20 19:18:57.679931 kubelet[2785]: I0420 19:18:57.679016 2785 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 20 19:18:57.712087 kubelet[2785]: I0420 19:18:57.692453 2785 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 20 19:18:57.712087 kubelet[2785]: I0420 19:18:57.694871 2785 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 20 19:18:57.712087 kubelet[2785]: I0420 19:18:57.708580 2785 topology_manager.go:138] "Creating topology manager with none policy" Apr 20 19:18:57.712087 kubelet[2785]: I0420 19:18:57.708786 2785 container_manager_linux.go:303] "Creating device plugin manager" Apr 20 19:18:57.713525 kubelet[2785]: I0420 19:18:57.712883 2785 state_mem.go:36] "Initialized new in-memory state store" Apr 20 19:18:57.761993 kubelet[2785]: I0420 19:18:57.754770 2785 kubelet.go:480] "Attempting to sync node with API server" Apr 20 19:18:57.771798 kubelet[2785]: I0420 19:18:57.768855 2785 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 20 19:18:57.771798 kubelet[2785]: I0420 19:18:57.771442 2785 kubelet.go:386] "Adding apiserver pod source" Apr 20 19:18:57.771798 kubelet[2785]: I0420 19:18:57.771665 2785 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 20 19:18:57.834023 kubelet[2785]: E0420 19:18:57.833360 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:18:57.834023 kubelet[2785]: E0420 19:18:57.833643 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:18:57.903972 kubelet[2785]: I0420 19:18:57.901829 2785 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.2.1" apiVersion="v1" Apr 20 19:18:57.911376 kubelet[2785]: I0420 19:18:57.911306 2785 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 20 19:18:57.915542 kubelet[2785]: W0420 19:18:57.914847 2785 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 20 19:18:58.080171 kubelet[2785]: I0420 19:18:58.079214 2785 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 20 19:18:58.111746 kubelet[2785]: I0420 19:18:58.083440 2785 server.go:1289] "Started kubelet" Apr 20 19:18:58.111746 kubelet[2785]: I0420 19:18:58.109069 2785 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 20 19:18:58.111746 kubelet[2785]: I0420 19:18:58.108775 2785 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 20 19:18:58.172456 kubelet[2785]: I0420 19:18:58.167356 2785 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 20 19:18:58.185191 kubelet[2785]: E0420 19:18:58.167307 2785 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.18:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.18:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18a826d2652bbeb6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:18:58.080014006 +0000 UTC m=+3.732389473,LastTimestamp:2026-04-20 19:18:58.080014006 +0000 UTC m=+3.732389473,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:18:58.262841 kubelet[2785]: I0420 19:18:58.259477 2785 server.go:317] "Adding debug handlers to kubelet server" Apr 20 19:18:58.262841 kubelet[2785]: I0420 19:18:58.260295 2785 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 20 19:18:58.278093 kubelet[2785]: E0420 19:18:58.273572 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.279954 kubelet[2785]: I0420 19:18:58.278922 2785 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 20 19:18:58.279954 kubelet[2785]: I0420 19:18:58.279500 2785 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 20 19:18:58.279954 kubelet[2785]: I0420 19:18:58.279796 2785 reconciler.go:26] "Reconciler: start to sync state" Apr 20 19:18:58.280964 kubelet[2785]: I0420 19:18:58.272020 2785 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 20 19:18:58.280964 kubelet[2785]: E0420 19:18:58.280616 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 20 19:18:58.280964 kubelet[2785]: E0420 19:18:58.280732 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" interval="200ms" Apr 20 19:18:58.381111 kubelet[2785]: E0420 19:18:58.378744 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.401752 kubelet[2785]: E0420 19:18:58.401271 2785 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 20 19:18:58.413445 kubelet[2785]: I0420 19:18:58.412842 2785 factory.go:223] Registration of the containerd container factory successfully Apr 20 19:18:58.413445 kubelet[2785]: I0420 19:18:58.413107 2785 factory.go:223] Registration of the systemd container factory successfully Apr 20 19:18:58.412000 audit[2805]: NETFILTER_CFG table=mangle:40 family=2 entries=2 op=nft_register_chain pid=2805 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.412000 audit[2805]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd7534dda0 a2=0 a3=0 items=0 ppid=2785 pid=2805 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.423418 kernel: kauditd_printk_skb: 32 callbacks suppressed Apr 20 19:18:58.423521 kubelet[2785]: I0420 19:18:58.413867 2785 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 20 19:18:58.423561 kernel: audit: type=1325 audit(1776712738.412:308): table=mangle:40 family=2 entries=2 op=nft_register_chain pid=2805 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.423580 kernel: audit: type=1300 audit(1776712738.412:308): arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd7534dda0 a2=0 a3=0 items=0 ppid=2785 pid=2805 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.412000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Apr 20 19:18:58.435724 kernel: audit: type=1327 audit(1776712738.412:308): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Apr 20 19:18:58.432000 audit[2806]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_chain pid=2806 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.432000 audit[2806]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd0c573110 a2=0 a3=0 items=0 ppid=2785 pid=2806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.432000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Apr 20 19:18:58.459108 kernel: audit: type=1325 audit(1776712738.432:309): table=filter:41 family=2 entries=1 op=nft_register_chain pid=2806 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.459289 kernel: audit: type=1300 audit(1776712738.432:309): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd0c573110 a2=0 a3=0 items=0 ppid=2785 pid=2806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.459310 kernel: audit: type=1327 audit(1776712738.432:309): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Apr 20 19:18:58.488068 kubelet[2785]: E0420 19:18:58.486797 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.524000 audit[2809]: NETFILTER_CFG table=filter:42 family=2 entries=2 op=nft_register_chain pid=2809 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.530571 kernel: audit: type=1325 audit(1776712738.524:310): table=filter:42 family=2 entries=2 op=nft_register_chain pid=2809 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.524000 audit[2809]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7fffc5cd4030 a2=0 a3=0 items=0 ppid=2785 pid=2809 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.537009 kubelet[2785]: E0420 19:18:58.531845 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" interval="400ms" Apr 20 19:18:58.524000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Apr 20 19:18:58.544012 kernel: audit: type=1300 audit(1776712738.524:310): arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7fffc5cd4030 a2=0 a3=0 items=0 ppid=2785 pid=2809 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.544329 kernel: audit: type=1327 audit(1776712738.524:310): proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Apr 20 19:18:58.572000 audit[2811]: NETFILTER_CFG table=filter:43 family=2 entries=2 op=nft_register_chain pid=2811 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.580029 kernel: audit: type=1325 audit(1776712738.572:311): table=filter:43 family=2 entries=2 op=nft_register_chain pid=2811 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.572000 audit[2811]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffc967c7970 a2=0 a3=0 items=0 ppid=2785 pid=2811 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.572000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Apr 20 19:18:58.587911 kubelet[2785]: E0420 19:18:58.587586 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.597000 audit[2816]: NETFILTER_CFG table=filter:44 family=2 entries=1 op=nft_register_rule pid=2816 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.597000 audit[2816]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7ffea2adcf90 a2=0 a3=0 items=0 ppid=2785 pid=2816 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.597000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 Apr 20 19:18:58.605347 kubelet[2785]: I0420 19:18:58.599608 2785 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 20 19:18:58.621000 audit[2818]: NETFILTER_CFG table=mangle:45 family=10 entries=2 op=nft_register_chain pid=2818 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:18:58.621000 audit[2818]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7fff6efaada0 a2=0 a3=0 items=0 ppid=2785 pid=2818 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.622000 audit[2817]: NETFILTER_CFG table=mangle:46 family=2 entries=1 op=nft_register_chain pid=2817 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.622000 audit[2817]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fffcd3a77c0 a2=0 a3=0 items=0 ppid=2785 pid=2817 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.622000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Apr 20 19:18:58.621000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Apr 20 19:18:58.737126 kubelet[2785]: I0420 19:18:58.633810 2785 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 20 19:18:58.737126 kubelet[2785]: I0420 19:18:58.638143 2785 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 20 19:18:58.737126 kubelet[2785]: I0420 19:18:58.638835 2785 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 20 19:18:58.737126 kubelet[2785]: I0420 19:18:58.647124 2785 kubelet.go:2436] "Starting kubelet main sync loop" Apr 20 19:18:58.737126 kubelet[2785]: E0420 19:18:58.648053 2785 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:18:58.738822 kubelet[2785]: E0420 19:18:58.738100 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.739209 kubelet[2785]: E0420 19:18:58.739147 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:18:58.753770 kubelet[2785]: E0420 19:18:58.753044 2785 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:18:58.754000 audit[2821]: NETFILTER_CFG table=mangle:47 family=10 entries=1 op=nft_register_chain pid=2821 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:18:58.754000 audit[2821]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fffea9ed840 a2=0 a3=0 items=0 ppid=2785 pid=2821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.754000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Apr 20 19:18:58.761000 audit[2820]: NETFILTER_CFG table=nat:48 family=2 entries=1 op=nft_register_chain pid=2820 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.761000 audit[2820]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffee51cd650 a2=0 a3=0 items=0 ppid=2785 pid=2820 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.761000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Apr 20 19:18:58.770729 kubelet[2785]: I0420 19:18:58.755779 2785 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 20 19:18:58.770729 kubelet[2785]: I0420 19:18:58.755791 2785 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 20 19:18:58.770729 kubelet[2785]: I0420 19:18:58.755898 2785 state_mem.go:36] "Initialized new in-memory state store" Apr 20 19:18:58.801000 audit[2823]: NETFILTER_CFG table=nat:49 family=10 entries=1 op=nft_register_chain pid=2823 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:18:58.801000 audit[2823]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0329d600 a2=0 a3=0 items=0 ppid=2785 pid=2823 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.801000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Apr 20 19:18:58.813436 kubelet[2785]: I0420 19:18:58.813406 2785 policy_none.go:49] "None policy: Start" Apr 20 19:18:58.813846 kubelet[2785]: I0420 19:18:58.813626 2785 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 20 19:18:58.813846 kubelet[2785]: I0420 19:18:58.813644 2785 state_mem.go:35] "Initializing new in-memory state store" Apr 20 19:18:58.812000 audit[2824]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_chain pid=2824 subj=system_u:system_r:kernel_t:s0 comm="iptables" Apr 20 19:18:58.812000 audit[2824]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcb67d86b0 a2=0 a3=0 items=0 ppid=2785 pid=2824 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.812000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Apr 20 19:18:58.824000 audit[2825]: NETFILTER_CFG table=filter:51 family=10 entries=1 op=nft_register_chain pid=2825 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Apr 20 19:18:58.824000 audit[2825]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe0cec9040 a2=0 a3=0 items=0 ppid=2785 pid=2825 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:18:58.824000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Apr 20 19:18:58.842296 kubelet[2785]: E0420 19:18:58.841918 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.935418 kubelet[2785]: E0420 19:18:58.934868 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" interval="800ms" Apr 20 19:18:58.943647 kubelet[2785]: E0420 19:18:58.943455 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:58.960623 kubelet[2785]: E0420 19:18:58.958945 2785 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:18:58.962745 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 20 19:18:59.046182 kubelet[2785]: E0420 19:18:59.045725 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:59.069478 kubelet[2785]: E0420 19:18:59.069376 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:18:59.129466 kubelet[2785]: E0420 19:18:59.129396 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:18:59.145357 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 20 19:18:59.171209 kubelet[2785]: E0420 19:18:59.150353 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:59.180514 kubelet[2785]: E0420 19:18:59.174746 2785 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.18:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 20 19:18:59.252288 kubelet[2785]: E0420 19:18:59.251805 2785 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 20 19:18:59.252886 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 20 19:18:59.304745 kubelet[2785]: E0420 19:18:59.304574 2785 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 20 19:18:59.305696 kubelet[2785]: I0420 19:18:59.305344 2785 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 20 19:18:59.305696 kubelet[2785]: I0420 19:18:59.305358 2785 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 20 19:18:59.313775 kubelet[2785]: I0420 19:18:59.313600 2785 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 20 19:18:59.323957 kubelet[2785]: E0420 19:18:59.323776 2785 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 20 19:18:59.323957 kubelet[2785]: E0420 19:18:59.323866 2785 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 20 19:18:59.426609 kubelet[2785]: I0420 19:18:59.426443 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:18:59.432429 kubelet[2785]: E0420 19:18:59.432308 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": dial tcp 10.0.0.18:6443: connect: connection refused" node="localhost" Apr 20 19:18:59.450728 kubelet[2785]: I0420 19:18:59.450189 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fde15e5d4357efe755cb792ce81eb1a9-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fde15e5d4357efe755cb792ce81eb1a9\") " pod="kube-system/kube-apiserver-localhost" Apr 20 19:18:59.450728 kubelet[2785]: I0420 19:18:59.450652 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fde15e5d4357efe755cb792ce81eb1a9-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fde15e5d4357efe755cb792ce81eb1a9\") " pod="kube-system/kube-apiserver-localhost" Apr 20 19:18:59.450728 kubelet[2785]: I0420 19:18:59.450668 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fde15e5d4357efe755cb792ce81eb1a9-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fde15e5d4357efe755cb792ce81eb1a9\") " pod="kube-system/kube-apiserver-localhost" Apr 20 19:18:59.489765 kubelet[2785]: E0420 19:18:59.489666 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 20 19:18:59.566470 kubelet[2785]: I0420 19:18:59.557549 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:18:59.566470 kubelet[2785]: I0420 19:18:59.564047 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:18:59.566470 kubelet[2785]: I0420 19:18:59.564895 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:18:59.566470 kubelet[2785]: I0420 19:18:59.565004 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:18:59.566470 kubelet[2785]: I0420 19:18:59.565017 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:18:59.578637 kubelet[2785]: I0420 19:18:59.565175 2785 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/33fee6ba1581201eda98a989140db110-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"33fee6ba1581201eda98a989140db110\") " pod="kube-system/kube-scheduler-localhost" Apr 20 19:18:59.579718 systemd[1]: Created slice kubepods-burstable-podfde15e5d4357efe755cb792ce81eb1a9.slice - libcontainer container kubepods-burstable-podfde15e5d4357efe755cb792ce81eb1a9.slice. Apr 20 19:18:59.654639 kubelet[2785]: I0420 19:18:59.654558 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:18:59.655517 kubelet[2785]: E0420 19:18:59.655445 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": dial tcp 10.0.0.18:6443: connect: connection refused" node="localhost" Apr 20 19:18:59.655585 kubelet[2785]: E0420 19:18:59.655470 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:18:59.664823 kubelet[2785]: E0420 19:18:59.662385 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:18:59.685593 containerd[1648]: time="2026-04-20T19:18:59.685402382Z" level=info msg="RunPodSandbox for name:\"kube-apiserver-localhost\" uid:\"fde15e5d4357efe755cb792ce81eb1a9\" namespace:\"kube-system\"" Apr 20 19:18:59.698442 systemd[1]: Created slice kubepods-burstable-pode9ca41790ae21be9f4cbd451ade0acec.slice - libcontainer container kubepods-burstable-pode9ca41790ae21be9f4cbd451ade0acec.slice. Apr 20 19:18:59.742292 kubelet[2785]: E0420 19:18:59.742182 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" interval="1.6s" Apr 20 19:18:59.771374 kubelet[2785]: E0420 19:18:59.770791 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:18:59.852417 kubelet[2785]: E0420 19:18:59.846689 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:18:59.861675 containerd[1648]: time="2026-04-20T19:18:59.861537257Z" level=info msg="RunPodSandbox for name:\"kube-controller-manager-localhost\" uid:\"e9ca41790ae21be9f4cbd451ade0acec\" namespace:\"kube-system\"" Apr 20 19:18:59.949903 systemd[1]: Created slice kubepods-burstable-pod33fee6ba1581201eda98a989140db110.slice - libcontainer container kubepods-burstable-pod33fee6ba1581201eda98a989140db110.slice. Apr 20 19:18:59.986035 kubelet[2785]: E0420 19:18:59.985759 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:19:00.005065 kubelet[2785]: E0420 19:19:00.004656 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:00.006874 kubelet[2785]: E0420 19:19:00.005672 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:00.007610 containerd[1648]: time="2026-04-20T19:19:00.007557112Z" level=info msg="RunPodSandbox for name:\"kube-scheduler-localhost\" uid:\"33fee6ba1581201eda98a989140db110\" namespace:\"kube-system\"" Apr 20 19:19:00.078893 kubelet[2785]: I0420 19:19:00.078773 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:19:00.080200 kubelet[2785]: E0420 19:19:00.079919 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": dial tcp 10.0.0.18:6443: connect: connection refused" node="localhost" Apr 20 19:19:00.918033 kubelet[2785]: I0420 19:19:00.917548 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:19:00.920670 kubelet[2785]: E0420 19:19:00.919574 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": dial tcp 10.0.0.18:6443: connect: connection refused" node="localhost" Apr 20 19:19:01.393209 kubelet[2785]: E0420 19:19:01.392963 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" interval="3.2s" Apr 20 19:19:01.422022 kubelet[2785]: E0420 19:19:01.421104 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:19:01.969998 kubelet[2785]: E0420 19:19:01.969676 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:19:02.455336 kubelet[2785]: E0420 19:19:02.450606 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 20 19:19:02.498636 kubelet[2785]: E0420 19:19:02.497978 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:19:02.547183 kubelet[2785]: I0420 19:19:02.543144 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:19:02.566442 kubelet[2785]: E0420 19:19:02.566374 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": dial tcp 10.0.0.18:6443: connect: connection refused" node="localhost" Apr 20 19:19:02.820126 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount606816220.mount: Deactivated successfully. Apr 20 19:19:03.180620 containerd[1648]: time="2026-04-20T19:19:03.164142414Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 20 19:19:03.182192 containerd[1648]: time="2026-04-20T19:19:03.181215707Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=316649" Apr 20 19:19:03.267305 containerd[1648]: time="2026-04-20T19:19:03.266737959Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 20 19:19:03.381411 kubelet[2785]: E0420 19:19:03.380440 2785 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.18:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 20 19:19:03.383065 containerd[1648]: time="2026-04-20T19:19:03.382979956Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 20 19:19:03.383591 containerd[1648]: time="2026-04-20T19:19:03.383538584Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=0" Apr 20 19:19:03.383779 containerd[1648]: time="2026-04-20T19:19:03.383577418Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 20 19:19:03.405649 containerd[1648]: time="2026-04-20T19:19:03.404976637Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=0" Apr 20 19:19:03.450562 containerd[1648]: time="2026-04-20T19:19:03.443409647Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 20 19:19:03.588007 containerd[1648]: time="2026-04-20T19:19:03.586882859Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 3.866120432s" Apr 20 19:19:03.719036 containerd[1648]: time="2026-04-20T19:19:03.711163274Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 3.69979045s" Apr 20 19:19:03.808803 containerd[1648]: time="2026-04-20T19:19:03.807683518Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 3.91556442s" Apr 20 19:19:04.443156 containerd[1648]: time="2026-04-20T19:19:04.442526177Z" level=info msg="connecting to shim f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a" address="unix:///run/containerd/s/3eeaa10a372a85df02c901b3a154e40b8957c2f16506966d3fde9602859768e2" namespace=k8s.io protocol=ttrpc version=3 Apr 20 19:19:04.524447 containerd[1648]: time="2026-04-20T19:19:04.523647138Z" level=info msg="connecting to shim 5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7" address="unix:///run/containerd/s/b93351f384226afa3532a45fb30e5855b56c808eda1352c27be1433f7918c805" namespace=k8s.io protocol=ttrpc version=3 Apr 20 19:19:04.687667 kubelet[2785]: E0420 19:19:04.687032 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" interval="6.4s" Apr 20 19:19:05.688716 containerd[1648]: time="2026-04-20T19:19:05.688301414Z" level=info msg="connecting to shim 61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8" address="unix:///run/containerd/s/69f68cf67abb641c73f5591b7c1ef32ac2d1779cbaa7f3b47e6001986c597532" namespace=k8s.io protocol=ttrpc version=3 Apr 20 19:19:05.915000 kubelet[2785]: E0420 19:19:05.914840 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 20 19:19:05.989751 kubelet[2785]: E0420 19:19:05.989186 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:19:06.051971 kubelet[2785]: I0420 19:19:06.050154 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:19:06.066774 kubelet[2785]: E0420 19:19:06.066526 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": dial tcp 10.0.0.18:6443: connect: connection refused" node="localhost" Apr 20 19:19:06.264795 systemd[1]: Started cri-containerd-5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7.scope - libcontainer container 5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7. Apr 20 19:19:06.298057 systemd[1]: Started cri-containerd-f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a.scope - libcontainer container f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a. Apr 20 19:19:06.440154 kubelet[2785]: E0420 19:19:06.439735 2785 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.18:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.18:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18a826d2652bbeb6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:18:58.080014006 +0000 UTC m=+3.732389473,LastTimestamp:2026-04-20 19:18:58.080014006 +0000 UTC m=+3.732389473,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:19:06.491914 systemd[1]: Started cri-containerd-61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8.scope - libcontainer container 61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8. Apr 20 19:19:06.538000 audit: BPF prog-id=59 op=LOAD Apr 20 19:19:06.540952 kernel: kauditd_printk_skb: 26 callbacks suppressed Apr 20 19:19:06.541003 kernel: audit: type=1334 audit(1776712746.538:320): prog-id=59 op=LOAD Apr 20 19:19:06.561000 audit: BPF prog-id=60 op=LOAD Apr 20 19:19:06.561000 audit[2870]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000194240 a2=98 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.561000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.645670 kernel: audit: type=1334 audit(1776712746.561:321): prog-id=60 op=LOAD Apr 20 19:19:06.645732 kernel: audit: type=1300 audit(1776712746.561:321): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000194240 a2=98 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.645744 kernel: audit: type=1327 audit(1776712746.561:321): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.580000 audit: BPF prog-id=60 op=UNLOAD Apr 20 19:19:06.580000 audit[2870]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.580000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.670992 kernel: audit: type=1334 audit(1776712746.580:322): prog-id=60 op=UNLOAD Apr 20 19:19:06.671020 kernel: audit: type=1300 audit(1776712746.580:322): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.671039 kernel: audit: type=1327 audit(1776712746.580:322): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.642000 audit: BPF prog-id=61 op=LOAD Apr 20 19:19:06.642000 audit[2870]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000194490 a2=98 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.642000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.644000 audit: BPF prog-id=62 op=LOAD Apr 20 19:19:06.644000 audit[2870]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000194220 a2=98 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.644000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.644000 audit: BPF prog-id=62 op=UNLOAD Apr 20 19:19:06.644000 audit[2870]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.644000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.644000 audit: BPF prog-id=61 op=UNLOAD Apr 20 19:19:06.644000 audit[2870]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.644000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.681000 audit: BPF prog-id=63 op=LOAD Apr 20 19:19:06.668000 audit: BPF prog-id=64 op=LOAD Apr 20 19:19:06.694000 audit: BPF prog-id=65 op=LOAD Apr 20 19:19:06.694000 audit[2868]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a240 a2=98 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.694000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.694000 audit: BPF prog-id=65 op=UNLOAD Apr 20 19:19:06.694000 audit[2868]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.694000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.694000 audit: BPF prog-id=66 op=LOAD Apr 20 19:19:06.668000 audit[2870]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001946f0 a2=98 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.668000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.694000 audit[2868]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a490 a2=98 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.694000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.706407 kernel: audit: type=1334 audit(1776712746.642:323): prog-id=61 op=LOAD Apr 20 19:19:06.706433 kernel: audit: type=1300 audit(1776712746.642:323): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000194490 a2=98 a3=0 items=0 ppid=2849 pid=2870 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.706446 kernel: audit: type=1327 audit(1776712746.642:323): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562386365643635636235643335313737316165643963386637633432 Apr 20 19:19:06.704000 audit: BPF prog-id=67 op=LOAD Apr 20 19:19:06.704000 audit[2868]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a220 a2=98 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.704000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.706000 audit: BPF prog-id=67 op=UNLOAD Apr 20 19:19:06.706000 audit[2868]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.706000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.706000 audit: BPF prog-id=66 op=UNLOAD Apr 20 19:19:06.706000 audit[2868]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.706000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.706000 audit: BPF prog-id=68 op=LOAD Apr 20 19:19:06.706000 audit[2868]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6f0 a2=98 a3=0 items=0 ppid=2845 pid=2868 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.706000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6637646531316438383831653665333038613365643933643933393636 Apr 20 19:19:06.863000 audit: BPF prog-id=69 op=LOAD Apr 20 19:19:06.898000 audit: BPF prog-id=70 op=LOAD Apr 20 19:19:06.898000 audit[2913]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000198240 a2=98 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.898000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:06.903000 audit: BPF prog-id=70 op=UNLOAD Apr 20 19:19:06.903000 audit[2913]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.903000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:06.904000 audit: BPF prog-id=71 op=LOAD Apr 20 19:19:06.904000 audit[2913]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000198490 a2=98 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.904000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:06.904000 audit: BPF prog-id=72 op=LOAD Apr 20 19:19:06.904000 audit[2913]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000198220 a2=98 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.904000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:06.905000 audit: BPF prog-id=72 op=UNLOAD Apr 20 19:19:06.905000 audit[2913]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.905000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:06.906000 audit: BPF prog-id=71 op=UNLOAD Apr 20 19:19:06.906000 audit[2913]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.906000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:06.906000 audit: BPF prog-id=73 op=LOAD Apr 20 19:19:06.906000 audit[2913]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001986f0 a2=98 a3=0 items=0 ppid=2892 pid=2913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:06.906000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3631613838353062613738376437613535376238623162313838373134 Apr 20 19:19:07.148449 containerd[1648]: time="2026-04-20T19:19:07.148168387Z" level=info msg="RunPodSandbox for name:\"kube-apiserver-localhost\" uid:\"fde15e5d4357efe755cb792ce81eb1a9\" namespace:\"kube-system\" returns sandbox id \"5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7\"" Apr 20 19:19:07.148449 containerd[1648]: time="2026-04-20T19:19:07.148193168Z" level=info msg="RunPodSandbox for name:\"kube-scheduler-localhost\" uid:\"33fee6ba1581201eda98a989140db110\" namespace:\"kube-system\" returns sandbox id \"f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a\"" Apr 20 19:19:07.232049 kubelet[2785]: E0420 19:19:07.231778 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:07.238634 kubelet[2785]: E0420 19:19:07.233202 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:07.342385 containerd[1648]: time="2026-04-20T19:19:07.341887928Z" level=info msg="RunPodSandbox for name:\"kube-controller-manager-localhost\" uid:\"e9ca41790ae21be9f4cbd451ade0acec\" namespace:\"kube-system\" returns sandbox id \"61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8\"" Apr 20 19:19:07.372688 containerd[1648]: time="2026-04-20T19:19:07.371779076Z" level=info msg="CreateContainer within sandbox \"5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7\" for container name:\"kube-apiserver\"" Apr 20 19:19:07.448920 containerd[1648]: time="2026-04-20T19:19:07.371801438Z" level=info msg="CreateContainer within sandbox \"f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a\" for container name:\"kube-scheduler\"" Apr 20 19:19:07.473822 kubelet[2785]: E0420 19:19:07.473095 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:07.623767 containerd[1648]: time="2026-04-20T19:19:07.622633055Z" level=info msg="CreateContainer within sandbox \"61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8\" for container name:\"kube-controller-manager\"" Apr 20 19:19:07.634312 containerd[1648]: time="2026-04-20T19:19:07.633766887Z" level=info msg="Container 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7: CDI devices from CRI Config.CDIDevices: []" Apr 20 19:19:07.634312 containerd[1648]: time="2026-04-20T19:19:07.633815032Z" level=info msg="Container 5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a: CDI devices from CRI Config.CDIDevices: []" Apr 20 19:19:08.078786 kubelet[2785]: E0420 19:19:08.078037 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:19:08.171943 containerd[1648]: time="2026-04-20T19:19:08.170706844Z" level=info msg="CreateContainer within sandbox \"5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7\" for name:\"kube-apiserver\" returns container id \"5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a\"" Apr 20 19:19:08.186590 containerd[1648]: time="2026-04-20T19:19:08.186174670Z" level=info msg="CreateContainer within sandbox \"f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a\" for name:\"kube-scheduler\" returns container id \"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\"" Apr 20 19:19:08.195030 containerd[1648]: time="2026-04-20T19:19:08.186879565Z" level=info msg="StartContainer for \"5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a\"" Apr 20 19:19:08.277634 containerd[1648]: time="2026-04-20T19:19:08.277001336Z" level=info msg="StartContainer for \"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\"" Apr 20 19:19:08.289056 containerd[1648]: time="2026-04-20T19:19:08.287657975Z" level=info msg="Container edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00: CDI devices from CRI Config.CDIDevices: []" Apr 20 19:19:08.438965 containerd[1648]: time="2026-04-20T19:19:08.420443472Z" level=info msg="connecting to shim 5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a" address="unix:///run/containerd/s/b93351f384226afa3532a45fb30e5855b56c808eda1352c27be1433f7918c805" protocol=ttrpc version=3 Apr 20 19:19:08.522747 kubelet[2785]: E0420 19:19:08.521195 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.18:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:19:08.541544 containerd[1648]: time="2026-04-20T19:19:08.541036060Z" level=info msg="connecting to shim 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" address="unix:///run/containerd/s/3eeaa10a372a85df02c901b3a154e40b8957c2f16506966d3fde9602859768e2" protocol=ttrpc version=3 Apr 20 19:19:09.078296 containerd[1648]: time="2026-04-20T19:19:09.077910126Z" level=info msg="CreateContainer within sandbox \"61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8\" for name:\"kube-controller-manager\" returns container id \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\"" Apr 20 19:19:09.080958 containerd[1648]: time="2026-04-20T19:19:09.080896300Z" level=info msg="StartContainer for \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\"" Apr 20 19:19:09.083033 containerd[1648]: time="2026-04-20T19:19:09.082975742Z" level=info msg="connecting to shim edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" address="unix:///run/containerd/s/69f68cf67abb641c73f5591b7c1ef32ac2d1779cbaa7f3b47e6001986c597532" protocol=ttrpc version=3 Apr 20 19:19:09.105802 systemd[1]: Started cri-containerd-5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a.scope - libcontainer container 5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a. Apr 20 19:19:09.123688 systemd[1]: Started cri-containerd-78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7.scope - libcontainer container 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7. Apr 20 19:19:09.304000 audit: BPF prog-id=74 op=LOAD Apr 20 19:19:09.316000 audit: BPF prog-id=75 op=LOAD Apr 20 19:19:09.316000 audit[2974]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106240 a2=98 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.316000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.356000 audit: BPF prog-id=75 op=UNLOAD Apr 20 19:19:09.356000 audit[2974]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.356000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.370832 kubelet[2785]: E0420 19:19:09.354176 2785 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 20 19:19:09.372000 audit: BPF prog-id=76 op=LOAD Apr 20 19:19:09.372000 audit[2974]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106490 a2=98 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.372000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.373000 audit: BPF prog-id=77 op=LOAD Apr 20 19:19:09.373000 audit[2974]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000106220 a2=98 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.373000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.373000 audit: BPF prog-id=77 op=UNLOAD Apr 20 19:19:09.373000 audit[2974]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.373000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.373000 audit: BPF prog-id=76 op=UNLOAD Apr 20 19:19:09.373000 audit[2974]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.373000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.373000 audit: BPF prog-id=78 op=LOAD Apr 20 19:19:09.373000 audit: BPF prog-id=79 op=LOAD Apr 20 19:19:09.373000 audit[2974]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001066f0 a2=98 a3=0 items=0 ppid=2849 pid=2974 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.373000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3563336162383963613964376238613039303966316233613735646363 Apr 20 19:19:09.382000 audit: BPF prog-id=80 op=LOAD Apr 20 19:19:09.382000 audit[2975]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000170240 a2=98 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.382000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.433000 audit: BPF prog-id=80 op=UNLOAD Apr 20 19:19:09.433000 audit[2975]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.433000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.435000 audit: BPF prog-id=81 op=LOAD Apr 20 19:19:09.435000 audit[2975]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000170490 a2=98 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.435000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.436000 audit: BPF prog-id=82 op=LOAD Apr 20 19:19:09.436000 audit[2975]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000170220 a2=98 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.436000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.436000 audit: BPF prog-id=82 op=UNLOAD Apr 20 19:19:09.436000 audit[2975]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.436000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.436000 audit: BPF prog-id=81 op=UNLOAD Apr 20 19:19:09.436000 audit[2975]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.436000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.436000 audit: BPF prog-id=83 op=LOAD Apr 20 19:19:09.436000 audit[2975]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001706f0 a2=98 a3=0 items=0 ppid=2845 pid=2975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.436000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3738653061616638333831333838656233613931623235376133626130 Apr 20 19:19:09.487877 systemd[1]: Started cri-containerd-edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00.scope - libcontainer container edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00. Apr 20 19:19:09.656000 audit: BPF prog-id=84 op=LOAD Apr 20 19:19:09.701000 audit: BPF prog-id=85 op=LOAD Apr 20 19:19:09.701000 audit[2997]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a240 a2=98 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.701000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:09.703000 audit: BPF prog-id=85 op=UNLOAD Apr 20 19:19:09.703000 audit[2997]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.703000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:09.703000 audit: BPF prog-id=86 op=LOAD Apr 20 19:19:09.703000 audit[2997]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a490 a2=98 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.703000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:09.703000 audit: BPF prog-id=87 op=LOAD Apr 20 19:19:09.703000 audit[2997]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a220 a2=98 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.703000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:09.703000 audit: BPF prog-id=87 op=UNLOAD Apr 20 19:19:09.703000 audit[2997]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.703000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:09.703000 audit: BPF prog-id=86 op=UNLOAD Apr 20 19:19:09.703000 audit[2997]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.703000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:09.704000 audit: BPF prog-id=88 op=LOAD Apr 20 19:19:09.704000 audit[2997]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6f0 a2=98 a3=0 items=0 ppid=2892 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 20 19:19:09.704000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6564656561303764353133356263656365383535386366643862646539 Apr 20 19:19:10.294826 containerd[1648]: time="2026-04-20T19:19:10.286881590Z" level=info msg="StartContainer for \"5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a\" returns successfully" Apr 20 19:19:11.134096 containerd[1648]: time="2026-04-20T19:19:11.131794948Z" level=info msg="StartContainer for \"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" returns successfully" Apr 20 19:19:11.563717 containerd[1648]: time="2026-04-20T19:19:11.518135099Z" level=info msg="StartContainer for \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" returns successfully" Apr 20 19:19:12.436322 kubelet[2785]: E0420 19:19:12.425750 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:12.622355 kubelet[2785]: E0420 19:19:12.617683 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:12.948068 kubelet[2785]: I0420 19:19:12.946582 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:19:13.373211 kubelet[2785]: E0420 19:19:13.372609 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:13.386434 kubelet[2785]: E0420 19:19:13.383951 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:15.481094 kubelet[2785]: E0420 19:19:15.480090 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:15.497291 kubelet[2785]: E0420 19:19:15.480101 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:15.497291 kubelet[2785]: E0420 19:19:15.496447 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:15.497291 kubelet[2785]: E0420 19:19:15.496482 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:15.497291 kubelet[2785]: E0420 19:19:15.496989 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:15.497291 kubelet[2785]: E0420 19:19:15.497111 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:16.556273 kubelet[2785]: E0420 19:19:16.552387 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:16.596287 kubelet[2785]: E0420 19:19:16.593567 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:16.596287 kubelet[2785]: E0420 19:19:16.593876 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:16.596287 kubelet[2785]: E0420 19:19:16.593883 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:16.596287 kubelet[2785]: E0420 19:19:16.594095 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:16.596287 kubelet[2785]: E0420 19:19:16.594143 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:17.819595 kubelet[2785]: E0420 19:19:17.802899 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:17.896905 kubelet[2785]: E0420 19:19:17.874878 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:19.391732 kubelet[2785]: E0420 19:19:19.390864 2785 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 20 19:19:21.155729 kubelet[2785]: E0420 19:19:21.150438 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="7s" Apr 20 19:19:21.713749 kubelet[2785]: E0420 19:19:21.713122 2785 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.18:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 20 19:19:23.261379 kubelet[2785]: E0420 19:19:23.260698 2785 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.18:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Apr 20 19:19:23.861613 kubelet[2785]: E0420 19:19:23.860823 2785 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:19:25.597533 kubelet[2785]: I0420 19:19:25.567816 2785 apiserver.go:52] "Watching apiserver" Apr 20 19:19:25.722879 kubelet[2785]: E0420 19:19:25.716521 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:25.723590 kubelet[2785]: E0420 19:19:25.723536 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:25.760749 kubelet[2785]: E0420 19:19:25.743066 2785 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18a826d2652bbeb6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:18:58.080014006 +0000 UTC m=+3.732389473,LastTimestamp:2026-04-20 19:18:58.080014006 +0000 UTC m=+3.732389473,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:19:25.914305 kubelet[2785]: I0420 19:19:25.904859 2785 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 20 19:19:26.443506 kubelet[2785]: E0420 19:19:26.442605 2785 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.18a826d2784f03fa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:18:58.401092602 +0000 UTC m=+4.053468073,LastTimestamp:2026-04-20 19:18:58.401092602 +0000 UTC m=+4.053468073,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:19:27.127912 kubelet[2785]: E0420 19:19:27.125631 2785 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Apr 20 19:19:27.270855 kubelet[2785]: E0420 19:19:27.164459 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:27.724741 kubelet[2785]: E0420 19:19:27.723033 2785 csi_plugin.go:397] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "localhost" not found Apr 20 19:19:29.096355 kubelet[2785]: E0420 19:19:29.094810 2785 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Apr 20 19:19:29.446294 kubelet[2785]: E0420 19:19:29.444774 2785 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 20 19:19:29.731610 kubelet[2785]: E0420 19:19:29.724324 2785 csi_plugin.go:397] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "localhost" not found Apr 20 19:19:30.722259 kubelet[2785]: I0420 19:19:30.720937 2785 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:19:30.794094 kubelet[2785]: I0420 19:19:30.778319 2785 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Apr 20 19:19:30.794094 kubelet[2785]: E0420 19:19:30.785595 2785 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Apr 20 19:19:30.921910 kubelet[2785]: I0420 19:19:30.882121 2785 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Apr 20 19:19:31.277514 kubelet[2785]: I0420 19:19:31.263535 2785 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Apr 20 19:19:31.497672 kubelet[2785]: I0420 19:19:31.497421 2785 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Apr 20 19:19:31.502053 kubelet[2785]: E0420 19:19:31.497478 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:31.506908 kubelet[2785]: E0420 19:19:31.506843 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:31.596848 kubelet[2785]: E0420 19:19:31.588580 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:19:36.512184 kubelet[2785]: E0420 19:19:36.511525 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.839s" Apr 20 19:19:40.741980 kubelet[2785]: I0420 19:19:40.734971 2785 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=9.734952969 podStartE2EDuration="9.734952969s" podCreationTimestamp="2026-04-20 19:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-20 19:19:40.478156034 +0000 UTC m=+46.130531512" watchObservedRunningTime="2026-04-20 19:19:40.734952969 +0000 UTC m=+46.387328454" Apr 20 19:19:42.130772 kubelet[2785]: I0420 19:19:42.128555 2785 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=11.127905169 podStartE2EDuration="11.127905169s" podCreationTimestamp="2026-04-20 19:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-20 19:19:42.120934938 +0000 UTC m=+47.773310410" watchObservedRunningTime="2026-04-20 19:19:42.127905169 +0000 UTC m=+47.780280637" Apr 20 19:19:42.294094 kubelet[2785]: E0420 19:19:42.153900 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.486s" Apr 20 19:19:42.294094 kubelet[2785]: I0420 19:19:42.163170 2785 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=11.162409302 podStartE2EDuration="11.162409302s" podCreationTimestamp="2026-04-20 19:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-20 19:19:41.494187755 +0000 UTC m=+47.146563235" watchObservedRunningTime="2026-04-20 19:19:42.162409302 +0000 UTC m=+47.814784780" Apr 20 19:19:55.670455 kubelet[2785]: E0420 19:19:55.655723 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="8.911s" Apr 20 19:20:01.292503 systemd[1]: cri-containerd-edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00.scope: Deactivated successfully. Apr 20 19:20:01.431000 audit: BPF prog-id=84 op=UNLOAD Apr 20 19:20:01.431000 audit: BPF prog-id=88 op=UNLOAD Apr 20 19:20:01.587504 kernel: kauditd_printk_skb: 122 callbacks suppressed Apr 20 19:20:01.436026 systemd[1]: cri-containerd-edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00.scope: Consumed 2.886s CPU time, 20.1M memory peak. Apr 20 19:20:01.602945 kernel: audit: type=1334 audit(1776712801.431:368): prog-id=84 op=UNLOAD Apr 20 19:20:01.607037 kernel: audit: type=1334 audit(1776712801.431:369): prog-id=88 op=UNLOAD Apr 20 19:20:12.280160 containerd[1648]: time="2026-04-20T19:20:12.261588762Z" level=info msg="received container exit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:20:22.447418 kubelet[2785]: E0420 19:20:22.322057 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="26.04s" Apr 20 19:20:29.048364 containerd[1648]: time="2026-04-20T19:20:29.032414306Z" level=error msg="ttrpc: received message on inactive stream" stream=27 Apr 20 19:20:29.409057 containerd[1648]: time="2026-04-20T19:20:29.386707787Z" level=error msg="ttrpc: received message on inactive stream" stream=29 Apr 20 19:20:30.069048 containerd[1648]: time="2026-04-20T19:20:29.966446498Z" level=error msg="failed to handle container TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" error="failed to stop container: context deadline exceeded" Apr 20 19:20:32.638045 containerd[1648]: time="2026-04-20T19:20:32.536077902Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:20:40.146507 kubelet[2785]: E0420 19:20:39.996178 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="16.963s" Apr 20 19:20:43.028093 kubelet[2785]: E0420 19:20:42.894627 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:20:43.403912 containerd[1648]: time="2026-04-20T19:20:43.170737772Z" level=error msg="failed to drain init process edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:20:43.784200 containerd[1648]: time="2026-04-20T19:20:43.179069840Z" level=error msg="failed to delete task" error="context deadline exceeded" id=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 Apr 20 19:20:45.060848 containerd[1648]: time="2026-04-20T19:20:44.983833738Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:20:45.400857 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00-rootfs.mount: Deactivated successfully. Apr 20 19:20:45.817805 containerd[1648]: time="2026-04-20T19:20:45.751152824Z" level=error msg="ttrpc: received message on inactive stream" stream=43 Apr 20 19:20:46.360275 kubelet[2785]: E0420 19:20:46.350830 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="6.207s" Apr 20 19:20:47.903123 kubelet[2785]: E0420 19:20:47.884022 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:20:48.143699 containerd[1648]: time="2026-04-20T19:20:47.989055664Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:20:49.315942 kubelet[2785]: E0420 19:20:49.285926 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:21:00.969098 containerd[1648]: time="2026-04-20T19:20:59.821712802Z" level=error msg="ttrpc: received message on inactive stream" stream=51 Apr 20 19:21:01.354595 containerd[1648]: time="2026-04-20T19:21:00.536884461Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:21:01.465968 containerd[1648]: time="2026-04-20T19:21:01.353743817Z" level=error msg="ttrpc: received message on inactive stream" stream=49 Apr 20 19:21:07.960495 containerd[1648]: time="2026-04-20T19:21:07.938190630Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:21:12.278520 systemd[1]: Starting systemd-tmpfiles-clean.service - Cleanup of Temporary Directories... Apr 20 19:21:15.413568 kubelet[2785]: E0420 19:21:15.142582 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="28.175s" Apr 20 19:21:28.771044 systemd-tmpfiles[3101]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Apr 20 19:21:28.808015 systemd-tmpfiles[3101]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Apr 20 19:21:28.933969 containerd[1648]: time="2026-04-20T19:21:26.178200695Z" level=error msg="ttrpc: received message on inactive stream" stream=31 Apr 20 19:21:29.162667 systemd-tmpfiles[3101]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 20 19:21:32.044802 systemd-tmpfiles[3101]: ACLs are not supported, ignoring. Apr 20 19:21:32.338031 systemd-tmpfiles[3101]: ACLs are not supported, ignoring. Apr 20 19:21:33.266777 systemd-tmpfiles[3101]: Detected autofs mount point /boot during canonicalization of boot. Apr 20 19:21:33.271789 systemd-tmpfiles[3101]: Skipping /boot Apr 20 19:21:34.420199 containerd[1648]: time="2026-04-20T19:21:30.128958676Z" level=error msg="get state for 61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8" error="context deadline exceeded" Apr 20 19:21:34.647588 systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully. Apr 20 19:21:34.959726 systemd[1]: Finished systemd-tmpfiles-clean.service - Cleanup of Temporary Directories. Apr 20 19:21:35.167000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-clean comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:21:35.182000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-clean comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:21:35.199828 systemd[1]: systemd-tmpfiles-clean.service: Consumed 5.931s CPU time, 4.6M memory peak. Apr 20 19:21:36.364759 kernel: audit: type=1130 audit(1776712895.167:370): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-clean comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:21:36.559099 kernel: audit: type=1131 audit(1776712895.182:371): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-clean comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:21:38.882495 containerd[1648]: time="2026-04-20T19:21:34.417752247Z" level=warning msg="unknown status" status=0 Apr 20 19:21:49.182800 containerd[1648]: time="2026-04-20T19:21:48.447088564Z" level=error msg="ttrpc: received message on inactive stream" stream=57 Apr 20 19:21:50.895365 containerd[1648]: time="2026-04-20T19:21:50.882833952Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:21:51.549975 containerd[1648]: time="2026-04-20T19:21:50.955094335Z" level=error msg="ttrpc: received message on inactive stream" stream=59 Apr 20 19:21:53.263965 kubelet[2785]: E0420 19:21:52.940757 2785 kubelet_node_status.go:460] "Node not becoming ready in time after startup" Apr 20 19:21:56.033492 kubelet[2785]: E0420 19:21:56.024876 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="40.365s" Apr 20 19:21:56.537500 kubelet[2785]: E0420 19:21:56.537135 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:21:57.525040 kubelet[2785]: E0420 19:21:57.524579 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:21:59.795390 containerd[1648]: time="2026-04-20T19:21:59.793417265Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:22:05.341110 kubelet[2785]: E0420 19:22:05.335670 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:22:11.175435 containerd[1648]: time="2026-04-20T19:22:11.137131509Z" level=error msg="ttrpc: received message on inactive stream" stream=33 Apr 20 19:22:12.661837 containerd[1648]: time="2026-04-20T19:22:12.620095075Z" level=error msg="get state for 61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8" error="context deadline exceeded" Apr 20 19:22:13.039010 containerd[1648]: time="2026-04-20T19:22:12.623121658Z" level=warning msg="unknown status" status=0 Apr 20 19:22:13.209590 systemd[1]: Reload requested from client PID 3107 ('systemctl') (unit session-8.scope)... Apr 20 19:22:13.232278 systemd[1]: Reloading... Apr 20 19:22:20.338725 kubelet[2785]: E0420 19:22:20.323208 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:22:27.780674 containerd[1648]: time="2026-04-20T19:22:24.555130187Z" level=error msg="ttrpc: received message on inactive stream" stream=67 Apr 20 19:22:30.806788 containerd[1648]: time="2026-04-20T19:22:30.806171229Z" level=error msg="ttrpc: received message on inactive stream" stream=69 Apr 20 19:22:48.849894 containerd[1648]: time="2026-04-20T19:22:40.304941156Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:22:49.674654 update_engine[1624]: I20260420 19:22:48.945523 1624 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Apr 20 19:22:49.674654 update_engine[1624]: I20260420 19:22:49.046941 1624 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Apr 20 19:22:49.674654 update_engine[1624]: I20260420 19:22:49.630829 1624 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Apr 20 19:22:50.457749 update_engine[1624]: I20260420 19:22:50.444545 1624 omaha_request_params.cc:62] Current group set to alpha Apr 20 19:22:50.626645 update_engine[1624]: I20260420 19:22:50.598814 1624 update_attempter.cc:499] Already updated boot flags. Skipping. Apr 20 19:22:50.626645 update_engine[1624]: I20260420 19:22:50.603184 1624 update_attempter.cc:643] Scheduling an action processor start. Apr 20 19:22:51.112219 update_engine[1624]: I20260420 19:22:50.647033 1624 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 20 19:22:51.112219 update_engine[1624]: I20260420 19:22:50.825119 1624 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Apr 20 19:22:51.112219 update_engine[1624]: I20260420 19:22:50.954401 1624 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 20 19:22:51.112219 update_engine[1624]: I20260420 19:22:50.954840 1624 omaha_request_action.cc:272] Request: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: Apr 20 19:22:51.112219 update_engine[1624]: I20260420 19:22:50.954901 1624 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 20 19:22:53.106858 update_engine[1624]: I20260420 19:22:52.161839 1624 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 20 19:22:53.106858 update_engine[1624]: I20260420 19:22:52.721072 1624 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 20 19:22:53.106858 update_engine[1624]: E20260420 19:22:53.039933 1624 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Apr 20 19:22:53.697691 zram_generator::config[3162]: No configuration found. Apr 20 19:22:53.708854 update_engine[1624]: I20260420 19:22:53.221753 1624 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Apr 20 19:22:55.629795 kubelet[2785]: E0420 19:22:55.623509 2785 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:23:01.564920 locksmithd[1723]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Apr 20 19:23:02.187375 systemd-ssh-generator[3158]: Failed to query local AF_VSOCK CID: Cannot assign requested address Apr 20 19:23:03.826290 update_engine[1624]: I20260420 19:23:03.818580 1624 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 20 19:23:04.195851 update_engine[1624]: I20260420 19:23:04.034834 1624 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 20 19:23:04.615124 update_engine[1624]: I20260420 19:23:04.555773 1624 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 20 19:23:04.717559 update_engine[1624]: E20260420 19:23:04.656978 1624 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Apr 20 19:23:04.808712 update_engine[1624]: I20260420 19:23:04.802404 1624 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Apr 20 19:23:05.888545 kubelet[2785]: E0420 19:23:05.755876 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:23:08.193900 kubelet[2785]: E0420 19:23:08.075779 2785 controller.go:195] "Failed to update lease" err="Operation cannot be fulfilled on leases.coordination.k8s.io \"localhost\": the object has been modified; please apply your changes to the latest version and try again" Apr 20 19:23:08.941906 containerd[1648]: time="2026-04-20T19:23:07.630810848Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:23:09.708950 containerd[1648]: time="2026-04-20T19:23:07.641031849Z" level=error msg="post event" error="context deadline exceeded" Apr 20 19:23:10.058614 containerd[1648]: time="2026-04-20T19:23:10.039366129Z" level=info msg="received container exit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:23:10.655767 containerd[1648]: time="2026-04-20T19:23:09.475852670Z" level=error msg="ttrpc: received message on inactive stream" stream=9 Apr 20 19:23:10.903510 kubelet[2785]: E0420 19:23:08.780682 2785 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1m12.546s" Apr 20 19:23:14.850908 update_engine[1624]: I20260420 19:23:14.825711 1624 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 20 19:23:14.850908 update_engine[1624]: I20260420 19:23:14.838640 1624 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 20 19:23:15.674665 update_engine[1624]: I20260420 19:23:14.953065 1624 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 20 19:23:15.674665 update_engine[1624]: E20260420 19:23:15.577178 1624 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Apr 20 19:23:16.248136 containerd[1648]: time="2026-04-20T19:23:14.386741804Z" level=error msg="ttrpc: received message on inactive stream" stream=11 Apr 20 19:23:16.248136 containerd[1648]: time="2026-04-20T19:23:14.486308641Z" level=error msg="forward event" error="context deadline exceeded" Apr 20 19:23:16.912865 update_engine[1624]: I20260420 19:23:15.680567 1624 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Apr 20 19:23:15.683648 (sd-exec-[3139]: /usr/lib/systemd/system-generators/systemd-ssh-generator failed with exit status 1. Apr 20 19:23:17.345716 containerd[1648]: time="2026-04-20T19:23:15.948057436Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:23:17.345716 containerd[1648]: time="2026-04-20T19:23:16.579957327Z" level=warning msg="unknown status" status=0 Apr 20 19:23:20.553109 containerd[1648]: time="2026-04-20T19:23:19.653355089Z" level=error msg="ttrpc: received message on inactive stream" stream=25 Apr 20 19:23:21.813554 containerd[1648]: time="2026-04-20T19:23:21.554298034Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:23:22.006514 containerd[1648]: time="2026-04-20T19:23:21.955764009Z" level=warning msg="unknown status" status=0 Apr 20 19:23:22.020945 containerd[1648]: time="2026-04-20T19:23:21.526743661Z" level=error msg="ttrpc: received message on inactive stream" stream=71 Apr 20 19:23:25.843044 update_engine[1624]: I20260420 19:23:25.823590 1624 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 20 19:23:26.268347 update_engine[1624]: I20260420 19:23:25.941764 1624 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 20 19:23:26.387422 kubelet[2785]: E0420 19:23:25.143936 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.638640 1624 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 20 19:23:27.324194 update_engine[1624]: E20260420 19:23:26.804353 1624 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.805001 1624 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.805045 1624 omaha_request_action.cc:617] Omaha request response: Apr 20 19:23:27.324194 update_engine[1624]: E20260420 19:23:26.805507 1624 omaha_request_action.cc:636] Omaha request network transfer failed. Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.805809 1624 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.805819 1624 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.805825 1624 update_attempter.cc:306] Processing Done. Apr 20 19:23:27.324194 update_engine[1624]: E20260420 19:23:26.880788 1624 update_attempter.cc:619] Update failed. Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.918886 1624 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.935503 1624 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:26.948385 1624 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:27.109715 1624 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:27.154579 1624 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 20 19:23:27.324194 update_engine[1624]: I20260420 19:23:27.156736 1624 omaha_request_action.cc:272] Request: Apr 20 19:23:27.324194 update_engine[1624]: Apr 20 19:23:27.324194 update_engine[1624]: Apr 20 19:23:28.137498 locksmithd[1723]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Apr 20 19:23:28.137498 locksmithd[1723]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Apr 20 19:23:28.259383 kubelet[2785]: E0420 19:23:24.314024 2785 request.go:1360] "Unexpected error when reading response body" err="context deadline exceeded" Apr 20 19:23:28.259383 kubelet[2785]: E0420 19:23:28.106121 2785 controller.go:145] "Failed to ensure lease exists, will retry" err="unexpected error when reading response body. Please retry. Original error: context deadline exceeded" interval="200ms" Apr 20 19:23:28.932716 update_engine[1624]: Apr 20 19:23:28.932716 update_engine[1624]: Apr 20 19:23:28.932716 update_engine[1624]: Apr 20 19:23:28.932716 update_engine[1624]: Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.229995 1624 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.545575 1624 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.751798 1624 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 20 19:23:28.932716 update_engine[1624]: E20260420 19:23:27.763868 1624 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.769842 1624 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.770022 1624 omaha_request_action.cc:617] Omaha request response: Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.770032 1624 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.770037 1624 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.770042 1624 update_attempter.cc:306] Processing Done. Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.770049 1624 update_attempter.cc:310] Error event sent. Apr 20 19:23:28.932716 update_engine[1624]: I20260420 19:23:27.770097 1624 update_check_scheduler.cc:74] Next update check in 41m12s Apr 20 19:23:34.470653 kubelet[2785]: E0420 19:23:34.444063 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:23:36.619053 containerd[1648]: time="2026-04-20T19:23:36.618834081Z" level=info msg="StopContainer for \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" with timeout 30 (s)" Apr 20 19:23:39.156815 containerd[1648]: time="2026-04-20T19:23:38.463980687Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:23:40.368915 containerd[1648]: time="2026-04-20T19:23:39.771899091Z" level=error msg="ttrpc: received message on inactive stream" stream=75 Apr 20 19:23:42.088654 containerd[1648]: time="2026-04-20T19:23:41.515600109Z" level=warning msg="unknown status" status=0 Apr 20 19:23:42.785208 kubelet[2785]: E0420 19:23:42.776853 2785 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:23:42.952708 containerd[1648]: time="2026-04-20T19:23:42.789589780Z" level=error msg="ttrpc: received message on inactive stream" stream=29 Apr 20 19:23:44.903978 containerd[1648]: time="2026-04-20T19:23:43.124098130Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:23:45.551962 containerd[1648]: time="2026-04-20T19:23:45.173037162Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:23:48.225958 containerd[1648]: time="2026-04-20T19:23:46.320047016Z" level=warning msg="unknown status" status=0 Apr 20 19:23:48.953875 containerd[1648]: time="2026-04-20T19:23:48.046404400Z" level=error msg="failed to drain init process edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:23:49.476846 containerd[1648]: time="2026-04-20T19:23:46.517937308Z" level=warning msg="unknown status" status=0 Apr 20 19:23:50.224590 containerd[1648]: time="2026-04-20T19:23:48.157101912Z" level=error msg="failed to delete task" error="context deadline exceeded" id=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 Apr 20 19:23:51.054296 containerd[1648]: time="2026-04-20T19:23:50.276061231Z" level=error msg="ttrpc: received message on inactive stream" stream=79 Apr 20 19:23:51.224727 kubelet[2785]: E0420 19:23:51.221992 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:23:51.593044 containerd[1648]: time="2026-04-20T19:23:50.773280251Z" level=info msg="Stop container \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" with signal terminated" Apr 20 19:23:53.436182 containerd[1648]: time="2026-04-20T19:23:51.780652748Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:23:56.640988 systemd[1]: /usr/lib/systemd/system/update-engine.service:10: Support for option BlockIOWeight= has been removed and it is ignored Apr 20 19:24:19.463696 containerd[1648]: time="2026-04-20T19:24:15.554121514Z" level=info msg="container event discarded" container=5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7 type=CONTAINER_CREATED_EVENT Apr 20 19:24:20.306457 containerd[1648]: time="2026-04-20T19:24:19.463787068Z" level=info msg="container event discarded" container=f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a type=CONTAINER_CREATED_EVENT Apr 20 19:24:24.409642 containerd[1648]: time="2026-04-20T19:24:16.446669867Z" level=error msg="ttrpc: received message on inactive stream" stream=31 Apr 20 19:24:29.532731 containerd[1648]: time="2026-04-20T19:24:24.443899317Z" level=info msg="container event discarded" container=5b8ced65cb5d351771aed9c8f7c4259cb7f984c23f694bd073019371a7dbbce7 type=CONTAINER_STARTED_EVENT Apr 20 19:24:32.585054 containerd[1648]: time="2026-04-20T19:24:32.026522759Z" level=info msg="container event discarded" container=f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a type=CONTAINER_STARTED_EVENT Apr 20 19:24:34.678383 containerd[1648]: time="2026-04-20T19:24:32.299035528Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:24:35.945388 containerd[1648]: time="2026-04-20T19:24:33.013192790Z" level=info msg="container event discarded" container=61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8 type=CONTAINER_CREATED_EVENT Apr 20 19:24:36.762417 kubelet[2785]: E0420 19:24:36.440143 2785 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:24:37.305273 containerd[1648]: time="2026-04-20T19:24:37.218681505Z" level=info msg="container event discarded" container=61a8850ba787d7a557b8b1b18871448437761fa290d57f0993694b543bd107b8 type=CONTAINER_STARTED_EVENT Apr 20 19:24:41.121904 containerd[1648]: time="2026-04-20T19:24:41.047339833Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:24:43.333982 systemd[1]: Reloading finished in 149935 ms. Apr 20 19:24:43.910000 audit: BPF prog-id=89 op=LOAD Apr 20 19:24:43.911000 audit: BPF prog-id=39 op=UNLOAD Apr 20 19:24:43.911000 audit: BPF prog-id=90 op=LOAD Apr 20 19:24:43.916000 audit: BPF prog-id=91 op=LOAD Apr 20 19:24:43.916000 audit: BPF prog-id=40 op=UNLOAD Apr 20 19:24:43.916000 audit: BPF prog-id=41 op=UNLOAD Apr 20 19:24:43.938000 audit: BPF prog-id=92 op=LOAD Apr 20 19:24:43.950000 audit: BPF prog-id=42 op=UNLOAD Apr 20 19:24:44.174000 audit: BPF prog-id=93 op=LOAD Apr 20 19:24:44.180000 audit: BPF prog-id=43 op=UNLOAD Apr 20 19:24:44.254623 kernel: audit: type=1334 audit(1776713083.910:372): prog-id=89 op=LOAD Apr 20 19:24:44.290710 kernel: audit: type=1334 audit(1776713083.911:373): prog-id=39 op=UNLOAD Apr 20 19:24:44.324149 kernel: audit: type=1334 audit(1776713083.911:374): prog-id=90 op=LOAD Apr 20 19:24:44.323000 audit: BPF prog-id=94 op=LOAD Apr 20 19:24:44.336000 audit: BPF prog-id=44 op=UNLOAD Apr 20 19:24:44.461438 kernel: audit: type=1334 audit(1776713083.916:375): prog-id=91 op=LOAD Apr 20 19:24:44.479427 kernel: audit: type=1334 audit(1776713083.916:376): prog-id=40 op=UNLOAD Apr 20 19:24:44.568799 kernel: audit: type=1334 audit(1776713083.916:377): prog-id=41 op=UNLOAD Apr 20 19:24:44.580888 kernel: audit: type=1334 audit(1776713083.938:378): prog-id=92 op=LOAD Apr 20 19:24:44.586061 kernel: audit: type=1334 audit(1776713083.950:379): prog-id=42 op=UNLOAD Apr 20 19:24:44.652939 kernel: audit: type=1334 audit(1776713084.174:380): prog-id=93 op=LOAD Apr 20 19:24:44.665066 kernel: audit: type=1334 audit(1776713084.180:381): prog-id=43 op=UNLOAD Apr 20 19:24:44.727000 audit: BPF prog-id=95 op=LOAD Apr 20 19:24:44.739000 audit: BPF prog-id=69 op=UNLOAD Apr 20 19:24:45.360944 containerd[1648]: time="2026-04-20T19:24:41.579182032Z" level=warning msg="unknown status" status=0 Apr 20 19:24:45.564000 audit: BPF prog-id=96 op=LOAD Apr 20 19:24:45.564000 audit: BPF prog-id=45 op=UNLOAD Apr 20 19:24:45.565000 audit: BPF prog-id=97 op=LOAD Apr 20 19:24:45.565000 audit: BPF prog-id=98 op=LOAD Apr 20 19:24:45.565000 audit: BPF prog-id=46 op=UNLOAD Apr 20 19:24:45.565000 audit: BPF prog-id=47 op=UNLOAD Apr 20 19:24:45.956000 audit: BPF prog-id=99 op=LOAD Apr 20 19:24:45.966000 audit: BPF prog-id=63 op=UNLOAD Apr 20 19:24:46.073000 audit: BPF prog-id=100 op=LOAD Apr 20 19:24:46.073000 audit: BPF prog-id=101 op=LOAD Apr 20 19:24:46.073000 audit: BPF prog-id=48 op=UNLOAD Apr 20 19:24:46.073000 audit: BPF prog-id=49 op=UNLOAD Apr 20 19:24:46.570000 audit: BPF prog-id=102 op=LOAD Apr 20 19:24:46.578000 audit: BPF prog-id=50 op=UNLOAD Apr 20 19:24:46.579000 audit: BPF prog-id=103 op=LOAD Apr 20 19:24:46.584000 audit: BPF prog-id=104 op=LOAD Apr 20 19:24:46.609000 audit: BPF prog-id=51 op=UNLOAD Apr 20 19:24:46.613000 audit: BPF prog-id=52 op=UNLOAD Apr 20 19:24:47.105000 audit: BPF prog-id=105 op=LOAD Apr 20 19:24:47.135000 audit: BPF prog-id=74 op=UNLOAD Apr 20 19:24:47.319000 audit: BPF prog-id=106 op=LOAD Apr 20 19:24:47.329000 audit: BPF prog-id=53 op=UNLOAD Apr 20 19:24:47.335000 audit: BPF prog-id=107 op=LOAD Apr 20 19:24:47.353000 audit: BPF prog-id=108 op=LOAD Apr 20 19:24:47.368000 audit: BPF prog-id=54 op=UNLOAD Apr 20 19:24:47.378000 audit: BPF prog-id=55 op=UNLOAD Apr 20 19:24:47.431000 audit: BPF prog-id=109 op=LOAD Apr 20 19:24:47.446000 audit: BPF prog-id=78 op=UNLOAD Apr 20 19:24:47.580000 audit: BPF prog-id=110 op=LOAD Apr 20 19:24:47.580000 audit: BPF prog-id=56 op=UNLOAD Apr 20 19:24:47.580000 audit: BPF prog-id=111 op=LOAD Apr 20 19:24:47.580000 audit: BPF prog-id=112 op=LOAD Apr 20 19:24:47.580000 audit: BPF prog-id=57 op=UNLOAD Apr 20 19:24:47.580000 audit: BPF prog-id=58 op=UNLOAD Apr 20 19:24:48.153078 containerd[1648]: time="2026-04-20T19:24:41.934126851Z" level=error msg="ttrpc: received message on inactive stream" stream=85 Apr 20 19:24:48.224000 audit: BPF prog-id=113 op=LOAD Apr 20 19:24:48.224000 audit: BPF prog-id=59 op=UNLOAD Apr 20 19:24:53.850150 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7-rootfs.mount: Deactivated successfully. Apr 20 19:24:58.117724 systemd[1]: cri-containerd-78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7.scope: Deactivated successfully. Apr 20 19:24:58.210000 audit: BPF prog-id=83 op=UNLOAD Apr 20 19:24:58.247000 audit: BPF prog-id=109 op=UNLOAD Apr 20 19:24:58.440531 kernel: kauditd_printk_skb: 40 callbacks suppressed Apr 20 19:24:58.288758 systemd[1]: cri-containerd-78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7.scope: Consumed 14.023s CPU time, 17.6M memory peak. Apr 20 19:24:58.825004 kernel: audit: type=1334 audit(1776713098.210:422): prog-id=83 op=UNLOAD Apr 20 19:24:58.834011 kernel: audit: type=1334 audit(1776713098.247:423): prog-id=109 op=UNLOAD Apr 20 19:25:01.123122 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:25:02.318182 systemd[1]: kubelet.service: Deactivated successfully. Apr 20 19:25:02.363000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:25:02.447330 kernel: audit: type=1131 audit(1776713102.363:424): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:25:02.359889 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:25:02.372873 systemd[1]: kubelet.service: Consumed 2min 17.723s CPU time, 137.7M memory peak. Apr 20 19:25:02.762134 containerd[1648]: time="2026-04-20T19:24:16.040734526Z" level=error msg="failed to delete task" error="context deadline exceeded" id=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 Apr 20 19:25:03.030778 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 20 19:25:03.600589 containerd[1648]: time="2026-04-20T19:25:03.580167749Z" level=error msg="StopContainer for \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" failed" error="rpc error: code = Canceled desc = context canceled" Apr 20 19:25:04.389147 containerd[1648]: time="2026-04-20T19:25:03.863253465Z" level=error msg="failed to handle container TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" error="failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:25:06.640029 containerd[1648]: time="2026-04-20T19:25:06.633602113Z" level=info msg="container event discarded" container=5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a type=CONTAINER_CREATED_EVENT Apr 20 19:25:06.640029 containerd[1648]: time="2026-04-20T19:25:06.638092773Z" level=info msg="container event discarded" container=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 type=CONTAINER_CREATED_EVENT Apr 20 19:25:06.742817 containerd[1648]: time="2026-04-20T19:25:06.683088004Z" level=info msg="container event discarded" container=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 type=CONTAINER_CREATED_EVENT Apr 20 19:25:06.742817 containerd[1648]: time="2026-04-20T19:25:06.718075467Z" level=info msg="container event discarded" container=5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a type=CONTAINER_STARTED_EVENT Apr 20 19:25:06.742817 containerd[1648]: time="2026-04-20T19:25:06.728825884Z" level=info msg="container event discarded" container=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 type=CONTAINER_STARTED_EVENT Apr 20 19:25:06.937625 containerd[1648]: time="2026-04-20T19:25:06.840061182Z" level=info msg="container event discarded" container=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 type=CONTAINER_STARTED_EVENT Apr 20 19:25:06.963881 containerd[1648]: time="2026-04-20T19:25:06.937816522Z" level=error msg="ttrpc: received message on inactive stream" stream=89 Apr 20 19:25:06.983053 containerd[1648]: time="2026-04-20T19:25:06.952485697Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:25:07.072131 containerd[1648]: time="2026-04-20T19:25:07.070133250Z" level=warning msg="unknown status" status=0 Apr 20 19:25:07.159797 containerd[1648]: time="2026-04-20T19:25:07.104144184Z" level=error msg="failed to delete task" error="context deadline exceeded" id=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 Apr 20 19:25:07.263525 containerd[1648]: time="2026-04-20T19:25:07.163767056Z" level=error msg="failed to drain init process edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:25:07.359724 containerd[1648]: time="2026-04-20T19:25:07.341906070Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:25:07.739613 containerd[1648]: time="2026-04-20T19:25:07.529612451Z" level=error msg="ttrpc: received message on inactive stream" stream=91 Apr 20 19:25:07.919846 containerd[1648]: time="2026-04-20T19:25:07.915475924Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:25:18.544279 containerd[1648]: time="2026-04-20T19:25:18.386215771Z" level=error msg="failed to delete task" error="context deadline exceeded" id=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 Apr 20 19:25:19.174158 containerd[1648]: time="2026-04-20T19:25:19.151059082Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:25:19.639193 containerd[1648]: time="2026-04-20T19:25:19.589205646Z" level=error msg="ttrpc: received message on inactive stream" stream=47 Apr 20 19:25:21.876129 containerd[1648]: time="2026-04-20T19:25:21.837149855Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:25:30.931578 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 20 19:25:31.542000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:25:31.656636 kernel: audit: type=1130 audit(1776713131.542:425): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:25:32.888204 containerd[1648]: time="2026-04-20T19:25:32.362212825Z" level=error msg="failed to delete task" error="context deadline exceeded" id=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 Apr 20 19:25:33.918591 containerd[1648]: time="2026-04-20T19:25:33.910870461Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:25:35.082987 containerd[1648]: time="2026-04-20T19:25:33.876461721Z" level=error msg="ttrpc: received message on inactive stream" stream=61 Apr 20 19:25:38.114508 (kubelet)[3245]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 20 19:25:38.928400 containerd[1648]: time="2026-04-20T19:25:38.918510836Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:25:51.285796 containerd[1648]: time="2026-04-20T19:25:50.765810967Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:25:51.770951 containerd[1648]: time="2026-04-20T19:25:51.549309380Z" level=error msg="ttrpc: received message on inactive stream" stream=69 Apr 20 19:25:52.928629 containerd[1648]: time="2026-04-20T19:25:52.428474600Z" level=error msg="ttrpc: received message on inactive stream" stream=67 Apr 20 19:26:00.399536 containerd[1648]: time="2026-04-20T19:26:00.341876486Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:26:20.722669 containerd[1648]: time="2026-04-20T19:26:20.702312796Z" level=error msg="ttrpc: received message on inactive stream" stream=75 Apr 20 19:26:20.826382 containerd[1648]: time="2026-04-20T19:26:20.704472227Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:26:21.078935 containerd[1648]: time="2026-04-20T19:26:21.018174120Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:26:21.253361 containerd[1648]: time="2026-04-20T19:26:21.033738897Z" level=error msg="ttrpc: received message on inactive stream" stream=77 Apr 20 19:26:28.637000 audit[3267]: AUDIT1101 pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock,pam_permit acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:26:29.325000 audit[3267]: AUDIT1110 pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:26:29.806660 kernel: audit: type=1101 audit(1776713188.637:426): pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock,pam_permit acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:26:29.820000 audit[3267]: AUDIT1105 pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:26:29.078849 sudo[3267]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Apr 20 19:26:30.509951 kernel: audit: type=1110 audit(1776713189.325:427): pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:26:29.796011 sudo[3267]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Apr 20 19:26:30.742345 kernel: audit: type=1105 audit(1776713189.820:428): pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:26:41.800036 containerd[1648]: time="2026-04-20T19:26:41.520581928Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:26:42.347122 containerd[1648]: time="2026-04-20T19:26:41.850207100Z" level=error msg="ttrpc: received message on inactive stream" stream=99 Apr 20 19:26:42.766204 containerd[1648]: time="2026-04-20T19:26:42.374054988Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:26:43.232170 containerd[1648]: time="2026-04-20T19:26:42.673121241Z" level=error msg="ttrpc: received message on inactive stream" stream=97 Apr 20 19:27:00.059972 containerd[1648]: time="2026-04-20T19:27:00.034953683Z" level=error msg="ttrpc: received message on inactive stream" stream=83 Apr 20 19:27:01.304868 containerd[1648]: time="2026-04-20T19:27:01.169976545Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:27:01.721191 containerd[1648]: time="2026-04-20T19:27:01.234003420Z" level=error msg="ttrpc: received message on inactive stream" stream=85 Apr 20 19:27:34.216028 containerd[1648]: time="2026-04-20T19:27:34.215063031Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:27:45.746211 sudo[3267]: pam_unix(sudo:session): session closed for user root Apr 20 19:27:45.844000 audit[3267]: AUDIT1106 pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:27:46.108000 audit[3267]: AUDIT1104 pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:27:47.251874 kernel: audit: type=1106 audit(1776713265.844:429): pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:27:47.413195 kernel: audit: type=1104 audit(1776713266.108:430): pid=3267 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:28:06.757794 containerd[1648]: time="2026-04-20T19:28:06.255955373Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:28:07.666881 containerd[1648]: time="2026-04-20T19:28:07.656946243Z" level=error msg="ttrpc: received message on inactive stream" stream=91 Apr 20 19:28:08.831644 containerd[1648]: time="2026-04-20T19:28:08.556618558Z" level=error msg="ttrpc: received message on inactive stream" stream=93 Apr 20 19:28:24.034190 kubelet[3245]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 20 19:28:24.730330 kubelet[3245]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 20 19:28:24.730330 kubelet[3245]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 20 19:28:24.905625 kubelet[3245]: I0420 19:28:24.440016 3245 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 20 19:28:45.553208 kubelet[3245]: I0420 19:28:45.540703 3245 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 20 19:28:46.683786 kubelet[3245]: I0420 19:28:45.666392 3245 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 20 19:28:48.367485 kubelet[3245]: I0420 19:28:48.341927 3245 server.go:956] "Client rotation is on, will bootstrap in background" Apr 20 19:28:51.307107 containerd[1648]: time="2026-04-20T19:28:51.215047976Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:28:56.640007 kubelet[3245]: I0420 19:28:56.625992 3245 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 20 19:29:06.451187 containerd[1648]: time="2026-04-20T19:29:06.166123968Z" level=error msg="ttrpc: received message on inactive stream" stream=105 Apr 20 19:29:07.224188 containerd[1648]: time="2026-04-20T19:29:07.072018840Z" level=error msg="ttrpc: received message on inactive stream" stream=107 Apr 20 19:29:07.562783 containerd[1648]: time="2026-04-20T19:29:07.424207700Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:29:08.121151 kubelet[3245]: I0420 19:29:07.989169 3245 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 20 19:29:12.081540 containerd[1648]: time="2026-04-20T19:29:11.445187892Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:29:26.540841 containerd[1648]: time="2026-04-20T19:29:26.502602347Z" level=error msg="ttrpc: received message on inactive stream" stream=99 Apr 20 19:29:27.239776 containerd[1648]: time="2026-04-20T19:29:27.147950220Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:29:27.539723 containerd[1648]: time="2026-04-20T19:29:27.512185918Z" level=error msg="ttrpc: received message on inactive stream" stream=101 Apr 20 19:29:40.381057 kubelet[3245]: I0420 19:29:40.379876 3245 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 20 19:30:08.697940 kubelet[3245]: I0420 19:30:08.515794 3245 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 20 19:30:09.660046 kubelet[3245]: I0420 19:30:09.577011 3245 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 20 19:30:10.575752 kubelet[3245]: I0420 19:30:09.755568 3245 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 20 19:30:10.985769 kubelet[3245]: I0420 19:30:10.662479 3245 topology_manager.go:138] "Creating topology manager with none policy" Apr 20 19:30:10.985769 kubelet[3245]: I0420 19:30:10.667771 3245 container_manager_linux.go:303] "Creating device plugin manager" Apr 20 19:30:10.985769 kubelet[3245]: I0420 19:30:10.774151 3245 state_mem.go:36] "Initialized new in-memory state store" Apr 20 19:30:11.592492 kubelet[3245]: I0420 19:30:11.464011 3245 kubelet.go:480] "Attempting to sync node with API server" Apr 20 19:30:12.048962 kubelet[3245]: I0420 19:30:11.673969 3245 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 20 19:30:12.048962 kubelet[3245]: I0420 19:30:11.874996 3245 kubelet.go:386] "Adding apiserver pod source" Apr 20 19:30:12.048962 kubelet[3245]: I0420 19:30:11.884380 3245 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 20 19:30:19.725802 kubelet[3245]: I0420 19:30:19.712983 3245 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.2.1" apiVersion="v1" Apr 20 19:30:22.913851 kubelet[3245]: I0420 19:30:22.875737 3245 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 20 19:30:23.757123 kubelet[3245]: I0420 19:30:23.695729 3245 apiserver.go:52] "Watching apiserver" Apr 20 19:30:37.453044 kubelet[3245]: I0420 19:30:37.443025 3245 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 20 19:30:38.083782 kubelet[3245]: I0420 19:30:37.668210 3245 server.go:1289] "Started kubelet" Apr 20 19:30:38.083782 kubelet[3245]: I0420 19:30:37.810941 3245 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 20 19:30:40.074147 kubelet[3245]: I0420 19:30:39.188107 3245 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 20 19:30:40.289557 kubelet[3245]: I0420 19:30:40.161048 3245 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 20 19:30:48.282751 kubelet[3245]: I0420 19:30:48.278064 3245 server.go:317] "Adding debug handlers to kubelet server" Apr 20 19:30:48.716097 kubelet[3245]: I0420 19:30:48.352428 3245 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 20 19:30:48.716097 kubelet[3245]: I0420 19:30:48.578260 3245 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 20 19:30:48.716097 kubelet[3245]: I0420 19:30:48.635139 3245 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 20 19:30:50.045857 kubelet[3245]: I0420 19:30:49.733041 3245 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 20 19:30:53.410692 kubelet[3245]: I0420 19:30:53.404840 3245 reconciler.go:26] "Reconciler: start to sync state" Apr 20 19:31:21.164397 kubelet[3245]: I0420 19:31:21.058101 3245 factory.go:223] Registration of the systemd container factory successfully Apr 20 19:31:22.070050 kubelet[3245]: I0420 19:31:22.017044 3245 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 20 19:31:24.755878 kubelet[3245]: W0420 19:31:24.748120 3245 logging.go:55] [core] [Channel #9 SubChannel #10]grpc: addrConn.createTransport failed to connect to {Addr: "/run/containerd/containerd.sock", ServerName: "localhost", Attributes: {"<%!p(networktype.keyType=grpc.internal.transport.networktype)>": "unix" }, }. Err: connection error: desc = "transport: Error while dialing: dial unix:///run/containerd/containerd.sock: timeout" Apr 20 19:31:28.441165 kubelet[3245]: I0420 19:31:28.245696 3245 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: context deadline exceeded Apr 20 19:31:28.942582 kubelet[3245]: W0420 19:31:28.367055 3245 logging.go:55] [core] [Channel #9 SubChannel #10]grpc: addrConn.createTransport failed to connect to {Addr: "/run/containerd/containerd.sock", ServerName: "localhost", Attributes: {"<%!p(networktype.keyType=grpc.internal.transport.networktype)>": "unix" }, }. Err: connection error: desc = "transport: Error while dialing: dial unix:///run/containerd/containerd.sock: timeout" Apr 20 19:31:41.142369 containerd[1648]: time="2026-04-20T19:31:40.768669365Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:31:44.336043 containerd[1648]: time="2026-04-20T19:31:44.021816950Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:31:45.492033 kubelet[3245]: E0420 19:31:43.982942 3245 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 20 19:31:46.126991 containerd[1648]: time="2026-04-20T19:31:44.556569619Z" level=error msg="ttrpc: received message on inactive stream" stream=103 Apr 20 19:31:46.614713 containerd[1648]: time="2026-04-20T19:31:45.462499856Z" level=warning msg="unknown status" status=0 Apr 20 19:32:03.367890 containerd[1648]: time="2026-04-20T19:32:03.062986944Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:32:04.038609 containerd[1648]: time="2026-04-20T19:32:03.227056996Z" level=error msg="ttrpc: received message on inactive stream" stream=107 Apr 20 19:32:04.442396 kubelet[3245]: I0420 19:32:04.429829 3245 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 20 19:32:05.250454 containerd[1648]: time="2026-04-20T19:32:04.275918982Z" level=warning msg="unknown status" status=0 Apr 20 19:32:07.959396 kubelet[3245]: I0420 19:32:07.866671 3245 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 20 19:32:08.350583 containerd[1648]: time="2026-04-20T19:32:07.451801031Z" level=error msg="failed to delete task" error="context deadline exceeded" id=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 Apr 20 19:32:08.571111 containerd[1648]: time="2026-04-20T19:32:08.234523533Z" level=error msg="failed to drain init process 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:32:08.751447 kubelet[3245]: I0420 19:32:08.623704 3245 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 20 19:32:08.928979 kubelet[3245]: I0420 19:32:08.641007 3245 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 20 19:32:09.565874 containerd[1648]: time="2026-04-20T19:32:09.389956875Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:32:10.370913 kubelet[3245]: I0420 19:32:09.356073 3245 kubelet.go:2436] "Starting kubelet main sync loop" Apr 20 19:32:10.966979 kubelet[3245]: E0420 19:32:10.656169 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:11.516425 kubelet[3245]: E0420 19:32:11.167064 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:12.020854 kubelet[3245]: E0420 19:32:11.924536 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:12.483743 containerd[1648]: time="2026-04-20T19:32:12.418922074Z" level=error msg="ttrpc: received message on inactive stream" stream=109 Apr 20 19:32:12.860394 kubelet[3245]: E0420 19:32:12.731096 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:14.830052 kubelet[3245]: E0420 19:32:14.331089 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:18.246639 kubelet[3245]: E0420 19:32:18.217526 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:23.305944 kubelet[3245]: E0420 19:32:22.884963 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:29.837942 kubelet[3245]: E0420 19:32:29.764958 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:36.263927 kubelet[3245]: E0420 19:32:36.229904 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:41.661964 kubelet[3245]: E0420 19:32:41.657961 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:49.143704 kubelet[3245]: E0420 19:32:48.271901 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:32:55.174095 kubelet[3245]: E0420 19:32:55.156954 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:33:00.483090 kubelet[3245]: E0420 19:33:00.291076 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:33:07.864989 kubelet[3245]: E0420 19:33:06.263192 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:33:15.153528 kubelet[3245]: E0420 19:33:15.148995 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 20 19:33:22.753769 kubelet[3245]: E0420 19:33:21.512939 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:33:29.186909 kubelet[3245]: E0420 19:33:29.155892 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:33:30.963882 containerd[1648]: time="2026-04-20T19:33:25.861541475Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:33:35.728665 kubelet[3245]: E0420 19:33:35.726334 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:33:44.328308 kubelet[3245]: E0420 19:33:44.315874 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:33:45.184034 containerd[1648]: time="2026-04-20T19:33:45.180151725Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:33:45.744281 containerd[1648]: time="2026-04-20T19:33:45.211454926Z" level=error msg="ttrpc: received message on inactive stream" stream=109 Apr 20 19:33:46.546540 containerd[1648]: time="2026-04-20T19:33:45.211462480Z" level=warning msg="unknown status" status=0 Apr 20 19:33:53.512745 kubelet[3245]: E0420 19:33:53.386153 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:33:59.282027 kubelet[3245]: E0420 19:33:59.276923 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:34:03.089954 kubelet[3245]: E0420 19:34:01.478023 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:08.720854 kubelet[3245]: E0420 19:34:08.483172 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:13.936919 kubelet[3245]: E0420 19:34:13.925385 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:20.062965 kubelet[3245]: E0420 19:34:19.927033 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:28.805934 containerd[1648]: time="2026-04-20T19:34:28.564785113Z" level=error msg="ttrpc: received message on inactive stream" stream=111 Apr 20 19:34:29.780566 containerd[1648]: time="2026-04-20T19:34:28.642945086Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:34:29.864313 kubelet[3245]: E0420 19:34:27.264882 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:30.609534 containerd[1648]: time="2026-04-20T19:34:29.864944302Z" level=warning msg="unknown status" status=0 Apr 20 19:34:35.144066 kubelet[3245]: E0420 19:34:35.130986 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:41.225687 kubelet[3245]: E0420 19:34:40.352056 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:46.657815 containerd[1648]: time="2026-04-20T19:34:46.653761528Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:34:47.685765 kubelet[3245]: E0420 19:34:46.702056 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:48.545345 containerd[1648]: time="2026-04-20T19:34:47.871555335Z" level=error msg="ttrpc: received message on inactive stream" stream=113 Apr 20 19:34:49.050540 containerd[1648]: time="2026-04-20T19:34:48.677274470Z" level=warning msg="unknown status" status=0 Apr 20 19:34:51.980547 kubelet[3245]: E0420 19:34:51.826039 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:34:58.029008 kubelet[3245]: E0420 19:34:57.990775 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:35:02.056757 containerd[1648]: time="2026-04-20T19:35:01.318632393Z" level=error msg="failed to delete task" error="context deadline exceeded" id=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 Apr 20 19:35:03.363819 containerd[1648]: time="2026-04-20T19:35:02.063034648Z" level=error msg="failed to drain init process edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:35:04.506867 kubelet[3245]: E0420 19:35:04.501974 3245 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 20 19:35:07.481724 kubelet[3245]: E0420 19:35:07.462981 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:35:08.553506 containerd[1648]: time="2026-04-20T19:35:08.551422231Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:35:10.842995 kubelet[3245]: E0420 19:35:10.805006 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m1.63687108s ago; threshold is 3m0s]" Apr 20 19:35:12.875070 containerd[1648]: time="2026-04-20T19:35:10.257027093Z" level=error msg="ttrpc: received message on inactive stream" stream=115 Apr 20 19:35:22.763205 kubelet[3245]: E0420 19:35:19.370078 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m9.663555291s ago; threshold is 3m0s]" Apr 20 19:35:27.960864 kubelet[3245]: E0420 19:35:27.942845 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m18.767625634s ago; threshold is 3m0s]" Apr 20 19:35:29.084465 kubelet[3245]: E0420 19:35:27.170833 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:35:36.766806 kubelet[3245]: E0420 19:35:35.384980 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m24.910890358s ago; threshold is 3m0s]" Apr 20 19:35:45.977925 kubelet[3245]: E0420 19:35:45.762837 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m34.481617851s ago; threshold is 3m0s]" Apr 20 19:35:49.536135 kubelet[3245]: E0420 19:35:45.847930 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:35:52.062081 kubelet[3245]: E0420 19:35:51.124772 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m41.777371342s ago; threshold is 3m0s]" Apr 20 19:36:00.167935 kubelet[3245]: E0420 19:36:00.144026 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m50.797878058s ago; threshold is 3m0s]" Apr 20 19:36:07.715161 kubelet[3245]: E0420 19:36:06.117212 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 3m56.096712291s ago; threshold is 3m0s]" Apr 20 19:36:16.843945 kubelet[3245]: E0420 19:36:16.747958 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 4m7.579816049s ago; threshold is 3m0s]" Apr 20 19:36:18.847459 kubelet[3245]: E0420 19:36:15.473200 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:36:34.132032 kubelet[3245]: E0420 19:36:30.790998 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 4m15.497289326s ago; threshold is 3m0s]" Apr 20 19:36:38.309346 containerd[1648]: time="2026-04-20T19:36:34.691369756Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:36:47.438972 kubelet[3245]: E0420 19:36:43.301970 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 4m31.392570124s ago; threshold is 3m0s]" Apr 20 19:36:53.236009 kubelet[3245]: E0420 19:36:53.228389 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:36:53.426763 containerd[1648]: time="2026-04-20T19:36:51.544025651Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:36:53.426763 containerd[1648]: time="2026-04-20T19:36:53.377342587Z" level=warning msg="unknown status" status=0 Apr 20 19:36:53.426763 containerd[1648]: time="2026-04-20T19:36:53.378492675Z" level=error msg="ttrpc: received message on inactive stream" stream=111 Apr 20 19:36:54.271718 kubelet[3245]: I0420 19:36:54.068833 3245 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Apr 20 19:36:54.883960 kubelet[3245]: E0420 19:36:54.584759 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 4m45.015813132s ago; threshold is 3m0s]" Apr 20 19:37:01.955192 kubelet[3245]: E0420 19:37:01.951469 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 4m50.564742815s ago; threshold is 3m0s]" Apr 20 19:37:10.350786 kubelet[3245]: E0420 19:37:09.323925 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 4m57.918520374s ago; threshold is 3m0s]" Apr 20 19:37:16.313136 kubelet[3245]: E0420 19:37:16.291833 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m6.2409748s ago; threshold is 3m0s]" Apr 20 19:37:23.268811 kubelet[3245]: E0420 19:37:23.217132 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 20 19:37:23.621580 kubelet[3245]: E0420 19:37:21.708133 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m12.355524548s ago; threshold is 3m0s]" Apr 20 19:37:24.661989 containerd[1648]: time="2026-04-20T19:37:24.417869074Z" level=error msg="ttrpc: received message on inactive stream" stream=113 Apr 20 19:37:25.576697 containerd[1648]: time="2026-04-20T19:37:25.491129379Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:37:26.139539 containerd[1648]: time="2026-04-20T19:37:25.960628582Z" level=warning msg="unknown status" status=0 Apr 20 19:37:30.332138 kubelet[3245]: E0420 19:37:30.144757 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m20.089618s ago; threshold is 3m0s]" Apr 20 19:37:36.243600 kubelet[3245]: E0420 19:37:36.234968 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m26.940576899s ago; threshold is 3m0s]" Apr 20 19:37:36.876366 kubelet[3245]: E0420 19:37:35.865130 3245 controller.go:195] "Failed to update lease" err="Operation cannot be fulfilled on leases.coordination.k8s.io \"localhost\": the object has been modified; please apply your changes to the latest version and try again" Apr 20 19:37:41.967658 kubelet[3245]: E0420 19:37:41.366070 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m32.184851603s ago; threshold is 3m0s]" Apr 20 19:37:50.454924 kubelet[3245]: E0420 19:37:50.252990 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m38.444375276s ago; threshold is 3m0s]" Apr 20 19:37:56.001695 kubelet[3245]: E0420 19:37:56.001271 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m46.47621423s ago; threshold is 3m0s]" Apr 20 19:37:56.502296 containerd[1648]: time="2026-04-20T19:37:55.813916261Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:37:56.502296 containerd[1648]: time="2026-04-20T19:37:56.468592799Z" level=warning msg="unknown status" status=0 Apr 20 19:37:57.653923 containerd[1648]: time="2026-04-20T19:37:55.978482806Z" level=error msg="ttrpc: received message on inactive stream" stream=115 Apr 20 19:37:59.816958 containerd[1648]: time="2026-04-20T19:37:59.437765965Z" level=error msg="failed to delete task" error="context deadline exceeded" id=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 Apr 20 19:38:01.848641 containerd[1648]: time="2026-04-20T19:38:00.376142159Z" level=error msg="failed to drain init process 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:38:02.437849 kubelet[3245]: E0420 19:38:01.839820 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m51.844734607s ago; threshold is 3m0s]" Apr 20 19:38:04.282034 containerd[1648]: time="2026-04-20T19:38:04.262885847Z" level=error msg="ttrpc: received message on inactive stream" stream=117 Apr 20 19:38:05.758643 kubelet[3245]: I0420 19:38:05.754473 3245 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 20 19:38:05.758643 kubelet[3245]: I0420 19:38:05.756923 3245 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 20 19:38:06.584961 kubelet[3245]: I0420 19:38:06.244127 3245 state_mem.go:36] "Initialized new in-memory state store" Apr 20 19:38:07.473594 kubelet[3245]: E0420 19:38:07.441726 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 5m58.088614806s ago; threshold is 3m0s]" Apr 20 19:38:08.279023 containerd[1648]: time="2026-04-20T19:38:04.963053956Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:38:08.860038 kubelet[3245]: I0420 19:38:08.146955 3245 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 20 19:38:09.755986 kubelet[3245]: I0420 19:38:09.150875 3245 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 20 19:38:09.755986 kubelet[3245]: I0420 19:38:09.348802 3245 policy_none.go:49] "None policy: Start" Apr 20 19:38:09.755986 kubelet[3245]: I0420 19:38:09.350153 3245 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 20 19:38:09.755986 kubelet[3245]: I0420 19:38:09.369101 3245 state_mem.go:35] "Initializing new in-memory state store" Apr 20 19:38:13.227160 kubelet[3245]: E0420 19:38:13.176814 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 6m4.008990395s ago; threshold is 3m0s]" Apr 20 19:38:15.308281 kubelet[3245]: I0420 19:38:15.274681 3245 state_mem.go:75] "Updated machine memory state" Apr 20 19:38:19.352769 kubelet[3245]: E0420 19:38:18.580153 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 6m9.287789888s ago; threshold is 3m0s]" Apr 20 19:38:25.382661 kubelet[3245]: E0420 19:38:25.241987 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 6m15.505260076s ago; threshold is 3m0s]" Apr 20 19:38:31.164061 kubelet[3245]: E0420 19:38:31.091210 3245 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 20 19:38:38.078194 kubelet[3245]: E0420 19:38:38.066935 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 6m22.148121649s ago; threshold is 3m0s]" Apr 20 19:38:42.025325 kubelet[3245]: I0420 19:38:42.012959 3245 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 20 19:38:45.565735 kubelet[3245]: I0420 19:38:45.540347 3245 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 20 19:38:48.480198 kubelet[3245]: E0420 19:38:48.467907 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg was last seen active 6m34.471595967s ago; threshold is 3m0s]" Apr 20 19:38:48.873134 kubelet[3245]: E0420 19:38:48.519101 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:38:53.308090 kubelet[3245]: I0420 19:38:52.932059 3245 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 20 19:38:55.473081 kubelet[3245]: E0420 19:38:55.470599 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 6m45.808764935s ago; threshold is 3m0s" Apr 20 19:39:02.557561 kubelet[3245]: E0420 19:39:02.526150 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 6m51.567883694s ago; threshold is 3m0s" Apr 20 19:39:07.754044 kubelet[3245]: E0420 19:39:06.517454 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:39:10.016213 kubelet[3245]: E0420 19:39:10.015209 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 7m0.117903791s ago; threshold is 3m0s" Apr 20 19:39:16.017706 kubelet[3245]: E0420 19:39:16.008610 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 7m6.839624274s ago; threshold is 3m0s" Apr 20 19:39:21.353884 kubelet[3245]: E0420 19:39:21.315071 3245 controller.go:195] "Failed to update lease" err="Operation cannot be fulfilled on leases.coordination.k8s.io \"localhost\": the object has been modified; please apply your changes to the latest version and try again" Apr 20 19:39:22.700709 kubelet[3245]: E0420 19:39:22.687086 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 7m12.758073821s ago; threshold is 3m0s" Apr 20 19:39:24.793012 kubelet[3245]: E0420 19:39:24.173044 3245 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 20 19:39:45.518147 kubelet[3245]: E0420 19:39:43.142978 3245 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="200ms" Apr 20 19:39:47.339180 kubelet[3245]: I0420 19:39:45.854107 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fde15e5d4357efe755cb792ce81eb1a9-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fde15e5d4357efe755cb792ce81eb1a9\") " pod="kube-system/kube-apiserver-localhost" Apr 20 19:39:52.339871 kubelet[3245]: I0420 19:39:50.444138 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fde15e5d4357efe755cb792ce81eb1a9-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fde15e5d4357efe755cb792ce81eb1a9\") " pod="kube-system/kube-apiserver-localhost" Apr 20 19:40:00.764672 kubelet[3245]: I0420 19:40:00.681335 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fde15e5d4357efe755cb792ce81eb1a9-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fde15e5d4357efe755cb792ce81eb1a9\") " pod="kube-system/kube-apiserver-localhost" Apr 20 19:40:05.759829 kubelet[3245]: I0420 19:40:05.727175 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:40:09.123126 kubelet[3245]: E0420 19:40:09.116949 3245 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="400ms" Apr 20 19:40:11.056798 kubelet[3245]: I0420 19:40:11.052120 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:40:12.667521 containerd[1648]: time="2026-04-20T19:40:11.768872346Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:40:17.641854 kubelet[3245]: I0420 19:40:17.628632 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:40:18.542990 kubelet[3245]: I0420 19:40:18.542514 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:40:19.901665 kubelet[3245]: E0420 19:40:19.534578 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:40:20.529829 kubelet[3245]: I0420 19:40:19.916877 3245 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 20 19:40:22.264543 kubelet[3245]: I0420 19:40:22.241845 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e9ca41790ae21be9f4cbd451ade0acec-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"e9ca41790ae21be9f4cbd451ade0acec\") " pod="kube-system/kube-controller-manager-localhost" Apr 20 19:40:23.628097 kubelet[3245]: E0420 19:40:22.484417 3245 kubelet_node_status.go:460] "Node not becoming ready in time after startup" Apr 20 19:40:26.088515 kubelet[3245]: I0420 19:40:24.088521 3245 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/33fee6ba1581201eda98a989140db110-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"33fee6ba1581201eda98a989140db110\") " pod="kube-system/kube-scheduler-localhost" Apr 20 19:40:27.605771 containerd[1648]: time="2026-04-20T19:40:25.049659649Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:40:30.459500 containerd[1648]: time="2026-04-20T19:40:25.700118982Z" level=error msg="ttrpc: received message on inactive stream" stream=117 Apr 20 19:40:32.439884 containerd[1648]: time="2026-04-20T19:40:28.880342360Z" level=warning msg="unknown status" status=0 Apr 20 19:40:36.062534 sudo[1957]: pam_unix(sudo:session): session closed for user root Apr 20 19:40:36.091000 audit[1957]: AUDIT1106 pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:40:36.219000 audit[1957]: AUDIT1104 pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:40:37.462824 kernel: audit: type=1106 audit(1776714036.091:431): pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:40:37.477937 sshd[1956]: Connection closed by 10.0.0.1 port 52626 Apr 20 19:40:37.346106 sshd-session[1952]: pam_unix(sshd:session): session closed for user core Apr 20 19:40:37.645000 audit[1952]: AUDIT1106 pid=1952 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:40:37.868000 audit[1952]: AUDIT1104 pid=1952 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:40:38.218854 kernel: audit: type=1104 audit(1776714036.219:432): pid=1957 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Apr 20 19:40:38.245010 kernel: audit: type=1106 audit(1776714037.645:433): pid=1952 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:40:38.359937 kernel: audit: type=1104 audit(1776714037.868:434): pid=1952 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Apr 20 19:40:39.303844 systemd[1]: sshd@6-12289-10.0.0.18:22-10.0.0.1:52626.service: Deactivated successfully. Apr 20 19:40:39.637000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-12289-10.0.0.18:22-10.0.0.1:52626 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:40:40.253063 kernel: audit: type=1131 audit(1776714039.637:435): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-12289-10.0.0.18:22-10.0.0.1:52626 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 20 19:40:39.699114 systemd[1]: sshd@6-12289-10.0.0.18:22-10.0.0.1:52626.service: Consumed 1.064s CPU time, 4.6M memory peak. Apr 20 19:40:41.163528 systemd[1]: session-8.scope: Deactivated successfully. Apr 20 19:40:41.448282 systemd[1]: session-8.scope: Consumed 6min 45.770s CPU time, 254.9M memory peak. Apr 20 19:40:42.490170 systemd-logind[1618]: Session 8 logged out. Waiting for processes to exit. Apr 20 19:40:44.861461 systemd-logind[1618]: Removed session 8. Apr 20 19:40:45.925203 kubelet[3245]: E0420 19:40:44.089190 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:40:46.728613 kubelet[3245]: E0420 19:40:46.639173 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:40:55.226890 containerd[1648]: time="2026-04-20T19:40:54.346174814Z" level=error msg="get state for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="context deadline exceeded" Apr 20 19:40:57.404195 containerd[1648]: time="2026-04-20T19:40:55.223706884Z" level=error msg="ttrpc: received message on inactive stream" stream=121 Apr 20 19:40:58.168782 containerd[1648]: time="2026-04-20T19:40:56.359132848Z" level=warning msg="unknown status" status=0 Apr 20 19:40:58.707620 kubelet[3245]: I0420 19:40:58.547841 3245 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Apr 20 19:41:00.294024 containerd[1648]: time="2026-04-20T19:40:59.557178707Z" level=error msg="failed to delete task" error="context deadline exceeded" id=edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 Apr 20 19:41:00.848848 containerd[1648]: time="2026-04-20T19:41:00.201991718Z" level=error msg="failed to drain init process edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:41:01.969760 containerd[1648]: time="2026-04-20T19:41:01.961955890Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:41:03.963161 containerd[1648]: time="2026-04-20T19:41:03.778397421Z" level=error msg="ttrpc: received message on inactive stream" stream=123 Apr 20 19:41:13.708839 kubelet[3245]: E0420 19:41:10.568100 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:41:24.567849 kubelet[3245]: E0420 19:41:24.459077 3245 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1m9.964s" Apr 20 19:41:25.778673 kubelet[3245]: I0420 19:41:25.777723 3245 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Apr 20 19:41:25.938144 kubelet[3245]: I0420 19:41:25.826187 3245 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Apr 20 19:41:30.790958 kubelet[3245]: E0420 19:41:30.779931 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:41:31.569042 kubelet[3245]: E0420 19:41:30.651525 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:42:14.089205 kubelet[3245]: E0420 19:42:13.995828 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:42:16.419502 kubelet[3245]: E0420 19:42:13.527850 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:42:19.255930 kubelet[3245]: E0420 19:42:19.213064 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:42:37.410187 kubelet[3245]: E0420 19:42:35.741158 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:42:59.002947 kubelet[3245]: E0420 19:42:58.080139 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:43:05.928910 kubelet[3245]: E0420 19:43:05.454587 3245 kubelet_node_status.go:548] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:42:06Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:42:06Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:42:06Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:42:06Z\\\",\\\"type\\\":\\\"Ready\\\"}]}}\" for node \"localhost\": Patch \"https://10.0.0.18:6443/api/v1/nodes/localhost/status?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:43:14.139609 containerd[1648]: time="2026-04-20T19:43:13.670670723Z" level=info msg="TaskExit event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401}" Apr 20 19:43:24.831722 containerd[1648]: time="2026-04-20T19:43:24.795911462Z" level=error msg="ttrpc: received message on inactive stream" stream=121 Apr 20 19:43:25.879151 containerd[1648]: time="2026-04-20T19:43:24.799603585Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:43:27.089633 containerd[1648]: time="2026-04-20T19:43:26.725043127Z" level=warning msg="unknown status" status=0 Apr 20 19:43:46.896217 kubelet[3245]: E0420 19:43:46.555622 3245 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": Get \"https://10.0.0.18:6443/api/v1/nodes/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:43:51.473142 kubelet[3245]: E0420 19:43:49.426004 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:44:30.444628 containerd[1648]: time="2026-04-20T19:44:28.274502743Z" level=error msg="get state for f7de11d8881e6e308a3ed93d93966d03d6d1f9a6eb22bb9614f5a20a49cd100a" error="context deadline exceeded" Apr 20 19:44:34.481615 containerd[1648]: time="2026-04-20T19:44:33.277206478Z" level=error msg="ttrpc: received message on inactive stream" stream=45 Apr 20 19:44:41.670450 containerd[1648]: time="2026-04-20T19:44:32.580171840Z" level=warning msg="unknown status" status=0 Apr 20 19:45:04.157121 kubelet[3245]: I0420 19:45:04.094574 3245 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" Apr 20 19:45:13.381630 kubelet[3245]: E0420 19:45:13.174600 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:45:15.359711 kubelet[3245]: I0420 19:45:11.728056 3245 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" Apr 20 19:45:20.423378 kubelet[3245]: E0420 19:45:11.168132 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:45:27.489704 kubelet[3245]: I0420 19:45:03.637788 3245 reflector.go:556] "Warning: watch ended with error" reflector="pkg/kubelet/config/apiserver.go:66" type="*v1.Pod" err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" Apr 20 19:45:29.711478 kubelet[3245]: E0420 19:45:20.990472 3245 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": Get \"https://10.0.0.18:6443/api/v1/nodes/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:45:29.711478 kubelet[3245]: I0420 19:45:14.738516 3245 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" Apr 20 19:45:34.380112 kubelet[3245]: I0420 19:45:22.055507 3245 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" Apr 20 19:45:37.781144 kubelet[3245]: E0420 19:45:20.840124 3245 log.go:32] "ListImages with filter from image service failed" err="rpc error: code = DeadlineExceeded desc = context deadline exceeded" filter="nil" Apr 20 19:45:39.717220 kubelet[3245]: E0420 19:45:39.696278 3245 kuberuntime_image.go:104] "Failed to list images" err="rpc error: code = DeadlineExceeded desc = context deadline exceeded" Apr 20 19:45:44.486693 kubelet[3245]: E0420 19:45:03.690210 3245 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/events/kube-scheduler-localhost.18a828038d95e43d\": http2: client connection lost" event="&Event{ObjectMeta:{kube-scheduler-localhost.18a828038d95e43d kube-system 678 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-scheduler-localhost,UID:33fee6ba1581201eda98a989140db110,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://127.0.0.1:10259/readyz\": dial tcp 127.0.0.1:10259: connect: connection refused,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:40:48 +0000 UTC,LastTimestamp:2026-04-20 19:40:51.056876277 +0000 UTC m=+909.966783905,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:45:47.748662 kubelet[3245]: E0420 19:45:47.716267 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:45:49.275423 kubelet[3245]: I0420 19:45:43.169138 3245 image_gc_manager.go:222] "Failed to monitor images" err="rpc error: code = DeadlineExceeded desc = context deadline exceeded" Apr 20 19:46:01.893979 kubelet[3245]: E0420 19:46:01.883879 3245 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4m33.293s" Apr 20 19:46:04.937127 kubelet[3245]: E0420 19:46:04.893358 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m43.554145859s ago; threshold is 3m0s" Apr 20 19:46:06.770672 kubelet[3245]: E0420 19:46:05.381134 3245 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": Get \"https://10.0.0.18:6443/api/v1/nodes/localhost?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Apr 20 19:46:07.068201 kubelet[3245]: E0420 19:46:05.882770 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Apr 20 19:46:07.815592 kubelet[3245]: E0420 19:46:07.786913 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m46.415237256s ago; threshold is 3m0s" Apr 20 19:46:09.153094 containerd[1648]: time="2026-04-20T19:46:08.554207646Z" level=error msg="get state for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="context deadline exceeded" Apr 20 19:46:10.344262 containerd[1648]: time="2026-04-20T19:46:08.563920254Z" level=warning msg="unknown status" status=0 Apr 20 19:46:11.299213 containerd[1648]: time="2026-04-20T19:46:09.653664162Z" level=error msg="ttrpc: received message on inactive stream" stream=125 Apr 20 19:46:13.956905 containerd[1648]: time="2026-04-20T19:46:13.633576691Z" level=error msg="failed to delete task" error="context deadline exceeded" id=78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 Apr 20 19:46:14.828974 containerd[1648]: time="2026-04-20T19:46:14.634738689Z" level=error msg="failed to drain init process 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7 io" error="context deadline exceeded" runtime=io.containerd.runc.v2 Apr 20 19:46:15.372839 containerd[1648]: time="2026-04-20T19:46:14.820032551Z" level=error msg="Failed to handle backOff event container_id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" id:\"78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7\" pid:3006 exit_status:1 exited_at:{seconds:1776712981 nanos:448571401} for 78e0aaf8381388eb3a91b257a3ba0713bc82b4b5f8e15af8b94cec76d08e4ae7" error="failed to handle container TaskExit event: failed to stop container: failed to delete task: context deadline exceeded" Apr 20 19:46:18.119702 containerd[1648]: time="2026-04-20T19:46:18.065381926Z" level=error msg="ttrpc: received message on inactive stream" stream=127 Apr 20 19:46:18.325624 kubelet[3245]: E0420 19:46:16.823901 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m46.760736628s ago; threshold is 3m0s" Apr 20 19:46:18.890908 kubelet[3245]: E0420 19:46:18.803893 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m57.466567047s ago; threshold is 3m0s" Apr 20 19:46:20.181886 kubelet[3245]: E0420 19:46:19.351074 3245 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": Get \"https://10.0.0.18:6443/api/v1/nodes/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:46:20.679369 kubelet[3245]: E0420 19:46:18.627620 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?resourceVersion=640\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:46:20.679369 kubelet[3245]: E0420 19:46:17.925912 3245 log.go:32] "ListContainers with filter from runtime service failed" err="rpc error: code = DeadlineExceeded desc = context deadline exceeded" filter="&ContainerFilter{Id:,State:nil,PodSandboxId:,LabelSelector:map[string]string{},}" Apr 20 19:46:20.868825 kubelet[3245]: E0420 19:46:20.724699 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m58.336643253s ago; threshold is 3m0s" Apr 20 19:46:21.947914 containerd[1648]: time="2026-04-20T19:46:20.520355019Z" level=info msg="TaskExit event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942}" Apr 20 19:46:22.944566 kubelet[3245]: E0420 19:46:20.520346 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:46:24.600702 kubelet[3245]: I0420 19:46:24.575891 3245 status_manager.go:895] "Failed to get status for pod" podUID="e9ca41790ae21be9f4cbd451ade0acec" pod="kube-system/kube-controller-manager-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-localhost\": net/http: TLS handshake timeout - error from a previous attempt: http2: client connection lost" Apr 20 19:46:24.865920 kubelet[3245]: E0420 19:46:20.253626 3245 kubelet_node_status.go:535] "Unable to update node status" err="update node status exceeds retry count" Apr 20 19:46:25.224424 kubelet[3245]: E0420 19:46:25.123483 3245 container_log_manager.go:197] "Failed to rotate container logs" err="failed to list containers: rpc error: code = DeadlineExceeded desc = context deadline exceeded" Apr 20 19:46:26.541690 kubelet[3245]: E0420 19:46:25.481011 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 4m1.111629419s ago; threshold is 3m0s" Apr 20 19:46:30.783145 kubelet[3245]: E0420 19:46:30.752749 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime is down, PLEG is not healthy: pleg was last seen active 4m8.47694231s ago; threshold is 3m0s]" Apr 20 19:46:34.207205 kubelet[3245]: E0420 19:46:32.648115 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:46:37.809081 kubelet[3245]: E0420 19:46:37.659608 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&resourceVersion=642\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:46:41.361283 kubelet[3245]: E0420 19:46:41.350040 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&resourceVersion=656\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:46:42.929160 kubelet[3245]: E0420 19:46:28.030751 3245 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/events/kube-scheduler-localhost.18a828038d95e43d\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{kube-scheduler-localhost.18a828038d95e43d kube-system 678 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-scheduler-localhost,UID:33fee6ba1581201eda98a989140db110,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://127.0.0.1:10259/readyz\": dial tcp 127.0.0.1:10259: connect: connection refused,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:40:48 +0000 UTC,LastTimestamp:2026-04-20 19:40:51.056876277 +0000 UTC m=+909.966783905,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:46:42.929160 kubelet[3245]: E0420 19:46:41.787787 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m28.890320803s ago; threshold is 3m0s" Apr 20 19:46:42.929160 kubelet[3245]: E0420 19:46:41.852091 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Pod: Get \"https://10.0.0.18:6443/api/v1/pods?fieldSelector=spec.nodeName%3Dlocalhost&resourceVersion=686\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="pkg/kubelet/config/apiserver.go:66" type="*v1.Pod" Apr 20 19:46:45.976980 kubelet[3245]: E0420 19:46:36.946695 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.18:6443/apis/storage.k8s.io/v1/csidrivers?resourceVersion=657\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 20 19:46:46.765910 kubelet[3245]: I0420 19:46:46.756939 3245 status_manager.go:895] "Failed to get status for pod" podUID="33fee6ba1581201eda98a989140db110" pod="kube-system/kube-scheduler-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-scheduler-localhost\": net/http: TLS handshake timeout" Apr 20 19:46:46.883164 kubelet[3245]: E0420 19:46:46.882423 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m39.869275008s ago; threshold is 3m0s" Apr 20 19:46:48.978344 kubelet[3245]: E0420 19:46:48.947189 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" Apr 20 19:46:52.407654 kubelet[3245]: I0420 19:46:52.372762 3245 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Apr 20 19:46:54.623625 containerd[1648]: time="2026-04-20T19:46:54.566137665Z" level=error msg="ttrpc: received message on inactive stream" stream=131 Apr 20 19:46:54.969492 containerd[1648]: time="2026-04-20T19:46:54.738009559Z" level=error msg="Failed to handle backOff event container_id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" id:\"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" pid:3024 exit_status:1 exited_at:{seconds:1776712806 nanos:309134942} for edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00" error="failed to handle container TaskExit event: failed to stop container: context deadline exceeded" Apr 20 19:46:54.998951 kubelet[3245]: E0420 19:46:51.606705 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?resourceVersion=640\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:46:55.051773 kubelet[3245]: E0420 19:46:55.017126 3245 kubelet.go:2460] "Skipping pod synchronization" err="PLEG is not healthy: pleg was last seen active 3m44.888774721s ago; threshold is 3m0s" Apr 20 19:46:55.148041 containerd[1648]: time="2026-04-20T19:46:54.928034900Z" level=error msg="ttrpc: received message on inactive stream" stream=133 Apr 20 19:47:13.034563 kubelet[3245]: E0420 19:47:13.022874 3245 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="200ms" Apr 20 19:47:15.481740 kubelet[3245]: E0420 19:47:14.319768 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime is down, PLEG is not healthy: pleg was last seen active 3m58.952751761s ago; threshold is 3m0s]" Apr 20 19:47:16.296399 kubelet[3245]: E0420 19:47:16.116495 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Pod: Get \"https://10.0.0.18:6443/api/v1/pods?fieldSelector=spec.nodeName%3Dlocalhost&resourceVersion=686\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="pkg/kubelet/config/apiserver.go:66" type="*v1.Pod" Apr 20 19:47:19.162292 kubelet[3245]: E0420 19:47:19.160844 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&resourceVersion=642\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:47:19.708545 kubelet[3245]: I0420 19:47:19.153813 3245 status_manager.go:895] "Failed to get status for pod" podUID="fde15e5d4357efe755cb792ce81eb1a9" pod="kube-system/kube-apiserver-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-localhost\": net/http: TLS handshake timeout" Apr 20 19:47:33.035107 kubelet[3245]: E0420 19:47:12.568622 3245 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/events/kube-scheduler-localhost.18a828038d95e43d\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{kube-scheduler-localhost.18a828038d95e43d kube-system 678 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-scheduler-localhost,UID:33fee6ba1581201eda98a989140db110,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://127.0.0.1:10259/readyz\": dial tcp 127.0.0.1:10259: connect: connection refused,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:40:48 +0000 UTC,LastTimestamp:2026-04-20 19:40:51.056876277 +0000 UTC m=+909.966783905,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:47:34.424977 kubelet[3245]: E0420 19:47:28.739901 3245 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime is down, PLEG is not healthy: pleg was last seen active 4m17.036904863s ago; threshold is 3m0s]" Apr 20 19:47:41.474874 kubelet[3245]: E0420 19:47:37.435529 3245 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="400ms" Apr 20 19:47:42.258974 kubelet[3245]: E0420 19:47:40.175016 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:47:42.965183 kubelet[3245]: E0420 19:47:40.888351 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.18:6443/apis/storage.k8s.io/v1/csidrivers?resourceVersion=657\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 20 19:47:49.396818 kubelet[3245]: E0420 19:47:49.383151 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.18:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&resourceVersion=656\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 20 19:47:52.089655 kubelet[3245]: E0420 19:47:49.385142 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.18:6443/apis/node.k8s.io/v1/runtimeclasses?resourceVersion=640\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 20 19:47:55.878156 kubelet[3245]: I0420 19:47:55.856734 3245 status_manager.go:895] "Failed to get status for pod" podUID="e9ca41790ae21be9f4cbd451ade0acec" pod="kube-system/kube-controller-manager-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-localhost\": net/http: TLS handshake timeout" Apr 20 19:47:57.246975 kubelet[3245]: E0420 19:47:57.241721 3245 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded" interval="800ms" Apr 20 19:48:01.054495 kubelet[3245]: E0420 19:48:01.054155 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Pod: Get \"https://10.0.0.18:6443/api/v1/pods?fieldSelector=spec.nodeName%3Dlocalhost&resourceVersion=686\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="pkg/kubelet/config/apiserver.go:66" type="*v1.Pod" Apr 20 19:48:03.124112 kubelet[3245]: E0420 19:48:01.054323 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:48:06.362172 kubelet[3245]: E0420 19:48:06.352540 3245 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&resourceVersion=642\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 20 19:48:18.066830 kubelet[3245]: E0420 19:48:16.466113 3245 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="1.6s" Apr 20 19:48:20.158954 kubelet[3245]: E0420 19:48:20.150435 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:48:31.226010 kubelet[3245]: E0420 19:48:29.777902 3245 kubelet_node_status.go:548] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:47:59Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:47:59Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:47:59Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-04-20T19:47:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized\\\",\\\"type\\\":\\\"Ready\\\"}]}}\" for node \"localhost\": Patch \"https://10.0.0.18:6443/api/v1/nodes/localhost/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 20 19:48:45.978049 kubelet[3245]: E0420 19:48:45.932046 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:48:57.456642 kubelet[3245]: E0420 19:48:57.454671 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:49:09.849761 kubelet[3245]: E0420 19:49:09.840959 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:49:25.955890 kubelet[3245]: E0420 19:49:24.708154 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:49:32.926965 kubelet[3245]: E0420 19:49:32.188970 3245 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1m43.963s" Apr 20 19:49:39.644962 kubelet[3245]: E0420 19:49:39.182329 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:49:45.138915 kubelet[3245]: E0420 19:49:45.094201 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:49:55.531052 kubelet[3245]: E0420 19:49:55.526608 3245 request.go:1360] "Unexpected error when reading response body" err="context deadline exceeded (Client.Timeout or context cancellation while reading body)" Apr 20 19:49:56.979349 kubelet[3245]: E0420 19:49:55.564771 3245 controller.go:195] "Failed to update lease" err="unexpected error when reading response body. Please retry. Original error: context deadline exceeded (Client.Timeout or context cancellation while reading body)" Apr 20 19:49:57.354754 kubelet[3245]: E0420 19:49:57.334073 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:50:04.495101 containerd[1648]: time="2026-04-20T19:50:04.141675528Z" level=info msg="StopContainer for \"5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a\" with timeout 30 (s)" Apr 20 19:50:09.010561 containerd[1648]: time="2026-04-20T19:50:08.970179493Z" level=info msg="Stop container \"5c3ab89ca9d7b8a0909f1b3a75dcc3de59959c712621badfb65c0bf1c6a0b15a\" with signal terminated" Apr 20 19:50:09.765463 kubelet[3245]: E0420 19:50:09.764750 3245 controller.go:195] "Failed to update lease" err="Operation cannot be fulfilled on leases.coordination.k8s.io \"localhost\": the object has been modified; please apply your changes to the latest version and try again" Apr 20 19:50:10.891844 containerd[1648]: time="2026-04-20T19:50:10.711611142Z" level=info msg="StopContainer for \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" with timeout 30 (s)" Apr 20 19:50:12.657947 kubelet[3245]: E0420 19:50:12.568995 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:50:17.073960 containerd[1648]: time="2026-04-20T19:50:15.700103556Z" level=info msg="Skipping the sending of signal terminated to container \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\" because a prior stop with timeout>0 request already sent the signal" Apr 20 19:50:32.583356 kubelet[3245]: E0420 19:50:32.547161 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:50:36.580189 kubelet[3245]: E0420 19:50:35.982064 3245 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 20 19:50:41.641522 kubelet[3245]: E0420 19:50:41.640927 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:50:42.131364 kubelet[3245]: E0420 19:50:42.128741 3245 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="55.834s" Apr 20 19:50:42.152982 kubelet[3245]: E0420 19:50:42.121765 3245 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/events/kube-controller-manager-localhost.18a82802c0debbc3\": dial tcp 10.0.0.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-controller-manager-localhost.18a82802c0debbc3 kube-system 779 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-controller-manager-localhost,UID:e9ca41790ae21be9f4cbd451ade0acec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://127.0.0.1:10257/healthz\": dial tcp 127.0.0.1:10257: connect: connection refused,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:40:45 +0000 UTC,LastTimestamp:2026-04-20 19:41:19.520761386 +0000 UTC m=+938.430669017,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:50:42.152982 kubelet[3245]: I0420 19:50:42.150864 3245 status_manager.go:895] "Failed to get status for pod" podUID="fde15e5d4357efe755cb792ce81eb1a9" pod="kube-system/kube-apiserver-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:42.835786 kubelet[3245]: I0420 19:50:42.820825 3245 status_manager.go:895] "Failed to get status for pod" podUID="e9ca41790ae21be9f4cbd451ade0acec" pod="kube-system/kube-controller-manager-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:43.012639 kubelet[3245]: I0420 19:50:43.011148 3245 status_manager.go:895] "Failed to get status for pod" podUID="fde15e5d4357efe755cb792ce81eb1a9" pod="kube-system/kube-apiserver-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:44.726293 kubelet[3245]: I0420 19:50:44.686909 3245 status_manager.go:895] "Failed to get status for pod" podUID="e9ca41790ae21be9f4cbd451ade0acec" pod="kube-system/kube-controller-manager-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:44.726293 kubelet[3245]: E0420 19:50:44.723724 3245 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.572s" Apr 20 19:50:45.646160 kubelet[3245]: I0420 19:50:45.568879 3245 status_manager.go:895] "Failed to get status for pod" podUID="fde15e5d4357efe755cb792ce81eb1a9" pod="kube-system/kube-apiserver-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:46.093060 kubelet[3245]: I0420 19:50:46.088963 3245 status_manager.go:895] "Failed to get status for pod" podUID="e9ca41790ae21be9f4cbd451ade0acec" pod="kube-system/kube-controller-manager-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:47.434785 containerd[1648]: time="2026-04-20T19:50:47.428001239Z" level=info msg="Kill container \"edeea07d5135bcece8558cfd8bde9f935ec8189ae62173e7929202e08a9cdd00\"" Apr 20 19:50:49.652812 kubelet[3245]: E0420 19:50:49.649315 3245 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 20 19:50:51.242404 kubelet[3245]: I0420 19:50:51.208714 3245 status_manager.go:895] "Failed to get status for pod" podUID="fde15e5d4357efe755cb792ce81eb1a9" pod="kube-system/kube-apiserver-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:51.701190 kubelet[3245]: E0420 19:50:51.566870 3245 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/events/kube-controller-manager-localhost.18a82802c0debbc3\": dial tcp 10.0.0.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-controller-manager-localhost.18a82802c0debbc3 kube-system 779 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-controller-manager-localhost,UID:e9ca41790ae21be9f4cbd451ade0acec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://127.0.0.1:10257/healthz\": dial tcp 127.0.0.1:10257: connect: connection refused,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-04-20 19:40:45 +0000 UTC,LastTimestamp:2026-04-20 19:41:19.520761386 +0000 UTC m=+938.430669017,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Apr 20 19:50:52.034089 kubelet[3245]: I0420 19:50:51.752402 3245 status_manager.go:895] "Failed to get status for pod" podUID="e9ca41790ae21be9f4cbd451ade0acec" pod="kube-system/kube-controller-manager-localhost" err="Get \"https://10.0.0.18:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-localhost\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:52.754738 kubelet[3245]: E0420 19:50:52.646019 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:53.443212 kubelet[3245]: E0420 19:50:53.430050 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:53.513793 kubelet[3245]: E0420 19:50:53.512850 3245 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="8.789s" Apr 20 19:50:53.530010 kubelet[3245]: E0420 19:50:53.514867 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused" Apr 20 19:50:53.661891 kubelet[3245]: E0420 19:50:53.661626 3245 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.18:6443: connect: connection refused"