Mar 2 13:26:01.125187 kernel: Linux version 6.12.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Mon Mar 2 10:54:34 -00 2026 Mar 2 13:26:01.125290 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ecd171dfe44947e01ea4ccf8f02472d94712b039722a1ae86ca9989f56ff7a41 Mar 2 13:26:01.125314 kernel: BIOS-provided physical RAM map: Mar 2 13:26:01.125323 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Mar 2 13:26:01.125332 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Mar 2 13:26:01.125341 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Mar 2 13:26:01.125354 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Mar 2 13:26:01.125364 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Mar 2 13:26:01.125423 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Mar 2 13:26:01.125433 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Mar 2 13:26:01.125447 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 2 13:26:01.125459 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Mar 2 13:26:01.125468 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 2 13:26:01.125477 kernel: NX (Execute Disable) protection: active Mar 2 13:26:01.125488 kernel: APIC: Static calls initialized Mar 2 13:26:01.125505 kernel: SMBIOS 2.8 present. Mar 2 13:26:01.125563 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Mar 2 13:26:01.125574 kernel: DMI: Memory slots populated: 1/1 Mar 2 13:26:01.125583 kernel: Hypervisor detected: KVM Mar 2 13:26:01.125595 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Mar 2 13:26:01.125608 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 2 13:26:01.125617 kernel: kvm-clock: using sched offset of 61257404550 cycles Mar 2 13:26:01.125628 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 2 13:26:01.125639 kernel: tsc: Detected 2445.426 MHz processor Mar 2 13:26:01.125658 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 2 13:26:01.125668 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 2 13:26:01.125678 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Mar 2 13:26:01.125689 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Mar 2 13:26:01.125703 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 2 13:26:01.125713 kernel: Using GB pages for direct mapping Mar 2 13:26:01.125723 kernel: ACPI: Early table checksum verification disabled Mar 2 13:26:01.125739 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Mar 2 13:26:01.125751 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126580 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126593 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126604 kernel: ACPI: FACS 0x000000009CFE0000 000040 Mar 2 13:26:01.126615 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126626 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126644 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126658 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:26:01.126674 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Mar 2 13:26:01.126688 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Mar 2 13:26:01.126699 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Mar 2 13:26:01.126717 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Mar 2 13:26:01.126728 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Mar 2 13:26:01.126741 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Mar 2 13:26:01.126754 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Mar 2 13:26:01.126866 kernel: No NUMA configuration found Mar 2 13:26:01.126881 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Mar 2 13:26:01.126898 kernel: NODE_DATA(0) allocated [mem 0x9cfd4dc0-0x9cfdbfff] Mar 2 13:26:01.126912 kernel: Zone ranges: Mar 2 13:26:01.126923 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 2 13:26:01.126935 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Mar 2 13:26:01.126948 kernel: Normal empty Mar 2 13:26:01.126962 kernel: Device empty Mar 2 13:26:01.126973 kernel: Movable zone start for each node Mar 2 13:26:01.126986 kernel: Early memory node ranges Mar 2 13:26:01.127003 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Mar 2 13:26:01.127017 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Mar 2 13:26:01.127028 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Mar 2 13:26:01.127042 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 2 13:26:01.127055 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Mar 2 13:26:01.127118 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Mar 2 13:26:01.127134 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 2 13:26:01.127150 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 2 13:26:01.127163 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 2 13:26:01.127177 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 2 13:26:01.128465 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 2 13:26:01.128484 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 2 13:26:01.128496 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 2 13:26:01.128508 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 2 13:26:01.128529 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 2 13:26:01.128540 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 2 13:26:01.128550 kernel: TSC deadline timer available Mar 2 13:26:01.128562 kernel: CPU topo: Max. logical packages: 1 Mar 2 13:26:01.128575 kernel: CPU topo: Max. logical dies: 1 Mar 2 13:26:01.128587 kernel: CPU topo: Max. dies per package: 1 Mar 2 13:26:01.128597 kernel: CPU topo: Max. threads per core: 1 Mar 2 13:26:01.128608 kernel: CPU topo: Num. cores per package: 4 Mar 2 13:26:01.128627 kernel: CPU topo: Num. threads per package: 4 Mar 2 13:26:01.128638 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Mar 2 13:26:01.128648 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 2 13:26:01.128659 kernel: kvm-guest: KVM setup pv remote TLB flush Mar 2 13:26:01.128672 kernel: kvm-guest: setup PV sched yield Mar 2 13:26:01.128684 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Mar 2 13:26:01.128694 kernel: Booting paravirtualized kernel on KVM Mar 2 13:26:01.128712 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 2 13:26:01.128725 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Mar 2 13:26:01.128736 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Mar 2 13:26:01.128746 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Mar 2 13:26:01.128758 kernel: pcpu-alloc: [0] 0 1 2 3 Mar 2 13:26:01.128877 kernel: kvm-guest: PV spinlocks enabled Mar 2 13:26:01.128890 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 2 13:26:01.128906 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ecd171dfe44947e01ea4ccf8f02472d94712b039722a1ae86ca9989f56ff7a41 Mar 2 13:26:01.128918 kernel: random: crng init done Mar 2 13:26:01.128931 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 2 13:26:01.128943 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 2 13:26:01.128953 kernel: Fallback order for Node 0: 0 Mar 2 13:26:01.128965 kernel: Built 1 zonelists, mobility grouping on. Total pages: 642938 Mar 2 13:26:01.128983 kernel: Policy zone: DMA32 Mar 2 13:26:01.128994 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 2 13:26:01.129004 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 2 13:26:01.129017 kernel: ftrace: allocating 40130 entries in 157 pages Mar 2 13:26:01.129029 kernel: ftrace: allocated 157 pages with 5 groups Mar 2 13:26:01.129040 kernel: Dynamic Preempt: voluntary Mar 2 13:26:01.129051 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 2 13:26:01.129072 kernel: rcu: RCU event tracing is enabled. Mar 2 13:26:01.129087 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 2 13:26:01.129099 kernel: Trampoline variant of Tasks RCU enabled. Mar 2 13:26:01.129158 kernel: Rude variant of Tasks RCU enabled. Mar 2 13:26:01.129172 kernel: Tracing variant of Tasks RCU enabled. Mar 2 13:26:01.129183 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 2 13:26:01.129194 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 2 13:26:01.129206 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 13:26:01.132387 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 13:26:01.132414 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 13:26:01.132426 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Mar 2 13:26:01.132438 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 2 13:26:01.132467 kernel: Console: colour VGA+ 80x25 Mar 2 13:26:01.132482 kernel: printk: legacy console [ttyS0] enabled Mar 2 13:26:01.132495 kernel: ACPI: Core revision 20240827 Mar 2 13:26:01.132508 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 2 13:26:01.132519 kernel: APIC: Switch to symmetric I/O mode setup Mar 2 13:26:01.132535 kernel: x2apic enabled Mar 2 13:26:01.132550 kernel: APIC: Switched APIC routing to: physical x2apic Mar 2 13:26:01.132615 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Mar 2 13:26:01.132628 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Mar 2 13:26:01.132651 kernel: kvm-guest: setup PV IPIs Mar 2 13:26:01.132664 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 2 13:26:01.132675 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Mar 2 13:26:01.132687 kernel: Calibrating delay loop (skipped) preset value.. 4890.85 BogoMIPS (lpj=2445426) Mar 2 13:26:01.132699 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 2 13:26:01.132713 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 2 13:26:01.132724 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 2 13:26:01.132740 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 2 13:26:01.132754 kernel: Spectre V2 : Mitigation: Retpolines Mar 2 13:26:01.132983 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Mar 2 13:26:01.133000 kernel: Speculative Store Bypass: Vulnerable Mar 2 13:26:01.133012 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Mar 2 13:26:01.133024 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Mar 2 13:26:01.133043 kernel: active return thunk: srso_alias_return_thunk Mar 2 13:26:01.133056 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Mar 2 13:26:01.133067 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Mar 2 13:26:01.133078 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Mar 2 13:26:01.133093 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 2 13:26:01.133105 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 2 13:26:01.133115 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 2 13:26:01.133133 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 2 13:26:01.133146 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 2 13:26:01.133157 kernel: Freeing SMP alternatives memory: 32K Mar 2 13:26:01.133168 kernel: pid_max: default: 32768 minimum: 301 Mar 2 13:26:01.133182 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Mar 2 13:26:01.133193 kernel: landlock: Up and running. Mar 2 13:26:01.133204 kernel: SELinux: Initializing. Mar 2 13:26:01.133285 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 13:26:01.133303 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 13:26:01.133366 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Mar 2 13:26:01.133380 kernel: Performance Events: PMU not available due to virtualization, using software events only. Mar 2 13:26:01.133393 kernel: signal: max sigframe size: 1776 Mar 2 13:26:01.133408 kernel: rcu: Hierarchical SRCU implementation. Mar 2 13:26:01.133421 kernel: rcu: Max phase no-delay instances is 400. Mar 2 13:26:01.133438 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Mar 2 13:26:01.139458 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 2 13:26:01.139475 kernel: smp: Bringing up secondary CPUs ... Mar 2 13:26:01.139487 kernel: smpboot: x86: Booting SMP configuration: Mar 2 13:26:01.139498 kernel: .... node #0, CPUs: #1 #2 #3 Mar 2 13:26:01.139509 kernel: smp: Brought up 1 node, 4 CPUs Mar 2 13:26:01.139522 kernel: smpboot: Total of 4 processors activated (19563.40 BogoMIPS) Mar 2 13:26:01.139537 kernel: Memory: 2445296K/2571752K available (14336K kernel code, 2445K rwdata, 31644K rodata, 15544K init, 2492K bss, 120520K reserved, 0K cma-reserved) Mar 2 13:26:01.139557 kernel: devtmpfs: initialized Mar 2 13:26:01.139568 kernel: x86/mm: Memory block size: 128MB Mar 2 13:26:01.139580 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 2 13:26:01.139591 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 2 13:26:01.139602 kernel: pinctrl core: initialized pinctrl subsystem Mar 2 13:26:01.139613 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 2 13:26:01.139624 kernel: audit: initializing netlink subsys (disabled) Mar 2 13:26:01.139639 kernel: audit: type=2000 audit(1772457905.262:1): state=initialized audit_enabled=0 res=1 Mar 2 13:26:01.139650 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 2 13:26:01.139661 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 2 13:26:01.139673 kernel: cpuidle: using governor menu Mar 2 13:26:01.139732 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 2 13:26:01.139745 kernel: dca service started, version 1.12.1 Mar 2 13:26:01.139758 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Mar 2 13:26:01.139860 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] reserved as E820 entry Mar 2 13:26:01.139872 kernel: PCI: Using configuration type 1 for base access Mar 2 13:26:01.139883 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 2 13:26:01.139894 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 2 13:26:01.139905 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 2 13:26:01.139917 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 2 13:26:01.139928 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 2 13:26:01.139942 kernel: ACPI: Added _OSI(Module Device) Mar 2 13:26:01.139953 kernel: ACPI: Added _OSI(Processor Device) Mar 2 13:26:01.139964 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 2 13:26:01.139976 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 2 13:26:01.139987 kernel: ACPI: Interpreter enabled Mar 2 13:26:01.139998 kernel: ACPI: PM: (supports S0 S3 S5) Mar 2 13:26:01.140009 kernel: ACPI: Using IOAPIC for interrupt routing Mar 2 13:26:01.140023 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 2 13:26:01.140034 kernel: PCI: Using E820 reservations for host bridge windows Mar 2 13:26:01.140045 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 2 13:26:01.140057 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 2 13:26:01.140572 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 2 13:26:01.141084 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 2 13:26:01.143590 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 2 13:26:01.143612 kernel: PCI host bridge to bus 0000:00 Mar 2 13:26:01.144066 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 2 13:26:01.144407 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 2 13:26:01.144675 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 2 13:26:01.145078 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Mar 2 13:26:01.145428 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Mar 2 13:26:01.145671 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Mar 2 13:26:01.146049 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 2 13:26:01.156879 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Mar 2 13:26:01.157549 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Mar 2 13:26:01.171547 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfd000000-0xfdffffff pref] Mar 2 13:26:01.171981 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfebd0000-0xfebd0fff] Mar 2 13:26:01.172361 kernel: pci 0000:00:01.0: ROM [mem 0xfebc0000-0xfebcffff pref] Mar 2 13:26:01.172657 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 2 13:26:01.173101 kernel: pci 0000:00:01.0: pci_fixup_video+0x0/0x100 took 50781 usecs Mar 2 13:26:01.173478 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Mar 2 13:26:01.173897 kernel: pci 0000:00:02.0: BAR 0 [io 0xc0c0-0xc0df] Mar 2 13:26:01.174212 kernel: pci 0000:00:02.0: BAR 1 [mem 0xfebd1000-0xfebd1fff] Mar 2 13:26:01.183293 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfe000000-0xfe003fff 64bit pref] Mar 2 13:26:01.183651 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Mar 2 13:26:01.184068 kernel: pci 0000:00:03.0: BAR 0 [io 0xc000-0xc07f] Mar 2 13:26:01.184446 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebd2000-0xfebd2fff] Mar 2 13:26:01.184746 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe004000-0xfe007fff 64bit pref] Mar 2 13:26:01.185151 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Mar 2 13:26:01.185523 kernel: pci 0000:00:04.0: BAR 0 [io 0xc0e0-0xc0ff] Mar 2 13:26:01.185949 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebd3000-0xfebd3fff] Mar 2 13:26:01.200087 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe008000-0xfe00bfff 64bit pref] Mar 2 13:26:01.200623 kernel: pci 0000:00:04.0: ROM [mem 0xfeb80000-0xfebbffff pref] Mar 2 13:26:01.202223 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Mar 2 13:26:01.202606 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 2 13:26:01.203080 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Mar 2 13:26:01.203458 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc100-0xc11f] Mar 2 13:26:01.203884 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfebd4000-0xfebd4fff] Mar 2 13:26:01.204217 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Mar 2 13:26:01.214100 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Mar 2 13:26:01.214132 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 2 13:26:01.214146 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 2 13:26:01.214159 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 2 13:26:01.214173 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 2 13:26:01.214202 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 2 13:26:01.214215 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 2 13:26:01.214580 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 2 13:26:01.214598 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 2 13:26:01.214612 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 2 13:26:01.214627 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 2 13:26:01.214696 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 2 13:26:01.214720 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 2 13:26:01.214735 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 2 13:26:01.214746 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 2 13:26:01.214757 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 2 13:26:01.214864 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 2 13:26:01.214877 kernel: iommu: Default domain type: Translated Mar 2 13:26:01.214888 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 2 13:26:01.214906 kernel: PCI: Using ACPI for IRQ routing Mar 2 13:26:01.214918 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 2 13:26:01.214930 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Mar 2 13:26:01.214944 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Mar 2 13:26:01.215336 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 2 13:26:01.215637 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 2 13:26:01.216030 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 2 13:26:01.216056 kernel: vgaarb: loaded Mar 2 13:26:01.216071 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 2 13:26:01.216084 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 2 13:26:01.216095 kernel: clocksource: Switched to clocksource kvm-clock Mar 2 13:26:01.216107 kernel: VFS: Disk quotas dquot_6.6.0 Mar 2 13:26:01.216122 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 2 13:26:01.216133 kernel: pnp: PnP ACPI init Mar 2 13:26:01.216592 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Mar 2 13:26:01.216614 kernel: pnp: PnP ACPI: found 6 devices Mar 2 13:26:01.216627 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 2 13:26:01.216639 kernel: NET: Registered PF_INET protocol family Mar 2 13:26:01.216652 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 2 13:26:01.216667 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 2 13:26:01.216685 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 2 13:26:01.216697 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 2 13:26:01.216711 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 2 13:26:01.216725 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 2 13:26:01.216737 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 13:26:01.216751 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 13:26:01.216871 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 2 13:26:01.216892 kernel: NET: Registered PF_XDP protocol family Mar 2 13:26:01.217184 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 2 13:26:01.229973 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 2 13:26:01.232647 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 2 13:26:01.233098 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Mar 2 13:26:01.236919 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Mar 2 13:26:01.237214 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Mar 2 13:26:01.237313 kernel: PCI: CLS 0 bytes, default 64 Mar 2 13:26:01.237326 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Mar 2 13:26:01.237338 kernel: Initialise system trusted keyrings Mar 2 13:26:01.237351 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 2 13:26:01.237365 kernel: Key type asymmetric registered Mar 2 13:26:01.237376 kernel: Asymmetric key parser 'x509' registered Mar 2 13:26:01.237387 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 2 13:26:01.237406 kernel: io scheduler mq-deadline registered Mar 2 13:26:01.237419 kernel: io scheduler kyber registered Mar 2 13:26:01.237430 kernel: io scheduler bfq registered Mar 2 13:26:01.237442 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 2 13:26:01.237457 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 2 13:26:01.237469 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 2 13:26:01.237480 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 2 13:26:01.237500 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 2 13:26:01.237512 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 2 13:26:01.237523 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 2 13:26:01.237537 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 2 13:26:01.237549 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 2 13:26:01.238025 kernel: rtc_cmos 00:04: RTC can wake from S4 Mar 2 13:26:01.243424 kernel: rtc_cmos 00:04: registered as rtc0 Mar 2 13:26:01.243460 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1 Mar 2 13:26:01.243753 kernel: rtc_cmos 00:04: setting system clock to 2026-03-02T13:25:34 UTC (1772457934) Mar 2 13:26:01.244159 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Mar 2 13:26:01.244182 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 2 13:26:01.244196 kernel: NET: Registered PF_INET6 protocol family Mar 2 13:26:01.244209 kernel: Segment Routing with IPv6 Mar 2 13:26:01.244222 kernel: In-situ OAM (IOAM) with IPv6 Mar 2 13:26:01.244311 kernel: NET: Registered PF_PACKET protocol family Mar 2 13:26:01.244325 kernel: Key type dns_resolver registered Mar 2 13:26:01.244338 kernel: IPI shorthand broadcast: enabled Mar 2 13:26:01.244352 kernel: sched_clock: Marking stable (19883044784, 6121046274)->(31758558422, -5754467364) Mar 2 13:26:01.244364 kernel: registered taskstats version 1 Mar 2 13:26:01.244377 kernel: Loading compiled-in X.509 certificates Mar 2 13:26:01.244390 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.74-flatcar: 9223f79438693c88653fb3077b0a0a0f10ebb9fb' Mar 2 13:26:01.244408 kernel: Demotion targets for Node 0: null Mar 2 13:26:01.244420 kernel: Key type .fscrypt registered Mar 2 13:26:01.244433 kernel: Key type fscrypt-provisioning registered Mar 2 13:26:01.244445 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 2 13:26:01.244457 kernel: ima: Allocated hash algorithm: sha1 Mar 2 13:26:01.244471 kernel: ima: No architecture policies found Mar 2 13:26:01.249414 kernel: clk: Disabling unused clocks Mar 2 13:26:01.249440 kernel: Freeing unused kernel image (initmem) memory: 15544K Mar 2 13:26:01.251449 kernel: Write protecting the kernel read-only data: 47104k Mar 2 13:26:01.251466 kernel: Freeing unused kernel image (rodata/data gap) memory: 1124K Mar 2 13:26:01.251479 kernel: Run /init as init process Mar 2 13:26:01.251490 kernel: with arguments: Mar 2 13:26:01.251503 kernel: /init Mar 2 13:26:01.251515 kernel: with environment: Mar 2 13:26:01.251538 kernel: HOME=/ Mar 2 13:26:01.251551 kernel: TERM=linux Mar 2 13:26:01.251563 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 1255450629 wd_nsec: 1255450461 Mar 2 13:26:01.251574 kernel: SCSI subsystem initialized Mar 2 13:26:01.251586 kernel: libata version 3.00 loaded. Mar 2 13:26:01.251600 kernel: hrtimer: interrupt took 27307622 ns Mar 2 13:26:01.252128 kernel: ahci 0000:00:1f.2: version 3.0 Mar 2 13:26:01.252159 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 2 13:26:01.252556 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Mar 2 13:26:01.252949 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Mar 2 13:26:01.261613 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 2 13:26:01.262107 kernel: scsi host0: ahci Mar 2 13:26:01.262625 kernel: scsi host1: ahci Mar 2 13:26:01.263076 kernel: scsi host2: ahci Mar 2 13:26:01.263485 kernel: scsi host3: ahci Mar 2 13:26:01.263923 kernel: scsi host4: ahci Mar 2 13:26:01.274513 kernel: scsi host5: ahci Mar 2 13:26:01.274559 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 26 lpm-pol 1 Mar 2 13:26:01.274583 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 26 lpm-pol 1 Mar 2 13:26:01.274596 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 26 lpm-pol 1 Mar 2 13:26:01.274608 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 26 lpm-pol 1 Mar 2 13:26:01.274621 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 26 lpm-pol 1 Mar 2 13:26:01.274634 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 26 lpm-pol 1 Mar 2 13:26:01.274648 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 2 13:26:01.274661 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 2 13:26:01.274678 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 2 13:26:01.274693 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 2 13:26:01.274705 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 2 13:26:01.274720 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 2 13:26:01.274733 kernel: ata3.00: LPM support broken, forcing max_power Mar 2 13:26:01.274747 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 2 13:26:01.274867 kernel: ata3.00: applying bridge limits Mar 2 13:26:01.274891 kernel: ata3.00: LPM support broken, forcing max_power Mar 2 13:26:01.274907 kernel: ata3.00: configured for UDMA/100 Mar 2 13:26:01.275433 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 2 13:26:01.275881 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Mar 2 13:26:01.276176 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 2 13:26:01.276196 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 2 13:26:01.276582 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Mar 2 13:26:01.277003 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Mar 2 13:26:01.277080 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 2 13:26:01.277096 kernel: GPT:16515071 != 27000831 Mar 2 13:26:01.277109 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 2 13:26:01.277123 kernel: GPT:16515071 != 27000831 Mar 2 13:26:01.277137 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 2 13:26:01.277158 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 13:26:01.277174 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 2 13:26:01.277188 kernel: device-mapper: uevent: version 1.0.3 Mar 2 13:26:01.277201 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Mar 2 13:26:01.277216 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Mar 2 13:26:01.285513 kernel: raid6: avx2x4 gen() 8831 MB/s Mar 2 13:26:01.285536 kernel: raid6: avx2x2 gen() 10565 MB/s Mar 2 13:26:01.285557 kernel: raid6: avx2x1 gen() 5638 MB/s Mar 2 13:26:01.285570 kernel: raid6: using algorithm avx2x2 gen() 10565 MB/s Mar 2 13:26:01.285581 kernel: raid6: .... xor() 7272 MB/s, rmw enabled Mar 2 13:26:01.285593 kernel: raid6: using avx2x2 recovery algorithm Mar 2 13:26:01.285605 kernel: xor: automatically using best checksumming function avx Mar 2 13:26:01.285617 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 2 13:26:01.285633 kernel: BTRFS: device fsid 9f58c137-afcb-4dd2-af86-bad6ccd312c3 devid 1 transid 37 /dev/mapper/usr (253:0) scanned by mount (182) Mar 2 13:26:01.285645 kernel: BTRFS info (device dm-0): first mount of filesystem 9f58c137-afcb-4dd2-af86-bad6ccd312c3 Mar 2 13:26:01.285657 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:26:01.285669 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Mar 2 13:26:01.285681 kernel: BTRFS info (device dm-0 state E): enabling free space tree Mar 2 13:26:01.285696 kernel: loop: module loaded Mar 2 13:26:01.285707 kernel: loop0: detected capacity change from 0 to 100544 Mar 2 13:26:01.285719 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 2 13:26:01.285733 systemd[1]: Successfully made /usr/ read-only. Mar 2 13:26:01.285748 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 13:26:01.285859 systemd[1]: Detected virtualization kvm. Mar 2 13:26:01.285882 systemd[1]: Detected architecture x86-64. Mar 2 13:26:01.285896 systemd[1]: Running in initrd. Mar 2 13:26:01.285911 systemd[1]: No hostname configured, using default hostname. Mar 2 13:26:01.285927 systemd[1]: Hostname set to . Mar 2 13:26:01.285942 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Mar 2 13:26:01.285956 systemd[1]: Queued start job for default target initrd.target. Mar 2 13:26:01.285970 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Mar 2 13:26:01.285993 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 13:26:01.286006 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 13:26:01.286023 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 2 13:26:01.286039 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 13:26:01.286057 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 2 13:26:01.286077 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 2 13:26:01.286092 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 13:26:01.286107 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 13:26:01.286121 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Mar 2 13:26:01.286136 systemd[1]: Reached target paths.target - Path Units. Mar 2 13:26:01.286150 systemd[1]: Reached target slices.target - Slice Units. Mar 2 13:26:01.286164 systemd[1]: Reached target swap.target - Swaps. Mar 2 13:26:01.286183 systemd[1]: Reached target timers.target - Timer Units. Mar 2 13:26:01.286199 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 13:26:01.286212 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 13:26:01.286286 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Mar 2 13:26:01.286301 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 2 13:26:01.286318 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 2 13:26:01.286337 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 13:26:01.286421 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 13:26:01.286436 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 13:26:01.286451 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 13:26:01.286466 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 2 13:26:01.286480 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 2 13:26:01.286493 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 13:26:01.286563 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 2 13:26:01.286577 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Mar 2 13:26:01.286590 systemd[1]: Starting systemd-fsck-usr.service... Mar 2 13:26:01.286603 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 13:26:01.286616 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 13:26:01.286672 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:26:01.286685 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 2 13:26:01.286698 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 13:26:01.286863 systemd-journald[319]: Collecting audit messages is enabled. Mar 2 13:26:01.286965 systemd[1]: Finished systemd-fsck-usr.service. Mar 2 13:26:01.286984 kernel: audit: type=1130 audit(1772457961.114:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:01.286998 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 13:26:01.288349 systemd-journald[319]: Journal started Mar 2 13:26:01.288429 systemd-journald[319]: Runtime Journal (/run/log/journal/c1ba91920c0241e8afa7e0d69a108bd3) is 6M, max 48.2M, 42.1M free. Mar 2 13:26:01.114000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:01.346865 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 13:26:01.346000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:01.408881 kernel: audit: type=1130 audit(1772457961.346:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:01.422559 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 13:26:01.554077 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 13:26:01.551000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:01.567000 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 13:26:02.416402 kernel: audit: type=1130 audit(1772457961.551:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.416454 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 2 13:26:02.416473 kernel: Bridge firewalling registered Mar 2 13:26:01.704116 systemd-tmpfiles[333]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Mar 2 13:26:01.758718 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 13:26:02.452000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.146650 systemd-modules-load[322]: Inserted module 'br_netfilter' Mar 2 13:26:02.511395 kernel: audit: type=1130 audit(1772457962.452:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.611632 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 13:26:02.670709 kernel: audit: type=1130 audit(1772457962.621:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.621000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.654436 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:26:02.722896 kernel: audit: type=1130 audit(1772457962.682:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.682000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.726210 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 13:26:02.844298 kernel: audit: type=1130 audit(1772457962.752:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.752000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:02.773878 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 13:26:02.896500 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 13:26:02.995568 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 13:26:03.024000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:03.050685 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 2 13:26:03.102606 kernel: audit: type=1130 audit(1772457963.024:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:03.126100 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:26:03.196912 kernel: audit: type=1130 audit(1772457963.149:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:03.149000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:03.223007 kernel: audit: type=1334 audit(1772457963.202:11): prog-id=6 op=LOAD Mar 2 13:26:03.202000 audit: BPF prog-id=6 op=LOAD Mar 2 13:26:03.211922 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 13:26:03.301696 dracut-cmdline[356]: dracut-109 Mar 2 13:26:03.318743 dracut-cmdline[356]: Using kernel command line parameters: SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ecd171dfe44947e01ea4ccf8f02472d94712b039722a1ae86ca9989f56ff7a41 Mar 2 13:26:03.678301 systemd-resolved[358]: Positive Trust Anchors: Mar 2 13:26:03.678326 systemd-resolved[358]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 13:26:03.678333 systemd-resolved[358]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Mar 2 13:26:03.678373 systemd-resolved[358]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 13:26:03.890092 systemd-resolved[358]: Defaulting to hostname 'linux'. Mar 2 13:26:03.911440 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 13:26:03.983713 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 13:26:03.983000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:04.332221 kernel: Loading iSCSI transport class v2.0-870. Mar 2 13:26:04.412747 kernel: iscsi: registered transport (tcp) Mar 2 13:26:04.500604 kernel: iscsi: registered transport (qla4xxx) Mar 2 13:26:04.501159 kernel: QLogic iSCSI HBA Driver Mar 2 13:26:04.750934 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 13:26:04.951686 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 13:26:04.988000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:05.004314 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 13:26:06.124933 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 2 13:26:06.222589 kernel: kauditd_printk_skb: 2 callbacks suppressed Mar 2 13:26:06.222635 kernel: audit: type=1130 audit(1772457966.156:14): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:06.156000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:06.226160 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 2 13:26:06.307489 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 2 13:26:06.892436 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 2 13:26:07.119887 kernel: audit: type=1130 audit(1772457966.950:15): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:07.120093 kernel: audit: type=1334 audit(1772457966.986:16): prog-id=7 op=LOAD Mar 2 13:26:07.120118 kernel: audit: type=1334 audit(1772457966.986:17): prog-id=8 op=LOAD Mar 2 13:26:06.950000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:06.986000 audit: BPF prog-id=7 op=LOAD Mar 2 13:26:06.986000 audit: BPF prog-id=8 op=LOAD Mar 2 13:26:07.000406 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 13:26:07.426697 systemd-udevd[573]: Using default interface naming scheme 'v257'. Mar 2 13:26:07.895059 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 13:26:08.046237 kernel: audit: type=1130 audit(1772457967.939:18): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:07.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:08.104891 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 2 13:26:08.501091 dracut-pre-trigger[630]: rd.md=0: removing MD RAID activation Mar 2 13:26:09.226517 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 13:26:09.539218 kernel: audit: type=1130 audit(1772457969.286:19): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:09.286000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:09.308195 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 13:26:09.597473 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 13:26:09.828618 kernel: audit: type=1130 audit(1772457969.675:20): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:09.675000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:09.842000 audit: BPF prog-id=9 op=LOAD Mar 2 13:26:09.885443 kernel: audit: type=1334 audit(1772457969.842:21): prog-id=9 op=LOAD Mar 2 13:26:09.895922 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 13:26:10.477083 systemd-networkd[727]: lo: Link UP Mar 2 13:26:10.477158 systemd-networkd[727]: lo: Gained carrier Mar 2 13:26:10.540000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:10.491598 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 13:26:10.659900 kernel: audit: type=1130 audit(1772457970.540:22): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:10.542706 systemd[1]: Reached target network.target - Network. Mar 2 13:26:10.819686 kernel: audit: type=1130 audit(1772457970.695:23): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:10.695000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:10.666614 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 13:26:10.744220 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 2 13:26:12.072970 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 2 13:26:12.544121 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 2 13:26:12.623241 kernel: cryptd: max_cpu_qlen set to 1000 Mar 2 13:26:12.815090 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 2 13:26:12.921972 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 2 13:26:13.022143 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 2 13:26:13.082704 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 13:26:13.218454 kernel: audit: type=1131 audit(1772457973.104:24): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:13.104000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:13.083042 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:26:13.115177 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:26:13.200033 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:26:13.506175 disk-uuid[770]: Primary Header is updated. Mar 2 13:26:13.506175 disk-uuid[770]: Secondary Entries is updated. Mar 2 13:26:13.506175 disk-uuid[770]: Secondary Header is updated. Mar 2 13:26:14.023572 kernel: AES CTR mode by8 optimization enabled Mar 2 13:26:14.125435 systemd-networkd[727]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Mar 2 13:26:16.014040 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Mar 2 13:26:14.125452 systemd-networkd[727]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 13:26:16.027711 disk-uuid[772]: Warning: The kernel is still using the old partition table. Mar 2 13:26:16.027711 disk-uuid[772]: The new table will be used at the next reboot or after you Mar 2 13:26:16.027711 disk-uuid[772]: run partprobe(8) or kpartx(8) Mar 2 13:26:16.027711 disk-uuid[772]: The operation has completed successfully. Mar 2 13:26:14.137210 systemd-networkd[727]: eth0: Link UP Mar 2 13:26:14.144036 systemd-networkd[727]: eth0: Gained carrier Mar 2 13:26:14.144057 systemd-networkd[727]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Mar 2 13:26:14.237966 systemd-networkd[727]: eth0: DHCPv4 address 10.0.0.31/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 2 13:26:15.414592 systemd-networkd[727]: eth0: Gained IPv6LL Mar 2 13:26:16.046000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:15.834463 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 2 13:26:16.407580 kernel: audit: type=1130 audit(1772457976.046:25): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:16.407899 kernel: audit: type=1131 audit(1772457976.082:26): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:16.082000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:15.834693 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 2 13:26:16.093049 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 2 13:26:16.532271 kernel: audit: type=1130 audit(1772457976.482:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:16.482000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:16.533592 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:26:16.690516 kernel: audit: type=1130 audit(1772457976.593:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:16.593000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:16.721200 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 13:26:16.750982 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 13:26:16.801352 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 13:26:16.893632 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 2 13:26:17.023363 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 2 13:26:17.517874 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 2 13:26:17.605000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:17.671046 kernel: audit: type=1130 audit(1772457977.605:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:17.729981 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (858) Mar 2 13:26:17.765626 kernel: BTRFS info (device vda6): first mount of filesystem fc057d22-d01e-405b-8e65-66e5821ef6dd Mar 2 13:26:17.781641 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:26:17.922458 kernel: BTRFS info (device vda6): turning on async discard Mar 2 13:26:17.923114 kernel: BTRFS info (device vda6): enabling free space tree Mar 2 13:26:18.046159 kernel: BTRFS info (device vda6): last unmount of filesystem fc057d22-d01e-405b-8e65-66e5821ef6dd Mar 2 13:26:18.102584 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 2 13:26:18.114000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:18.129921 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 2 13:26:18.228499 kernel: audit: type=1130 audit(1772457978.114:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:23.980515 ignition[879]: Ignition 2.24.0 Mar 2 13:26:24.001946 ignition[879]: Stage: fetch-offline Mar 2 13:26:24.347193 ignition[879]: no configs at "/usr/lib/ignition/base.d" Mar 2 13:26:24.387506 ignition[879]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:26:24.389703 ignition[879]: parsed url from cmdline: "" Mar 2 13:26:24.389714 ignition[879]: no config URL provided Mar 2 13:26:24.399491 ignition[879]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 13:26:24.399554 ignition[879]: no config at "/usr/lib/ignition/user.ign" Mar 2 13:26:24.400193 ignition[879]: op(1): [started] loading QEMU firmware config module Mar 2 13:26:24.400210 ignition[879]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 2 13:26:24.660122 ignition[879]: op(1): [finished] loading QEMU firmware config module Mar 2 13:26:24.660177 ignition[879]: QEMU firmware config was not found. Ignoring... Mar 2 13:26:25.732509 ignition[879]: parsing config with SHA512: 2e47d7477739e2add8cc887a3fb04becba9f56dc85560aba63a830a7384b1073eb54c7093e4364990e8744d3d497f9f8ee46ba6de427e8c475ab9523c6cdab09 Mar 2 13:26:25.902420 unknown[879]: fetched base config from "system" Mar 2 13:26:25.925750 unknown[879]: fetched user config from "qemu" Mar 2 13:26:26.007515 ignition[879]: fetch-offline: fetch-offline passed Mar 2 13:26:26.008080 ignition[879]: Ignition finished successfully Mar 2 13:26:26.125097 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 13:26:26.288204 kernel: audit: type=1130 audit(1772457986.136:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:26.136000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:26.188719 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 2 13:26:26.205054 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 2 13:26:27.615466 ignition[891]: Ignition 2.24.0 Mar 2 13:26:27.618721 ignition[891]: Stage: kargs Mar 2 13:26:27.619150 ignition[891]: no configs at "/usr/lib/ignition/base.d" Mar 2 13:26:27.619171 ignition[891]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:26:27.625041 ignition[891]: kargs: kargs passed Mar 2 13:26:27.625137 ignition[891]: Ignition finished successfully Mar 2 13:26:27.709070 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 2 13:26:27.865000 kernel: audit: type=1130 audit(1772457987.771:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:27.771000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:27.820020 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 2 13:26:28.457441 ignition[899]: Ignition 2.24.0 Mar 2 13:26:28.457577 ignition[899]: Stage: disks Mar 2 13:26:28.459651 ignition[899]: no configs at "/usr/lib/ignition/base.d" Mar 2 13:26:28.459667 ignition[899]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:26:28.503568 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 2 13:26:28.614654 kernel: audit: type=1130 audit(1772457988.524:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:28.524000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:28.472299 ignition[899]: disks: disks passed Mar 2 13:26:28.525891 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 2 13:26:28.472472 ignition[899]: Ignition finished successfully Mar 2 13:26:28.529403 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 2 13:26:28.633072 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 13:26:28.734952 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 13:26:28.765900 systemd[1]: Reached target basic.target - Basic System. Mar 2 13:26:28.807019 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 2 13:26:29.215112 systemd-fsck[909]: ROOT: clean, 15/456736 files, 38230/456704 blocks Mar 2 13:26:29.257420 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 2 13:26:29.324000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:29.337749 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 2 13:26:29.454743 kernel: audit: type=1130 audit(1772457989.324:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:31.422525 kernel: EXT4-fs (vda9): mounted filesystem 6291e4f9-48b3-49a4-8304-70ba6c554986 r/w with ordered data mode. Quota mode: none. Mar 2 13:26:31.430150 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 2 13:26:31.498953 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 2 13:26:31.601513 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 13:26:31.674102 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 2 13:26:31.781033 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 2 13:26:31.822958 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (918) Mar 2 13:26:31.781196 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 2 13:26:31.971696 kernel: BTRFS info (device vda6): first mount of filesystem fc057d22-d01e-405b-8e65-66e5821ef6dd Mar 2 13:26:31.971936 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:26:31.781252 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 13:26:31.894243 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 2 13:26:32.125646 kernel: BTRFS info (device vda6): turning on async discard Mar 2 13:26:32.125688 kernel: BTRFS info (device vda6): enabling free space tree Mar 2 13:26:32.053990 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 2 13:26:32.166279 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 13:26:38.754757 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 2 13:26:38.834000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:38.988644 kernel: audit: type=1130 audit(1772457998.834:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:38.990944 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 2 13:26:39.216108 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 2 13:26:39.619075 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 2 13:26:39.681298 kernel: BTRFS info (device vda6): last unmount of filesystem fc057d22-d01e-405b-8e65-66e5821ef6dd Mar 2 13:26:40.167623 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 2 13:26:40.268000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:40.384576 kernel: audit: type=1130 audit(1772458000.268:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:42.087064 ignition[1016]: INFO : Ignition 2.24.0 Mar 2 13:26:42.087064 ignition[1016]: INFO : Stage: mount Mar 2 13:26:42.087064 ignition[1016]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 13:26:42.087064 ignition[1016]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:26:42.627740 kernel: audit: type=1130 audit(1772458002.276:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:42.276000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:42.638018 ignition[1016]: INFO : mount: mount passed Mar 2 13:26:42.638018 ignition[1016]: INFO : Ignition finished successfully Mar 2 13:26:42.153145 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 2 13:26:42.633260 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 2 13:26:43.147359 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 13:26:43.485358 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1029) Mar 2 13:26:43.541195 kernel: BTRFS info (device vda6): first mount of filesystem fc057d22-d01e-405b-8e65-66e5821ef6dd Mar 2 13:26:43.541928 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:26:43.758171 kernel: BTRFS info (device vda6): turning on async discard Mar 2 13:26:43.772342 kernel: BTRFS info (device vda6): enabling free space tree Mar 2 13:26:43.847348 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 13:26:45.854968 ignition[1046]: INFO : Ignition 2.24.0 Mar 2 13:26:45.854968 ignition[1046]: INFO : Stage: files Mar 2 13:26:45.953044 ignition[1046]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 13:26:45.953044 ignition[1046]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:26:46.068329 ignition[1046]: DEBUG : files: compiled without relabeling support, skipping Mar 2 13:26:46.068329 ignition[1046]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 2 13:26:46.068329 ignition[1046]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 2 13:26:46.270747 ignition[1046]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 2 13:26:46.270747 ignition[1046]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 2 13:26:46.270747 ignition[1046]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 2 13:26:46.270747 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 2 13:26:46.270747 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 2 13:26:46.152620 unknown[1046]: wrote ssh authorized keys file for user: core Mar 2 13:26:47.295287 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 2 13:26:53.002358 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 2 13:26:53.123157 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 13:26:53.208560 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 13:26:53.769719 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 13:26:53.769719 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 13:26:53.769719 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 13:26:53.769719 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 13:26:53.769719 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 13:26:53.769719 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.4-x86-64.raw: attempt #1 Mar 2 13:26:54.970897 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 2 13:26:57.966928 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-x86-64.raw" Mar 2 13:26:57.966928 ignition[1046]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Mar 2 13:26:58.064092 ignition[1046]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Mar 2 13:26:58.857714 ignition[1046]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 2 13:26:59.059721 ignition[1046]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 2 13:26:59.059721 ignition[1046]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Mar 2 13:26:59.059721 ignition[1046]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Mar 2 13:26:59.059721 ignition[1046]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Mar 2 13:26:59.466200 kernel: audit: type=1130 audit(1772458019.305:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:59.305000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:26:59.466656 ignition[1046]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 2 13:26:59.466656 ignition[1046]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 2 13:26:59.466656 ignition[1046]: INFO : files: files passed Mar 2 13:26:59.466656 ignition[1046]: INFO : Ignition finished successfully Mar 2 13:26:59.197400 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 2 13:26:59.344174 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 2 13:26:59.629223 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 2 13:27:00.018294 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 2 13:27:00.239305 kernel: audit: type=1130 audit(1772458020.062:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:00.239365 kernel: audit: type=1131 audit(1772458020.062:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:00.062000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:00.062000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:00.242956 initrd-setup-root-after-ignition[1077]: grep: /sysroot/oem/oem-release: No such file or directory Mar 2 13:27:00.381974 kernel: audit: type=1130 audit(1772458020.283:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:00.283000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:00.018714 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 2 13:27:00.408304 initrd-setup-root-after-ignition[1080]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 13:27:00.408304 initrd-setup-root-after-ignition[1080]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 2 13:27:00.151702 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 13:27:00.639549 initrd-setup-root-after-ignition[1084]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 13:27:00.294194 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 2 13:27:00.474169 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 2 13:27:01.985391 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 2 13:27:02.434089 kernel: audit: type=1130 audit(1772458022.016:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:02.434172 kernel: audit: type=1131 audit(1772458022.016:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:02.016000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:02.016000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:01.999908 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 2 13:27:02.025703 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 2 13:27:02.531564 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 2 13:27:02.807411 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 2 13:27:02.825200 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 2 13:27:03.516024 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 13:27:03.786524 kernel: audit: type=1130 audit(1772458023.598:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:03.598000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:03.632349 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 2 13:27:04.389404 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Mar 2 13:27:04.408166 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 2 13:27:04.509234 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 13:27:04.522157 systemd[1]: Stopped target timers.target - Timer Units. Mar 2 13:27:04.582175 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 2 13:27:04.593283 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 13:27:05.046413 kernel: audit: type=1131 audit(1772458024.899:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:04.899000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:05.055079 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 2 13:27:05.127298 systemd[1]: Stopped target basic.target - Basic System. Mar 2 13:27:05.266337 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 2 13:27:05.425036 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 13:27:05.545929 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 2 13:27:05.819218 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Mar 2 13:27:05.913524 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 2 13:27:06.252419 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 13:27:06.486702 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 2 13:27:06.625759 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 2 13:27:06.681345 systemd[1]: Stopped target swap.target - Swaps. Mar 2 13:27:07.264138 kernel: audit: type=1131 audit(1772458026.944:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:06.944000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:06.685005 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 2 13:27:06.693748 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 2 13:27:07.651376 kernel: audit: type=1131 audit(1772458027.508:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:07.508000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:07.115088 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 2 13:27:07.133186 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 13:27:07.399665 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 2 13:27:07.444000 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 13:27:07.483025 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 2 13:27:07.491118 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 2 13:27:07.696437 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 2 13:27:08.102353 kernel: audit: type=1131 audit(1772458027.813:48): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:07.813000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:07.697347 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 13:27:07.820300 systemd[1]: Stopped target paths.target - Path Units. Mar 2 13:27:07.984348 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 2 13:27:08.003726 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 13:27:08.200128 systemd[1]: Stopped target slices.target - Slice Units. Mar 2 13:27:08.351696 systemd[1]: Stopped target sockets.target - Socket Units. Mar 2 13:27:08.441740 systemd[1]: iscsid.socket: Deactivated successfully. Mar 2 13:27:08.442328 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 13:27:08.607399 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 2 13:27:08.613114 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 13:27:08.702342 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Mar 2 13:27:08.705179 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Mar 2 13:27:09.124759 kernel: audit: type=1131 audit(1772458028.976:49): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:08.976000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:08.793275 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 2 13:27:09.338118 kernel: audit: type=1131 audit(1772458029.179:50): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:09.179000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:08.799191 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 13:27:09.005178 systemd[1]: ignition-files.service: Deactivated successfully. Mar 2 13:27:09.010133 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 2 13:27:09.197416 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 2 13:27:09.621646 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 2 13:27:09.846040 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 2 13:27:09.883679 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 13:27:10.145028 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 2 13:27:10.321096 kernel: audit: type=1131 audit(1772458030.143:51): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.143000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.148139 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 13:27:10.698277 kernel: audit: type=1131 audit(1772458030.410:52): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.410000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.698550 ignition[1104]: INFO : Ignition 2.24.0 Mar 2 13:27:10.698550 ignition[1104]: INFO : Stage: umount Mar 2 13:27:10.853443 kernel: audit: type=1131 audit(1772458030.710:53): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.710000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.414288 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 2 13:27:10.988109 ignition[1104]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 13:27:10.988109 ignition[1104]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:27:11.427079 kernel: audit: type=1131 audit(1772458031.127:54): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.127000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.454008 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 13:27:11.781733 kernel: audit: type=1130 audit(1772458031.531:55): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.784208 kernel: audit: type=1131 audit(1772458031.534:56): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.531000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.534000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.794206 ignition[1104]: INFO : umount: umount passed Mar 2 13:27:11.794206 ignition[1104]: INFO : Ignition finished successfully Mar 2 13:27:12.419253 kernel: audit: type=1131 audit(1772458031.819:57): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:12.421986 kernel: audit: type=1131 audit(1772458032.221:58): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.819000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:12.221000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:10.922756 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 2 13:27:12.550047 kernel: audit: type=1131 audit(1772458032.423:59): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:12.423000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.003663 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 2 13:27:11.004272 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 2 13:27:11.142932 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 2 13:27:11.143299 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 2 13:27:13.012055 kernel: audit: type=1131 audit(1772458032.845:60): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:12.845000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.548455 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 2 13:27:13.101000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.554591 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 2 13:27:11.947418 systemd[1]: Stopped target network.target - Network. Mar 2 13:27:13.211000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:11.954312 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 2 13:27:12.020192 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 2 13:27:12.223415 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 2 13:27:12.230185 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 2 13:27:13.895000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:12.425341 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 2 13:27:12.433118 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 2 13:27:12.846585 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 2 13:27:12.847077 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 2 13:27:14.400000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:13.104301 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 2 13:27:14.516000 audit: BPF prog-id=6 op=UNLOAD Mar 2 13:27:14.519000 audit: BPF prog-id=9 op=UNLOAD Mar 2 13:27:13.111136 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 2 13:27:13.226402 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 2 13:27:13.394899 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 2 13:27:13.775412 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 2 13:27:13.781935 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 2 13:27:14.119928 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 2 13:27:14.120167 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 2 13:27:14.520258 systemd[1]: Stopped target network-pre.target - Preparation for Network. Mar 2 13:27:14.633173 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 2 13:27:14.633285 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 2 13:27:14.719613 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 2 13:27:15.107281 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 2 13:27:15.112055 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 13:27:15.287697 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 13:27:15.555454 kernel: kauditd_printk_skb: 6 callbacks suppressed Mar 2 13:27:15.555692 kernel: audit: type=1131 audit(1772458035.282:67): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:15.282000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:15.299100 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:27:15.781000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:15.804143 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 2 13:27:16.061353 kernel: audit: type=1131 audit(1772458035.781:68): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:15.804377 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 2 13:27:16.414706 kernel: audit: type=1131 audit(1772458036.066:69): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:16.066000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:16.125414 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 13:27:16.683242 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 2 13:27:16.685184 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 13:27:16.989000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:17.103959 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 2 13:27:17.154128 kernel: audit: type=1131 audit(1772458036.989:70): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:17.104139 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 2 13:27:17.351226 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 2 13:27:17.351481 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 13:27:17.600709 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 2 13:27:17.608354 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 2 13:27:17.893000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:18.001972 kernel: audit: type=1131 audit(1772458037.893:71): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:18.035304 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 2 13:27:18.046353 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 2 13:27:18.411746 kernel: audit: type=1131 audit(1772458038.236:72): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:18.236000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:18.433289 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 2 13:27:18.455200 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 13:27:18.902002 kernel: audit: type=1131 audit(1772458038.682:73): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:18.682000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:18.929463 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 2 13:27:19.197100 kernel: audit: type=1131 audit(1772458039.033:74): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.197177 kernel: audit: type=1131 audit(1772458039.034:75): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.033000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.034000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.027680 systemd[1]: systemd-network-generator.service: Deactivated successfully. Mar 2 13:27:19.611221 kernel: audit: type=1131 audit(1772458039.300:76): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.300000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.028332 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 13:27:19.734000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.894000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.034893 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 2 13:27:19.035014 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 13:27:19.035209 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Mar 2 13:27:20.136000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.046064 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 13:27:19.308382 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 2 13:27:19.308665 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 13:27:20.450565 kernel: kauditd_printk_skb: 3 callbacks suppressed Mar 2 13:27:20.450615 kernel: audit: type=1130 audit(1772458040.299:80): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:20.450635 kernel: audit: type=1131 audit(1772458040.306:81): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:20.299000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:20.306000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:19.736359 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 13:27:19.736620 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:27:20.016179 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 2 13:27:20.022312 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 2 13:27:20.242648 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 2 13:27:20.243126 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 2 13:27:20.307332 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 2 13:27:20.599614 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 2 13:27:21.756177 systemd[1]: Switching root. Mar 2 13:27:24.756299 systemd-journald[319]: Received SIGTERM from PID 1 (systemd). Mar 2 13:27:24.795736 systemd-journald[319]: Journal stopped Mar 2 13:27:42.535939 kernel: audit: type=1335 audit(1772458044.806:82): pid=319 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=kernel comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" nl-mcgrp=1 op=disconnect res=1 Mar 2 13:27:42.540731 kernel: SELinux: policy capability network_peer_controls=1 Mar 2 13:27:42.541215 kernel: SELinux: policy capability open_perms=1 Mar 2 13:27:42.541312 kernel: SELinux: policy capability extended_socket_class=1 Mar 2 13:27:42.541461 kernel: SELinux: policy capability always_check_network=0 Mar 2 13:27:42.541737 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 2 13:27:42.542000 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 2 13:27:42.542141 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 2 13:27:42.542162 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 2 13:27:42.542298 kernel: SELinux: policy capability userspace_initial_context=0 Mar 2 13:27:42.542429 kernel: audit: type=1403 audit(1772458049.039:83): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 2 13:27:42.542453 systemd[1]: Successfully loaded SELinux policy in 882.975ms. Mar 2 13:27:42.545668 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 36.952ms. Mar 2 13:27:42.545699 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 13:27:42.545719 systemd[1]: Detected virtualization kvm. Mar 2 13:27:42.545941 systemd[1]: Detected architecture x86-64. Mar 2 13:27:42.546039 systemd[1]: Detected first boot. Mar 2 13:27:42.546125 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Mar 2 13:27:42.546146 kernel: audit: type=1334 audit(1772458049.630:84): prog-id=10 op=LOAD Mar 2 13:27:42.546235 kernel: audit: type=1334 audit(1772458049.635:85): prog-id=10 op=UNLOAD Mar 2 13:27:42.546325 kernel: audit: type=1334 audit(1772458049.635:86): prog-id=11 op=LOAD Mar 2 13:27:42.546415 kernel: audit: type=1334 audit(1772458049.635:87): prog-id=11 op=UNLOAD Mar 2 13:27:42.546438 zram_generator::config[1148]: No configuration found. Mar 2 13:27:42.546534 kernel: Guest personality initialized and is inactive Mar 2 13:27:42.546554 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Mar 2 13:27:42.550724 kernel: Initialized host personality Mar 2 13:27:42.550982 kernel: NET: Registered PF_VSOCK protocol family Mar 2 13:27:42.551063 systemd[1]: Populated /etc with preset unit settings. Mar 2 13:27:42.551199 kernel: audit: type=1334 audit(1772458055.796:88): prog-id=12 op=LOAD Mar 2 13:27:42.551222 kernel: audit: type=1334 audit(1772458055.799:89): prog-id=3 op=UNLOAD Mar 2 13:27:42.551246 kernel: audit: type=1334 audit(1772458055.799:90): prog-id=13 op=LOAD Mar 2 13:27:42.551269 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 2 13:27:42.551289 kernel: audit: type=1334 audit(1772458055.799:91): prog-id=14 op=LOAD Mar 2 13:27:42.551431 kernel: audit: type=1334 audit(1772458055.799:92): prog-id=4 op=UNLOAD Mar 2 13:27:42.551517 kernel: audit: type=1334 audit(1772458055.799:93): prog-id=5 op=UNLOAD Mar 2 13:27:42.551541 kernel: audit: type=1131 audit(1772458055.811:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:42.553742 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 2 13:27:42.553907 kernel: audit: type=1130 audit(1772458055.923:95): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:42.554009 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 2 13:27:42.554033 kernel: audit: type=1131 audit(1772458055.923:96): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:42.554122 kernel: audit: type=1334 audit(1772458056.019:97): prog-id=12 op=UNLOAD Mar 2 13:27:42.554228 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 2 13:27:42.554322 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 2 13:27:42.554345 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 2 13:27:42.554363 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 2 13:27:42.554452 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 2 13:27:42.554483 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 2 13:27:42.554502 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 2 13:27:42.554529 systemd[1]: Created slice user.slice - User and Session Slice. Mar 2 13:27:42.554552 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 13:27:42.554652 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 13:27:42.554676 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 2 13:27:42.554941 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 2 13:27:42.554976 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 2 13:27:42.554999 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 13:27:42.555093 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 2 13:27:42.556745 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 13:27:42.556909 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 13:27:42.556932 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 2 13:27:42.556953 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 2 13:27:42.556975 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 2 13:27:42.557125 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 2 13:27:42.557218 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 13:27:42.557240 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 13:27:42.557262 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Mar 2 13:27:42.557282 systemd[1]: Reached target slices.target - Slice Units. Mar 2 13:27:42.557299 systemd[1]: Reached target swap.target - Swaps. Mar 2 13:27:42.557319 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 2 13:27:42.557339 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 2 13:27:42.557356 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 2 13:27:42.557451 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Mar 2 13:27:42.557471 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Mar 2 13:27:42.557489 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 13:27:42.557510 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Mar 2 13:27:42.557531 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Mar 2 13:27:42.557549 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 13:27:42.559339 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 13:27:42.559379 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 2 13:27:42.559407 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 2 13:27:42.563287 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 2 13:27:42.563312 systemd[1]: Mounting media.mount - External Media Directory... Mar 2 13:27:42.563335 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:27:42.563352 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 2 13:27:42.563368 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 2 13:27:42.563385 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 2 13:27:42.563402 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 2 13:27:42.576906 systemd[1]: Reached target machines.target - Containers. Mar 2 13:27:42.576941 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 2 13:27:42.577031 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 13:27:42.577052 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 13:27:42.577072 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 2 13:27:42.577200 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 13:27:42.577221 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 13:27:42.577241 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 13:27:42.577268 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 2 13:27:42.577351 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 13:27:42.577433 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 2 13:27:42.577454 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 2 13:27:42.577473 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 2 13:27:42.577492 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 2 13:27:42.577513 kernel: kauditd_printk_skb: 1 callbacks suppressed Mar 2 13:27:42.577541 kernel: audit: type=1131 audit(1772458060.997:99): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:42.580913 systemd[1]: Stopped systemd-fsck-usr.service. Mar 2 13:27:42.580956 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 13:27:42.580979 kernel: audit: type=1131 audit(1772458061.164:100): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:42.581006 kernel: audit: type=1334 audit(1772458061.221:101): prog-id=14 op=UNLOAD Mar 2 13:27:42.581029 kernel: audit: type=1334 audit(1772458061.221:102): prog-id=13 op=UNLOAD Mar 2 13:27:42.581046 kernel: audit: type=1334 audit(1772458061.284:103): prog-id=15 op=LOAD Mar 2 13:27:42.581254 kernel: audit: type=1334 audit(1772458061.349:104): prog-id=16 op=LOAD Mar 2 13:27:42.581276 kernel: audit: type=1334 audit(1772458061.409:105): prog-id=17 op=LOAD Mar 2 13:27:42.581296 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 13:27:42.581324 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 13:27:42.581346 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 13:27:42.581366 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 2 13:27:42.581384 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 2 13:27:42.581407 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 13:27:42.581480 systemd-journald[1236]: Collecting audit messages is enabled. Mar 2 13:27:42.581722 kernel: fuse: init (API version 7.41) Mar 2 13:27:42.581751 kernel: audit: type=1305 audit(1772458062.508:106): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Mar 2 13:27:42.581903 systemd-journald[1236]: Journal started Mar 2 13:27:42.581945 systemd-journald[1236]: Runtime Journal (/run/log/journal/c1ba91920c0241e8afa7e0d69a108bd3) is 6M, max 48.2M, 42.1M free. Mar 2 13:27:38.053000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Mar 2 13:27:40.997000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:41.164000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:41.221000 audit: BPF prog-id=14 op=UNLOAD Mar 2 13:27:41.221000 audit: BPF prog-id=13 op=UNLOAD Mar 2 13:27:41.284000 audit: BPF prog-id=15 op=LOAD Mar 2 13:27:41.349000 audit: BPF prog-id=16 op=LOAD Mar 2 13:27:41.409000 audit: BPF prog-id=17 op=LOAD Mar 2 13:27:42.508000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Mar 2 13:27:35.700892 systemd[1]: Queued start job for default target multi-user.target. Mar 2 13:27:35.801642 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 2 13:27:35.811364 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 2 13:27:35.812610 systemd[1]: systemd-journald.service: Consumed 5.557s CPU time. Mar 2 13:27:42.680742 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:27:42.681184 kernel: audit: type=1300 audit(1772458062.508:106): arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7ffd14b16b70 a2=4000 a3=0 items=0 ppid=1 pid=1236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 2 13:27:42.681233 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 13:27:42.508000 audit[1236]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7ffd14b16b70 a2=4000 a3=0 items=0 ppid=1 pid=1236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 2 13:27:42.508000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Mar 2 13:27:42.684748 kernel: audit: type=1327 audit(1772458062.508:106): proctitle="/usr/lib/systemd/systemd-journald" Mar 2 13:27:42.774000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:42.790120 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 2 13:27:42.915080 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 2 13:27:42.986419 systemd[1]: Mounted media.mount - External Media Directory. Mar 2 13:27:43.042954 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 2 13:27:43.100354 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 2 13:27:43.134432 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 2 13:27:43.213547 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 2 13:27:43.494000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:43.804044 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 13:27:43.856000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:43.889140 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 2 13:27:43.894316 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 2 13:27:43.932000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:43.932000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:43.934360 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 13:27:43.941694 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 13:27:43.987000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:43.996000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.006916 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 13:27:44.012241 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 13:27:44.054675 kernel: ACPI: bus type drm_connector registered Mar 2 13:27:44.120000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.120000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.124018 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 13:27:44.124337 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 13:27:44.159000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.169000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.175189 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 2 13:27:44.177493 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 2 13:27:44.202435 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 13:27:44.200000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.200000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.213207 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 13:27:44.245232 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 13:27:44.241000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.241000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.290000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.302477 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 13:27:44.328000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.346359 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 2 13:27:44.385000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.393216 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 2 13:27:44.428000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.440159 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 13:27:44.478000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:44.620734 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 13:27:44.720930 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Mar 2 13:27:44.786098 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 2 13:27:44.853046 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 2 13:27:44.985079 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 2 13:27:44.985493 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 13:27:45.038474 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 2 13:27:45.139312 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 13:27:45.139902 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Mar 2 13:27:45.191251 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 2 13:27:45.303466 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 2 13:27:45.384334 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 13:27:45.400163 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 2 13:27:45.441747 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 13:27:45.487160 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 13:27:45.525088 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 2 13:27:45.597450 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 13:27:45.633308 systemd-journald[1236]: Time spent on flushing to /var/log/journal/c1ba91920c0241e8afa7e0d69a108bd3 is 75.986ms for 1179 entries. Mar 2 13:27:45.633308 systemd-journald[1236]: System Journal (/var/log/journal/c1ba91920c0241e8afa7e0d69a108bd3) is 8M, max 163.5M, 155.5M free. Mar 2 13:27:45.786676 systemd-journald[1236]: Received client request to flush runtime journal. Mar 2 13:27:45.674114 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 2 13:27:45.727325 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 2 13:27:45.798336 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 2 13:27:45.891716 kernel: loop1: detected capacity change from 0 to 111560 Mar 2 13:27:45.885000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:45.897302 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 2 13:27:45.967000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.051115 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 2 13:27:46.156036 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 2 13:27:46.291167 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:27:46.317145 systemd-tmpfiles[1273]: ACLs are not supported, ignoring. Mar 2 13:27:46.317174 systemd-tmpfiles[1273]: ACLs are not supported, ignoring. Mar 2 13:27:46.348000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.424145 kernel: kauditd_printk_skb: 22 callbacks suppressed Mar 2 13:27:46.424351 kernel: audit: type=1130 audit(1772458066.348:129): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.504493 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 13:27:46.545000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.551405 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 2 13:27:46.560527 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 2 13:27:46.658177 kernel: audit: type=1130 audit(1772458066.545:130): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.659000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.696163 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 2 13:27:46.781154 kernel: audit: type=1130 audit(1772458066.659:131): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:46.870267 kernel: loop2: detected capacity change from 0 to 50784 Mar 2 13:27:47.148352 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 2 13:27:47.167000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:47.187091 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Mar 2 13:27:47.249047 kernel: audit: type=1130 audit(1772458067.167:132): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:47.176000 audit: BPF prog-id=18 op=LOAD Mar 2 13:27:47.179000 audit: BPF prog-id=19 op=LOAD Mar 2 13:27:47.352018 kernel: audit: type=1334 audit(1772458067.176:133): prog-id=18 op=LOAD Mar 2 13:27:47.352158 kernel: audit: type=1334 audit(1772458067.179:134): prog-id=19 op=LOAD Mar 2 13:27:47.352193 kernel: audit: type=1334 audit(1772458067.179:135): prog-id=20 op=LOAD Mar 2 13:27:47.179000 audit: BPF prog-id=20 op=LOAD Mar 2 13:27:47.363000 audit: BPF prog-id=21 op=LOAD Mar 2 13:27:47.379267 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 13:27:47.393114 kernel: audit: type=1334 audit(1772458067.363:136): prog-id=21 op=LOAD Mar 2 13:27:47.434218 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 13:27:47.465906 kernel: loop3: detected capacity change from 0 to 219192 Mar 2 13:27:47.517000 audit: BPF prog-id=22 op=LOAD Mar 2 13:27:47.537176 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Mar 2 13:27:47.517000 audit: BPF prog-id=23 op=LOAD Mar 2 13:27:47.570023 kernel: audit: type=1334 audit(1772458067.517:137): prog-id=22 op=LOAD Mar 2 13:27:47.570189 kernel: audit: type=1334 audit(1772458067.517:138): prog-id=23 op=LOAD Mar 2 13:27:47.517000 audit: BPF prog-id=24 op=LOAD Mar 2 13:27:47.615000 audit: BPF prog-id=25 op=LOAD Mar 2 13:27:47.615000 audit: BPF prog-id=26 op=LOAD Mar 2 13:27:47.615000 audit: BPF prog-id=27 op=LOAD Mar 2 13:27:47.621372 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 2 13:27:47.702447 systemd-tmpfiles[1294]: ACLs are not supported, ignoring. Mar 2 13:27:47.702474 systemd-tmpfiles[1294]: ACLs are not supported, ignoring. Mar 2 13:27:47.756362 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 13:27:47.821000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:47.914990 kernel: loop4: detected capacity change from 0 to 111560 Mar 2 13:27:48.101922 systemd-nsresourced[1296]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Mar 2 13:27:48.144990 kernel: loop5: detected capacity change from 0 to 50784 Mar 2 13:27:48.177354 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Mar 2 13:27:48.220000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:48.912370 kernel: loop6: detected capacity change from 0 to 219192 Mar 2 13:27:48.915390 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 2 13:27:48.958000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:49.226281 (sd-merge)[1301]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw'. Mar 2 13:27:50.124162 (sd-merge)[1301]: Merged extensions into '/usr'. Mar 2 13:27:50.204026 systemd[1]: Reload requested from client PID 1272 ('systemd-sysext') (unit systemd-sysext.service)... Mar 2 13:27:50.204128 systemd[1]: Reloading... Mar 2 13:27:52.256712 zram_generator::config[1349]: No configuration found. Mar 2 13:27:52.381263 systemd-oomd[1292]: No swap; memory pressure usage will be degraded Mar 2 13:27:52.573315 systemd-resolved[1293]: Positive Trust Anchors: Mar 2 13:27:52.573416 systemd-resolved[1293]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 13:27:52.573426 systemd-resolved[1293]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Mar 2 13:27:52.573482 systemd-resolved[1293]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 13:27:52.846123 systemd-resolved[1293]: Defaulting to hostname 'linux'. Mar 2 13:27:55.141482 systemd[1]: Reloading finished in 4936 ms. Mar 2 13:27:55.815032 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Mar 2 13:27:55.895000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:55.897505 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 13:27:55.952487 kernel: kauditd_printk_skb: 7 callbacks suppressed Mar 2 13:27:55.952727 kernel: audit: type=1130 audit(1772458075.895:146): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.168000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.245539 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 2 13:27:56.279301 kernel: audit: type=1130 audit(1772458076.168:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.349000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.431389 kernel: audit: type=1130 audit(1772458076.349:148): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.430460 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 2 13:27:56.545000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.656986 kernel: audit: type=1130 audit(1772458076.545:149): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:27:56.722582 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 13:27:57.224596 systemd[1]: Starting ensure-sysext.service... Mar 2 13:27:57.452711 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 13:27:57.585000 audit: BPF prog-id=8 op=UNLOAD Mar 2 13:27:57.612562 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 13:27:57.585000 audit: BPF prog-id=7 op=UNLOAD Mar 2 13:27:57.690010 kernel: audit: type=1334 audit(1772458077.585:150): prog-id=8 op=UNLOAD Mar 2 13:27:57.690148 kernel: audit: type=1334 audit(1772458077.585:151): prog-id=7 op=UNLOAD Mar 2 13:27:57.690202 kernel: audit: type=1334 audit(1772458077.596:152): prog-id=28 op=LOAD Mar 2 13:27:57.596000 audit: BPF prog-id=28 op=LOAD Mar 2 13:27:57.596000 audit: BPF prog-id=29 op=LOAD Mar 2 13:27:57.735357 kernel: audit: type=1334 audit(1772458077.596:153): prog-id=29 op=LOAD Mar 2 13:27:57.624000 audit: BPF prog-id=30 op=LOAD Mar 2 13:27:57.825343 kernel: audit: type=1334 audit(1772458077.624:154): prog-id=30 op=LOAD Mar 2 13:27:57.830889 kernel: audit: type=1334 audit(1772458077.624:155): prog-id=21 op=UNLOAD Mar 2 13:27:57.624000 audit: BPF prog-id=21 op=UNLOAD Mar 2 13:27:57.870000 audit: BPF prog-id=31 op=LOAD Mar 2 13:27:57.870000 audit: BPF prog-id=22 op=UNLOAD Mar 2 13:27:57.870000 audit: BPF prog-id=32 op=LOAD Mar 2 13:27:57.870000 audit: BPF prog-id=33 op=LOAD Mar 2 13:27:57.870000 audit: BPF prog-id=23 op=UNLOAD Mar 2 13:27:57.870000 audit: BPF prog-id=24 op=UNLOAD Mar 2 13:27:57.876000 audit: BPF prog-id=34 op=LOAD Mar 2 13:27:57.876000 audit: BPF prog-id=18 op=UNLOAD Mar 2 13:27:57.876000 audit: BPF prog-id=35 op=LOAD Mar 2 13:27:57.876000 audit: BPF prog-id=36 op=LOAD Mar 2 13:27:57.876000 audit: BPF prog-id=19 op=UNLOAD Mar 2 13:27:57.876000 audit: BPF prog-id=20 op=UNLOAD Mar 2 13:27:57.897000 audit: BPF prog-id=37 op=LOAD Mar 2 13:27:57.897000 audit: BPF prog-id=25 op=UNLOAD Mar 2 13:27:57.920000 audit: BPF prog-id=38 op=LOAD Mar 2 13:27:57.920000 audit: BPF prog-id=39 op=LOAD Mar 2 13:27:57.920000 audit: BPF prog-id=26 op=UNLOAD Mar 2 13:27:57.920000 audit: BPF prog-id=27 op=UNLOAD Mar 2 13:27:57.939000 audit: BPF prog-id=40 op=LOAD Mar 2 13:27:57.939000 audit: BPF prog-id=15 op=UNLOAD Mar 2 13:27:57.955000 audit: BPF prog-id=41 op=LOAD Mar 2 13:27:57.955000 audit: BPF prog-id=42 op=LOAD Mar 2 13:27:57.955000 audit: BPF prog-id=16 op=UNLOAD Mar 2 13:27:57.955000 audit: BPF prog-id=17 op=UNLOAD Mar 2 13:27:58.193610 systemd[1]: Reload requested from client PID 1382 ('systemctl') (unit ensure-sysext.service)... Mar 2 13:27:58.203987 systemd[1]: Reloading... Mar 2 13:27:58.252057 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Mar 2 13:27:58.252109 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Mar 2 13:27:58.272980 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 2 13:27:58.290544 systemd-tmpfiles[1383]: ACLs are not supported, ignoring. Mar 2 13:27:58.290982 systemd-tmpfiles[1383]: ACLs are not supported, ignoring. Mar 2 13:27:58.447251 systemd-tmpfiles[1383]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 13:27:58.447275 systemd-tmpfiles[1383]: Skipping /boot Mar 2 13:27:58.694504 systemd-udevd[1384]: Using default interface naming scheme 'v257'. Mar 2 13:27:58.882179 systemd-tmpfiles[1383]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 13:27:58.882203 systemd-tmpfiles[1383]: Skipping /boot Mar 2 13:27:59.543999 zram_generator::config[1422]: No configuration found. Mar 2 13:28:01.829102 kernel: mousedev: PS/2 mouse device common for all mice Mar 2 13:28:02.193723 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 2 13:28:02.240436 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 2 13:28:02.319421 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 Mar 2 13:28:02.393935 kernel: ACPI: button: Power Button [PWRF] Mar 2 13:28:04.076020 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 2 13:28:04.076596 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 2 13:28:04.119252 systemd[1]: Reloading finished in 5914 ms. Mar 2 13:28:04.470276 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 13:28:04.698605 kernel: kauditd_printk_skb: 24 callbacks suppressed Mar 2 13:28:04.699608 kernel: audit: type=1130 audit(1772458084.646:180): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:04.646000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:05.417242 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 13:28:05.510000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:05.676963 kernel: audit: type=1130 audit(1772458085.510:181): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:05.677090 kernel: audit: type=1334 audit(1772458085.544:182): prog-id=43 op=LOAD Mar 2 13:28:05.544000 audit: BPF prog-id=43 op=LOAD Mar 2 13:28:05.544000 audit: BPF prog-id=44 op=LOAD Mar 2 13:28:05.795575 kernel: audit: type=1334 audit(1772458085.544:183): prog-id=44 op=LOAD Mar 2 13:28:05.890609 kernel: audit: type=1334 audit(1772458085.544:184): prog-id=45 op=LOAD Mar 2 13:28:05.890998 kernel: audit: type=1334 audit(1772458085.544:185): prog-id=40 op=UNLOAD Mar 2 13:28:05.544000 audit: BPF prog-id=45 op=LOAD Mar 2 13:28:05.544000 audit: BPF prog-id=40 op=UNLOAD Mar 2 13:28:05.979425 kernel: audit: type=1334 audit(1772458085.544:186): prog-id=41 op=UNLOAD Mar 2 13:28:06.079148 kernel: audit: type=1334 audit(1772458085.544:187): prog-id=42 op=UNLOAD Mar 2 13:28:06.281476 kernel: audit: type=1334 audit(1772458085.580:188): prog-id=46 op=LOAD Mar 2 13:28:06.317732 kernel: audit: type=1334 audit(1772458085.580:189): prog-id=37 op=UNLOAD Mar 2 13:28:05.544000 audit: BPF prog-id=41 op=UNLOAD Mar 2 13:28:05.544000 audit: BPF prog-id=42 op=UNLOAD Mar 2 13:28:05.580000 audit: BPF prog-id=46 op=LOAD Mar 2 13:28:05.580000 audit: BPF prog-id=37 op=UNLOAD Mar 2 13:28:05.580000 audit: BPF prog-id=47 op=LOAD Mar 2 13:28:05.580000 audit: BPF prog-id=48 op=LOAD Mar 2 13:28:05.580000 audit: BPF prog-id=38 op=UNLOAD Mar 2 13:28:05.580000 audit: BPF prog-id=39 op=UNLOAD Mar 2 13:28:05.580000 audit: BPF prog-id=49 op=LOAD Mar 2 13:28:05.580000 audit: BPF prog-id=50 op=LOAD Mar 2 13:28:05.580000 audit: BPF prog-id=28 op=UNLOAD Mar 2 13:28:05.580000 audit: BPF prog-id=29 op=UNLOAD Mar 2 13:28:05.604000 audit: BPF prog-id=51 op=LOAD Mar 2 13:28:05.604000 audit: BPF prog-id=30 op=UNLOAD Mar 2 13:28:05.611000 audit: BPF prog-id=52 op=LOAD Mar 2 13:28:05.611000 audit: BPF prog-id=34 op=UNLOAD Mar 2 13:28:05.611000 audit: BPF prog-id=53 op=LOAD Mar 2 13:28:05.611000 audit: BPF prog-id=54 op=LOAD Mar 2 13:28:05.625000 audit: BPF prog-id=35 op=UNLOAD Mar 2 13:28:05.625000 audit: BPF prog-id=36 op=UNLOAD Mar 2 13:28:05.639000 audit: BPF prog-id=55 op=LOAD Mar 2 13:28:05.639000 audit: BPF prog-id=31 op=UNLOAD Mar 2 13:28:05.639000 audit: BPF prog-id=56 op=LOAD Mar 2 13:28:05.639000 audit: BPF prog-id=57 op=LOAD Mar 2 13:28:05.639000 audit: BPF prog-id=32 op=UNLOAD Mar 2 13:28:05.639000 audit: BPF prog-id=33 op=UNLOAD Mar 2 13:28:07.379249 systemd[1]: Finished ensure-sysext.service. Mar 2 13:28:07.471000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:08.536020 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:28:08.776306 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 13:28:08.875440 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 2 13:28:08.943507 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 13:28:08.972536 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 13:28:09.692483 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 13:28:10.217065 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 13:28:10.320754 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 13:28:10.421154 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 13:28:10.421443 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Mar 2 13:28:10.477092 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 2 13:28:10.977247 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 2 13:28:10.991408 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 13:28:11.074758 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 2 13:28:11.308268 kernel: kauditd_printk_skb: 23 callbacks suppressed Mar 2 13:28:11.308417 kernel: audit: type=1334 audit(1772458091.201:213): prog-id=58 op=LOAD Mar 2 13:28:11.201000 audit: BPF prog-id=58 op=LOAD Mar 2 13:28:11.398010 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 13:28:12.725947 kernel: audit: type=1334 audit(1772458092.542:214): prog-id=59 op=LOAD Mar 2 13:28:12.542000 audit: BPF prog-id=59 op=LOAD Mar 2 13:28:12.887003 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 2 13:28:13.115514 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 2 13:28:13.550451 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:28:13.551188 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:28:13.584249 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 13:28:13.585436 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 13:28:14.092109 kernel: audit: type=1130 audit(1772458093.605:215): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:14.092299 kernel: audit: type=1131 audit(1772458093.696:216): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:13.605000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:13.696000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:14.035489 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 13:28:15.433022 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 13:28:15.882000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:15.490090 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 13:28:15.912231 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 2 13:28:15.893000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:16.040081 kernel: audit: type=1130 audit(1772458095.882:217): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:16.040144 kernel: audit: type=1131 audit(1772458095.893:218): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:16.107121 kernel: audit: type=1127 audit(1772458096.102:219): pid=1522 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Mar 2 13:28:16.102000 audit[1522]: SYSTEM_BOOT pid=1522 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Mar 2 13:28:16.792502 kernel: audit: type=1130 audit(1772458096.732:220): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:16.732000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:18.419267 kernel: audit: type=1130 audit(1772458098.184:221): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:18.184000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:17.382549 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 2 13:28:18.402646 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 13:28:18.403540 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 13:28:18.702368 kernel: audit: type=1130 audit(1772458098.452:222): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:18.452000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:18.711293 augenrules[1536]: No rules Mar 2 13:28:18.713226 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 2 13:28:18.452000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:18.701000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Mar 2 13:28:18.931240 kernel: audit: type=1131 audit(1772458098.452:223): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 2 13:28:18.931428 kernel: audit: type=1305 audit(1772458098.701:224): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Mar 2 13:28:18.931466 kernel: audit: type=1300 audit(1772458098.701:224): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffebd867af0 a2=420 a3=0 items=0 ppid=1498 pid=1536 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 2 13:28:18.701000 audit[1536]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffebd867af0 a2=420 a3=0 items=0 ppid=1498 pid=1536 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 2 13:28:18.982582 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 13:28:18.986985 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 13:28:19.097342 kernel: audit: type=1327 audit(1772458098.701:224): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Mar 2 13:28:18.701000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Mar 2 13:28:19.155607 systemd[1]: audit-rules.service: Consumed 1.382s CPU time, 3.7M memory peak. Mar 2 13:28:19.163258 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 13:28:19.164408 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 13:28:19.288990 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 2 13:28:19.574584 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 13:28:19.678504 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 2 13:28:21.257966 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 2 13:28:21.925279 systemd-networkd[1516]: lo: Link UP Mar 2 13:28:21.925289 systemd-networkd[1516]: lo: Gained carrier Mar 2 13:28:22.032057 systemd-networkd[1516]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Mar 2 13:28:22.032079 systemd-networkd[1516]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 13:28:22.124394 systemd-networkd[1516]: eth0: Link UP Mar 2 13:28:22.180370 systemd-networkd[1516]: eth0: Gained carrier Mar 2 13:28:22.180420 systemd-networkd[1516]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Mar 2 13:28:22.821226 systemd-networkd[1516]: eth0: DHCPv4 address 10.0.0.31/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 2 13:28:22.834288 systemd-timesyncd[1521]: Network configuration changed, trying to establish connection. Mar 2 13:28:23.552281 systemd-timesyncd[1521]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 2 13:28:23.552943 systemd-timesyncd[1521]: Initial clock synchronization to Mon 2026-03-02 13:28:23.551426 UTC. Mar 2 13:28:23.556001 systemd-resolved[1293]: Clock change detected. Flushing caches. Mar 2 13:28:23.778719 systemd-networkd[1516]: eth0: Gained IPv6LL Mar 2 13:28:27.961383 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 13:28:28.061834 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:28:28.133752 systemd[1]: Reached target network.target - Network. Mar 2 13:28:28.196292 systemd[1]: Reached target time-set.target - System Time Set. Mar 2 13:28:28.240070 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 2 13:28:28.532407 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 2 13:28:30.055115 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 2 13:28:30.124886 systemd[1]: Reached target network-online.target - Network is Online. Mar 2 13:28:30.760930 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 2 13:28:39.975354 ldconfig[1510]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 2 13:28:40.038845 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 2 13:28:40.338868 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 2 13:28:41.000413 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 2 13:28:41.318080 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 13:28:41.940299 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 2 13:28:42.067342 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 2 13:28:42.158332 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Mar 2 13:28:42.266893 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 2 13:28:42.328868 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 2 13:28:42.401231 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Mar 2 13:28:42.475075 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Mar 2 13:28:42.541529 kernel: kvm_amd: TSC scaling supported Mar 2 13:28:42.541990 kernel: kvm_amd: Nested Virtualization enabled Mar 2 13:28:42.542030 kernel: kvm_amd: Nested Paging enabled Mar 2 13:28:42.578429 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Mar 2 13:28:42.652700 kernel: kvm_amd: PMU virtualization is disabled Mar 2 13:28:42.545310 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 2 13:28:42.619318 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 2 13:28:42.620036 systemd[1]: Reached target paths.target - Path Units. Mar 2 13:28:42.654783 systemd[1]: Reached target timers.target - Timer Units. Mar 2 13:28:42.692495 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 2 13:28:42.754296 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 2 13:28:42.825320 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 2 13:28:42.852472 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 2 13:28:42.883150 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 2 13:28:42.918884 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 2 13:28:42.961144 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 2 13:28:42.993381 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 2 13:28:43.032472 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 13:28:43.058320 systemd[1]: Reached target basic.target - Basic System. Mar 2 13:28:43.091174 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 2 13:28:43.091323 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 2 13:28:43.112826 systemd[1]: Starting containerd.service - containerd container runtime... Mar 2 13:28:43.169182 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 2 13:28:43.224008 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 2 13:28:43.339388 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 2 13:28:43.424487 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 2 13:28:43.497399 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 2 13:28:43.558858 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 2 13:28:43.728467 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Mar 2 13:28:43.766459 jq[1568]: false Mar 2 13:28:43.901935 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:28:44.121731 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 2 13:28:44.415878 oslogin_cache_refresh[1570]: Refreshing passwd entry cache Mar 2 13:28:44.797531 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Refreshing passwd entry cache Mar 2 13:28:44.797531 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Failure getting users, quitting Mar 2 13:28:44.797531 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Mar 2 13:28:44.606460 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 2 13:28:44.861421 oslogin_cache_refresh[1570]: Failure getting users, quitting Mar 2 13:28:44.956446 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Refreshing group entry cache Mar 2 13:28:44.933016 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 2 13:28:44.861482 oslogin_cache_refresh[1570]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Mar 2 13:28:44.916985 oslogin_cache_refresh[1570]: Refreshing group entry cache Mar 2 13:28:45.062518 extend-filesystems[1569]: Found /dev/vda6 Mar 2 13:28:45.843470 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Failure getting groups, quitting Mar 2 13:28:45.843470 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Mar 2 13:28:45.110265 oslogin_cache_refresh[1570]: Failure getting groups, quitting Mar 2 13:28:45.130533 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 2 13:28:45.110296 oslogin_cache_refresh[1570]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Mar 2 13:28:45.852913 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 2 13:28:46.031494 extend-filesystems[1569]: Found /dev/vda9 Mar 2 13:28:46.199954 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 2 13:28:46.394144 extend-filesystems[1569]: Checking size of /dev/vda9 Mar 2 13:28:46.388894 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 2 13:28:46.739033 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 2 13:28:46.786024 systemd[1]: Starting update-engine.service - Update Engine... Mar 2 13:28:46.912498 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 2 13:28:46.992206 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 2 13:28:47.031333 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 2 13:28:47.033372 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 2 13:28:47.035463 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Mar 2 13:28:47.038164 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Mar 2 13:28:47.137146 systemd[1]: motdgen.service: Deactivated successfully. Mar 2 13:28:47.138122 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 2 13:28:47.618221 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 2 13:28:48.042444 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 2 13:28:48.509056 extend-filesystems[1569]: Resized partition /dev/vda9 Mar 2 13:28:48.834138 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Mar 2 13:28:48.693266 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 2 13:28:48.834364 extend-filesystems[1611]: resize2fs 1.47.3 (8-Jul-2025) Mar 2 13:28:49.182380 jq[1598]: true Mar 2 13:28:49.550524 jq[1621]: true Mar 2 13:28:49.485005 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 2 13:28:49.699953 tar[1604]: linux-amd64/LICENSE Mar 2 13:28:50.127504 tar[1604]: linux-amd64/helm Mar 2 13:28:51.937091 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 2 13:28:52.739920 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 2 13:28:53.991526 update_engine[1593]: I20260302 13:28:53.935360 1593 main.cc:92] Flatcar Update Engine starting Mar 2 13:28:52.836444 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 2 13:28:55.170874 dbus-daemon[1566]: [system] SELinux support is enabled Mar 2 13:28:55.188230 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 2 13:28:55.356464 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Mar 2 13:28:55.726079 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 2 13:28:55.922091 update_engine[1593]: I20260302 13:28:55.812727 1593 update_check_scheduler.cc:74] Next update check in 2m47s Mar 2 13:28:55.726188 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 2 13:28:55.951907 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 2 13:28:56.364095 extend-filesystems[1611]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 2 13:28:56.364095 extend-filesystems[1611]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 2 13:28:56.364095 extend-filesystems[1611]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Mar 2 13:28:55.951943 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 2 13:28:57.625901 extend-filesystems[1569]: Resized filesystem in /dev/vda9 Mar 2 13:28:55.966236 systemd-logind[1587]: Watching system buttons on /dev/input/event2 (Power Button) Mar 2 13:28:57.852740 sshd_keygen[1596]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 2 13:28:55.966276 systemd-logind[1587]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 2 13:28:55.987036 systemd-logind[1587]: New seat seat0. Mar 2 13:28:56.305217 systemd[1]: Started systemd-logind.service - User Login Management. Mar 2 13:28:56.406329 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 2 13:28:56.407509 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 2 13:28:56.609356 systemd[1]: Started update-engine.service - Update Engine. Mar 2 13:28:57.643480 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 2 13:28:57.902767 bash[1650]: Updated "/home/core/.ssh/authorized_keys" Mar 2 13:28:57.937851 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 2 13:28:58.657186 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 2 13:29:00.450523 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 2 13:29:00.739468 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 2 13:29:01.518345 systemd[1]: Started sshd@0-10.0.0.31:22-10.0.0.1:39046.service - OpenSSH per-connection server daemon (10.0.0.1:39046). Mar 2 13:29:02.930910 systemd[1]: issuegen.service: Deactivated successfully. Mar 2 13:29:02.933430 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 2 13:29:03.070200 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 2 13:29:03.465474 systemd-udevd[1384]: cpu2: Worker [1447] processing SEQNUM=1803 is taking a long time Mar 2 13:29:03.465496 systemd-udevd[1384]: cpu1: Worker [1448] processing SEQNUM=1802 is taking a long time Mar 2 13:29:03.501254 systemd-udevd[1384]: cpu0: Worker [1462] processing SEQNUM=1801 is taking a long time Mar 2 13:29:03.501427 systemd-udevd[1384]: cpu3: Worker [1442] processing SEQNUM=1804 is taking a long time Mar 2 13:29:03.824495 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 2 13:29:03.930332 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 2 13:29:04.031493 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 2 13:29:04.119494 systemd[1]: Reached target getty.target - Login Prompts. Mar 2 13:29:04.268148 locksmithd[1653]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 2 13:29:05.380977 sshd[1664]: Accepted publickey for core from 10.0.0.1 port 39046 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:05.428038 sshd-session[1664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:05.470994 containerd[1614]: time="2026-03-02T13:29:05Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 2 13:29:05.483266 containerd[1614]: time="2026-03-02T13:29:05.478375933Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Mar 2 13:29:05.583892 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 2 13:29:05.673232 containerd[1614]: time="2026-03-02T13:29:05.673098570Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="13.26µs" Mar 2 13:29:05.673389 containerd[1614]: time="2026-03-02T13:29:05.673361749Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 2 13:29:05.673825 containerd[1614]: time="2026-03-02T13:29:05.673798062Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 2 13:29:05.673931 containerd[1614]: time="2026-03-02T13:29:05.673910473Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 2 13:29:05.674361 containerd[1614]: time="2026-03-02T13:29:05.674334649Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 2 13:29:05.674452 containerd[1614]: time="2026-03-02T13:29:05.674432638Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 13:29:05.674821 containerd[1614]: time="2026-03-02T13:29:05.674788398Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 13:29:05.675119 containerd[1614]: time="2026-03-02T13:29:05.675095916Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.675729 containerd[1614]: time="2026-03-02T13:29:05.675524077Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.675828 containerd[1614]: time="2026-03-02T13:29:05.675803983Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 13:29:05.675904 containerd[1614]: time="2026-03-02T13:29:05.675882833Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 13:29:05.675972 containerd[1614]: time="2026-03-02T13:29:05.675954782Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.676524 containerd[1614]: time="2026-03-02T13:29:05.676495775Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.676803 containerd[1614]: time="2026-03-02T13:29:05.676780608Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 2 13:29:05.677387 containerd[1614]: time="2026-03-02T13:29:05.677263532Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.678482 containerd[1614]: time="2026-03-02T13:29:05.678455405Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.690753 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 2 13:29:05.707390 containerd[1614]: time="2026-03-02T13:29:05.705368299Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 13:29:05.707390 containerd[1614]: time="2026-03-02T13:29:05.705430744Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 2 13:29:05.707390 containerd[1614]: time="2026-03-02T13:29:05.705859094Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 2 13:29:05.745512 containerd[1614]: time="2026-03-02T13:29:05.744259674Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 2 13:29:05.745512 containerd[1614]: time="2026-03-02T13:29:05.744797251Z" level=info msg="metadata content store policy set" policy=shared Mar 2 13:29:05.800434 systemd-logind[1587]: New session 1 of user core. Mar 2 13:29:05.922401 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 2 13:29:05.959767 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.964185496Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.964873614Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965099736Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965127049Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965149242Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965167159Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965183645Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965199139Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965215523Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965333102Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965356858Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965390309Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965406664Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 2 13:29:05.973728 containerd[1614]: time="2026-03-02T13:29:05.965424450Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.966809129Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.966842820Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.966862519Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.966879185Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967008652Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967024856Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967054441Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967178560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967198350Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967213874Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967227625Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967257380Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967401269Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.967507200Z" level=info msg="Start snapshots syncer" Mar 2 13:29:05.974488 containerd[1614]: time="2026-03-02T13:29:05.968286775Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 2 13:29:05.979251 containerd[1614]: time="2026-03-02T13:29:05.970442814Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 2 13:29:05.979251 containerd[1614]: time="2026-03-02T13:29:05.970509585Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971012950Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971180004Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971209159Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971225164Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971238002Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971258073Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971274329Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971288460Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971302631Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971315861Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971459188Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971481102Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 13:29:05.979755 containerd[1614]: time="2026-03-02T13:29:05.971493512Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971507153Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971519991Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971533122Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971804354Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971843132Z" level=info msg="runtime interface created" Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971851615Z" level=info msg="created NRI interface" Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971869052Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.971985490Z" level=info msg="Connect containerd service" Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.972016888Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 2 13:29:05.980396 containerd[1614]: time="2026-03-02T13:29:05.975144954Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 13:29:06.040173 (systemd)[1692]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:06.067110 systemd-logind[1587]: New session 2 of user core. Mar 2 13:29:07.433226 containerd[1614]: time="2026-03-02T13:29:07.433167723Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 2 13:29:07.444328 containerd[1614]: time="2026-03-02T13:29:07.444286437Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.438672367Z" level=info msg="Start subscribing containerd event" Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.450737908Z" level=info msg="Start recovering state" Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.450954736Z" level=info msg="Start event monitor" Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.450974117Z" level=info msg="Start cni network conf syncer for default" Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.451069605Z" level=info msg="Start streaming server" Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.451086051Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.451096197Z" level=info msg="runtime interface starting up..." Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.451104641Z" level=info msg="starting plugins..." Mar 2 13:29:07.464802 containerd[1614]: time="2026-03-02T13:29:07.451129499Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 2 13:29:07.487268 systemd[1]: Started containerd.service - containerd container runtime. Mar 2 13:29:07.513274 containerd[1614]: time="2026-03-02T13:29:07.504017209Z" level=info msg="containerd successfully booted in 2.059038s" Mar 2 13:29:07.731718 tar[1604]: linux-amd64/README.md Mar 2 13:29:08.041058 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 2 13:29:08.050671 systemd[1692]: Queued start job for default target default.target. Mar 2 13:29:08.090036 systemd[1692]: Created slice app.slice - User Application Slice. Mar 2 13:29:08.091684 systemd[1692]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Mar 2 13:29:08.091712 systemd[1692]: Reached target paths.target - Paths. Mar 2 13:29:08.091812 systemd[1692]: Reached target timers.target - Timers. Mar 2 13:29:08.140991 systemd[1692]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 2 13:29:08.193969 systemd[1692]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Mar 2 13:29:08.328374 systemd[1692]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 2 13:29:08.328516 systemd[1692]: Reached target sockets.target - Sockets. Mar 2 13:29:08.381066 systemd[1692]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Mar 2 13:29:08.388047 systemd[1692]: Reached target basic.target - Basic System. Mar 2 13:29:08.388349 systemd[1692]: Reached target default.target - Main User Target. Mar 2 13:29:08.388410 systemd[1692]: Startup finished in 2.253s. Mar 2 13:29:08.426743 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 2 13:29:08.483926 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 2 13:29:08.786978 systemd[1]: Started sshd@1-10.0.0.31:22-10.0.0.1:40560.service - OpenSSH per-connection server daemon (10.0.0.1:40560). Mar 2 13:29:09.171803 kernel: EDAC MC: Ver: 3.0.0 Mar 2 13:29:20.873353 sshd[1722]: Accepted publickey for core from 10.0.0.1 port 40560 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:20.890318 sshd-session[1722]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:21.185661 systemd-logind[1587]: New session 3 of user core. Mar 2 13:29:21.248420 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 2 13:29:21.460335 sshd[1731]: Connection closed by 10.0.0.1 port 40560 Mar 2 13:29:21.459120 sshd-session[1722]: pam_unix(sshd:session): session closed for user core Mar 2 13:29:21.528294 systemd[1]: sshd@1-10.0.0.31:22-10.0.0.1:40560.service: Deactivated successfully. Mar 2 13:29:21.536535 systemd[1]: session-3.scope: Deactivated successfully. Mar 2 13:29:21.568054 systemd-logind[1587]: Session 3 logged out. Waiting for processes to exit. Mar 2 13:29:21.576097 systemd-logind[1587]: Removed session 3. Mar 2 13:29:21.598735 systemd[1]: Started sshd@2-10.0.0.31:22-10.0.0.1:43682.service - OpenSSH per-connection server daemon (10.0.0.1:43682). Mar 2 13:29:22.505792 sshd[1737]: Accepted publickey for core from 10.0.0.1 port 43682 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:22.509469 sshd-session[1737]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:22.591783 systemd-logind[1587]: New session 4 of user core. Mar 2 13:29:22.613123 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 2 13:29:23.520379 sshd[1741]: Connection closed by 10.0.0.1 port 43682 Mar 2 13:29:23.518231 sshd-session[1737]: pam_unix(sshd:session): session closed for user core Mar 2 13:29:23.983975 systemd[1]: sshd@2-10.0.0.31:22-10.0.0.1:43682.service: Deactivated successfully. Mar 2 13:29:24.011046 systemd[1]: session-4.scope: Deactivated successfully. Mar 2 13:29:24.039486 systemd-logind[1587]: Session 4 logged out. Waiting for processes to exit. Mar 2 13:29:24.067847 systemd-logind[1587]: Removed session 4. Mar 2 13:29:24.245949 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:29:24.306314 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 2 13:29:24.315195 systemd[1]: Startup finished in 39.219s (kernel) + 1min 33.979s (initrd) + 1min 55.500s (userspace) = 4min 8.698s. Mar 2 13:29:24.355897 (kubelet)[1750]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:29:32.655497 kubelet[1750]: E0302 13:29:32.655172 1750 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:29:32.676171 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:29:32.676787 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:29:32.684175 systemd[1]: kubelet.service: Consumed 5.683s CPU time, 260M memory peak. Mar 2 13:29:33.592506 systemd[1]: Started sshd@3-10.0.0.31:22-10.0.0.1:58182.service - OpenSSH per-connection server daemon (10.0.0.1:58182). Mar 2 13:29:33.991847 sshd[1761]: Accepted publickey for core from 10.0.0.1 port 58182 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:34.011496 sshd-session[1761]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:34.118270 systemd-logind[1587]: New session 5 of user core. Mar 2 13:29:34.159777 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 2 13:29:34.401398 sshd[1765]: Connection closed by 10.0.0.1 port 58182 Mar 2 13:29:34.407034 sshd-session[1761]: pam_unix(sshd:session): session closed for user core Mar 2 13:29:34.487947 systemd[1]: Started sshd@4-10.0.0.31:22-10.0.0.1:58190.service - OpenSSH per-connection server daemon (10.0.0.1:58190). Mar 2 13:29:34.489084 systemd[1]: sshd@3-10.0.0.31:22-10.0.0.1:58182.service: Deactivated successfully. Mar 2 13:29:34.513984 systemd[1]: session-5.scope: Deactivated successfully. Mar 2 13:29:34.538288 systemd-logind[1587]: Session 5 logged out. Waiting for processes to exit. Mar 2 13:29:34.568152 systemd-logind[1587]: Removed session 5. Mar 2 13:29:34.905355 sshd[1768]: Accepted publickey for core from 10.0.0.1 port 58190 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:34.912266 sshd-session[1768]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:34.957901 systemd-logind[1587]: New session 6 of user core. Mar 2 13:29:35.150284 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 2 13:29:35.570320 sshd[1775]: Connection closed by 10.0.0.1 port 58190 Mar 2 13:29:35.555307 sshd-session[1768]: pam_unix(sshd:session): session closed for user core Mar 2 13:29:35.621256 systemd[1]: sshd@4-10.0.0.31:22-10.0.0.1:58190.service: Deactivated successfully. Mar 2 13:29:35.635460 systemd[1]: session-6.scope: Deactivated successfully. Mar 2 13:29:35.699333 systemd[1]: Started sshd@5-10.0.0.31:22-10.0.0.1:58196.service - OpenSSH per-connection server daemon (10.0.0.1:58196). Mar 2 13:29:35.709256 systemd-logind[1587]: Session 6 logged out. Waiting for processes to exit. Mar 2 13:29:35.746139 systemd-logind[1587]: Removed session 6. Mar 2 13:29:35.990690 sshd[1781]: Accepted publickey for core from 10.0.0.1 port 58196 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:36.010874 sshd-session[1781]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:36.075415 systemd-logind[1587]: New session 7 of user core. Mar 2 13:29:36.102124 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 2 13:29:36.333773 sshd[1785]: Connection closed by 10.0.0.1 port 58196 Mar 2 13:29:36.330467 sshd-session[1781]: pam_unix(sshd:session): session closed for user core Mar 2 13:29:36.372414 systemd[1]: sshd@5-10.0.0.31:22-10.0.0.1:58196.service: Deactivated successfully. Mar 2 13:29:36.382746 systemd[1]: session-7.scope: Deactivated successfully. Mar 2 13:29:36.398733 systemd-logind[1587]: Session 7 logged out. Waiting for processes to exit. Mar 2 13:29:36.407932 systemd[1]: Started sshd@6-10.0.0.31:22-10.0.0.1:58202.service - OpenSSH per-connection server daemon (10.0.0.1:58202). Mar 2 13:29:36.416183 systemd-logind[1587]: Removed session 7. Mar 2 13:29:36.614917 sshd[1791]: Accepted publickey for core from 10.0.0.1 port 58202 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:29:36.624685 sshd-session[1791]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:29:36.661690 systemd-logind[1587]: New session 8 of user core. Mar 2 13:29:36.682269 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 2 13:29:36.852891 sudo[1796]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 2 13:29:36.856142 sudo[1796]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 13:29:40.631183 update_engine[1593]: I20260302 13:29:40.627023 1593 update_attempter.cc:509] Updating boot flags... Mar 2 13:29:43.018827 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 2 13:29:43.097008 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:29:48.122714 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 2 13:29:48.287990 (dockerd)[1840]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 2 13:29:51.049776 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:29:51.335298 (kubelet)[1846]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:29:55.417097 kubelet[1846]: E0302 13:29:55.415692 1846 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:29:55.479772 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:29:55.480216 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:29:55.485009 systemd[1]: kubelet.service: Consumed 2.291s CPU time, 111M memory peak. Mar 2 13:30:04.615838 dockerd[1840]: time="2026-03-02T13:30:04.606332113Z" level=info msg="Starting up" Mar 2 13:30:04.628904 dockerd[1840]: time="2026-03-02T13:30:04.625495506Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 2 13:30:04.894915 dockerd[1840]: time="2026-03-02T13:30:04.884140021Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Mar 2 13:30:05.816533 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 2 13:30:05.870150 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:30:06.148261 systemd[1]: var-lib-docker-metacopy\x2dcheck1929708419-merged.mount: Deactivated successfully. Mar 2 13:30:06.694250 dockerd[1840]: time="2026-03-02T13:30:06.680679598Z" level=info msg="Loading containers: start." Mar 2 13:30:07.049179 kernel: Initializing XFRM netlink socket Mar 2 13:30:08.665143 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:30:08.710330 (kubelet)[1914]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:30:10.591280 kubelet[1914]: E0302 13:30:10.586491 1914 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:30:10.795645 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:30:10.796289 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:30:10.832707 systemd[1]: kubelet.service: Consumed 1.603s CPU time, 110.5M memory peak. Mar 2 13:30:16.310495 systemd-networkd[1516]: docker0: Link UP Mar 2 13:30:16.357534 dockerd[1840]: time="2026-03-02T13:30:16.354895870Z" level=info msg="Loading containers: done." Mar 2 13:30:16.438293 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4004042631-merged.mount: Deactivated successfully. Mar 2 13:30:16.477756 dockerd[1840]: time="2026-03-02T13:30:16.475061920Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 2 13:30:16.480966 dockerd[1840]: time="2026-03-02T13:30:16.479042750Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Mar 2 13:30:16.487470 dockerd[1840]: time="2026-03-02T13:30:16.482217234Z" level=info msg="Initializing buildkit" Mar 2 13:30:17.120873 dockerd[1840]: time="2026-03-02T13:30:17.119932358Z" level=info msg="Completed buildkit initialization" Mar 2 13:30:17.221520 dockerd[1840]: time="2026-03-02T13:30:17.220353478Z" level=info msg="Daemon has completed initialization" Mar 2 13:30:17.222370 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 2 13:30:17.230417 dockerd[1840]: time="2026-03-02T13:30:17.223894007Z" level=info msg="API listen on /run/docker.sock" Mar 2 13:30:21.093066 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 2 13:30:21.154121 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:30:26.982061 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:30:27.006129 (kubelet)[2097]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:30:27.959247 containerd[1614]: time="2026-03-02T13:30:27.958527443Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\"" Mar 2 13:30:30.600199 kubelet[2097]: E0302 13:30:30.599959 2097 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:30:30.621373 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:30:30.621922 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:30:30.625034 systemd[1]: kubelet.service: Consumed 2.934s CPU time, 110.5M memory peak. Mar 2 13:30:35.083411 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3507485966.mount: Deactivated successfully. Mar 2 13:30:40.730805 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 2 13:30:40.874492 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:30:48.536059 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:30:48.794290 (kubelet)[2175]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:30:50.256306 kubelet[2175]: E0302 13:30:50.255316 2175 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:30:50.297374 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:30:50.310365 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:30:50.312826 systemd[1]: kubelet.service: Consumed 2.235s CPU time, 110.7M memory peak. Mar 2 13:30:57.102517 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 3781086601 wd_nsec: 3781086327 Mar 2 13:31:00.605052 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Mar 2 13:31:01.192375 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:31:06.298497 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:31:06.411750 (kubelet)[2191]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:31:09.031807 kubelet[2191]: E0302 13:31:09.031181 2191 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:31:09.051122 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:31:09.051500 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:31:09.059237 systemd[1]: kubelet.service: Consumed 2.220s CPU time, 110M memory peak. Mar 2 13:31:12.033748 containerd[1614]: time="2026-03-02T13:31:12.032672802Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:12.046331 containerd[1614]: time="2026-03-02T13:31:12.045834016Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.5: active requests=0, bytes read=27063593" Mar 2 13:31:12.065961 containerd[1614]: time="2026-03-02T13:31:12.060402445Z" level=info msg="ImageCreate event name:\"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:12.126767 containerd[1614]: time="2026-03-02T13:31:12.124935793Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:12.137468 containerd[1614]: time="2026-03-02T13:31:12.135497598Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.5\" with image id \"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\", size \"27071096\" in 44.152973609s" Mar 2 13:31:12.137468 containerd[1614]: time="2026-03-02T13:31:12.135861446Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\" returns image reference \"sha256:364ea2876e41b29691964751b6217cd2e343433690fbe16a5c6a236042684df3\"" Mar 2 13:31:12.153404 containerd[1614]: time="2026-03-02T13:31:12.151369912Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\"" Mar 2 13:31:19.226785 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Mar 2 13:31:19.254228 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:31:21.843290 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:31:21.924917 (kubelet)[2212]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:31:23.240461 kubelet[2212]: E0302 13:31:23.238735 2212 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:31:23.260849 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:31:23.273926 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:31:23.278030 systemd[1]: kubelet.service: Consumed 1.391s CPU time, 110.5M memory peak. Mar 2 13:31:26.412738 containerd[1614]: time="2026-03-02T13:31:26.409034018Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:26.484626 containerd[1614]: time="2026-03-02T13:31:26.482170755Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.5: active requests=0, bytes read=21157668" Mar 2 13:31:26.745300 containerd[1614]: time="2026-03-02T13:31:26.656150953Z" level=info msg="ImageCreate event name:\"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:26.937860 containerd[1614]: time="2026-03-02T13:31:26.934750475Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:26.940827 containerd[1614]: time="2026-03-02T13:31:26.939974029Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.5\" with image id \"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\", size \"22822771\" in 14.788551157s" Mar 2 13:31:26.940827 containerd[1614]: time="2026-03-02T13:31:26.940025507Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\" returns image reference \"sha256:8926c34822743bb97f9003f92c30127bfeaad8bed71cd36f1c861ed8fda2c154\"" Mar 2 13:31:26.947706 containerd[1614]: time="2026-03-02T13:31:26.947281535Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\"" Mar 2 13:31:33.487283 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Mar 2 13:31:33.525877 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:31:36.566519 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:31:36.882757 (kubelet)[2233]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:31:38.907821 kubelet[2233]: E0302 13:31:38.891121 2233 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:31:39.008157 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:31:39.008934 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:31:39.015512 systemd[1]: kubelet.service: Consumed 1.476s CPU time, 112.2M memory peak. Mar 2 13:31:41.097289 containerd[1614]: time="2026-03-02T13:31:41.095284700Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:41.107981 containerd[1614]: time="2026-03-02T13:31:41.107908092Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.5: active requests=0, bytes read=15721687" Mar 2 13:31:41.116450 containerd[1614]: time="2026-03-02T13:31:41.116319389Z" level=info msg="ImageCreate event name:\"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:41.134951 containerd[1614]: time="2026-03-02T13:31:41.134658320Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:31:41.150040 containerd[1614]: time="2026-03-02T13:31:41.139531457Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.5\" with image id \"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\", size \"17386790\" in 14.192204557s" Mar 2 13:31:41.150040 containerd[1614]: time="2026-03-02T13:31:41.140269318Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\" returns image reference \"sha256:f6b3520b1732b4980b2528fe5622e62be26bb6a8d38da81349cb6ccd3a1e6d65\"" Mar 2 13:31:41.151707 containerd[1614]: time="2026-03-02T13:31:41.151398745Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\"" Mar 2 13:31:42.720765 update_engine[1593]: I20260302 13:31:42.719519 1593 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Mar 2 13:31:42.720765 update_engine[1593]: I20260302 13:31:42.720078 1593 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Mar 2 13:31:42.779033 update_engine[1593]: I20260302 13:31:42.729158 1593 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Mar 2 13:31:42.823121 update_engine[1593]: I20260302 13:31:42.821493 1593 omaha_request_params.cc:62] Current group set to alpha Mar 2 13:31:42.844000 update_engine[1593]: I20260302 13:31:42.832136 1593 update_attempter.cc:499] Already updated boot flags. Skipping. Mar 2 13:31:42.844000 update_engine[1593]: I20260302 13:31:42.832265 1593 update_attempter.cc:643] Scheduling an action processor start. Mar 2 13:31:42.844000 update_engine[1593]: I20260302 13:31:42.832301 1593 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Mar 2 13:31:42.844000 update_engine[1593]: I20260302 13:31:42.833302 1593 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Mar 2 13:31:42.845113 update_engine[1593]: I20260302 13:31:42.839494 1593 omaha_request_action.cc:271] Posting an Omaha request to disabled Mar 2 13:31:42.845113 update_engine[1593]: I20260302 13:31:42.844288 1593 omaha_request_action.cc:272] Request: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.845113 update_engine[1593]: Mar 2 13:31:42.846004 update_engine[1593]: I20260302 13:31:42.844455 1593 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 2 13:31:42.946478 update_engine[1593]: I20260302 13:31:42.943357 1593 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 2 13:31:42.951725 update_engine[1593]: I20260302 13:31:42.951206 1593 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 2 13:31:43.009478 locksmithd[1653]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Mar 2 13:31:43.033755 update_engine[1593]: E20260302 13:31:43.033643 1593 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Mar 2 13:31:43.034300 update_engine[1593]: I20260302 13:31:43.034269 1593 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Mar 2 13:31:49.421034 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Mar 2 13:31:49.598117 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:31:53.646761 update_engine[1593]: I20260302 13:31:53.629355 1593 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 2 13:31:53.646761 update_engine[1593]: I20260302 13:31:53.635715 1593 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 2 13:31:53.689851 update_engine[1593]: I20260302 13:31:53.659684 1593 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 2 13:31:53.699254 update_engine[1593]: E20260302 13:31:53.699119 1593 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Mar 2 13:31:53.699839 update_engine[1593]: I20260302 13:31:53.699808 1593 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Mar 2 13:31:54.542174 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount438481308.mount: Deactivated successfully. Mar 2 13:31:55.287291 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:31:55.604389 (kubelet)[2254]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:31:56.907905 kubelet[2254]: E0302 13:31:56.906425 2254 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:31:56.930387 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:31:56.931974 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:31:56.944008 systemd[1]: kubelet.service: Consumed 2.777s CPU time, 110.6M memory peak. Mar 2 13:32:03.669778 update_engine[1593]: I20260302 13:32:03.656420 1593 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 2 13:32:03.669778 update_engine[1593]: I20260302 13:32:03.672024 1593 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 2 13:32:03.740492 update_engine[1593]: I20260302 13:32:03.737071 1593 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 2 13:32:03.772379 update_engine[1593]: E20260302 13:32:03.771093 1593 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Mar 2 13:32:03.776466 update_engine[1593]: I20260302 13:32:03.775079 1593 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Mar 2 13:32:03.786760 containerd[1614]: time="2026-03-02T13:32:03.785871828Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:03.805488 containerd[1614]: time="2026-03-02T13:32:03.799450360Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.5: active requests=0, bytes read=25858690" Mar 2 13:32:03.811818 containerd[1614]: time="2026-03-02T13:32:03.809824273Z" level=info msg="ImageCreate event name:\"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:03.835968 containerd[1614]: time="2026-03-02T13:32:03.832523512Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:03.835968 containerd[1614]: time="2026-03-02T13:32:03.834156753Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.5\" with image id \"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\", repo tag \"registry.k8s.io/kube-proxy:v1.34.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\", size \"25860789\" in 22.682648072s" Mar 2 13:32:03.835968 containerd[1614]: time="2026-03-02T13:32:03.834194555Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\" returns image reference \"sha256:38728cde323c302ed9eca4f1b7c0080d17db50144e39398fcf901d9df13f0c3e\"" Mar 2 13:32:03.863487 containerd[1614]: time="2026-03-02T13:32:03.860003104Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Mar 2 13:32:06.978245 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Mar 2 13:32:07.007239 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:32:07.619997 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3368106695.mount: Deactivated successfully. Mar 2 13:32:11.102938 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:32:11.181480 (kubelet)[2286]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:32:13.797389 update_engine[1593]: I20260302 13:32:13.731259 1593 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 2 13:32:13.797389 update_engine[1593]: I20260302 13:32:13.746776 1593 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 2 13:32:13.797389 update_engine[1593]: I20260302 13:32:13.749056 1593 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 2 13:32:13.941293 update_engine[1593]: E20260302 13:32:13.934376 1593 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Mar 2 13:32:13.941293 update_engine[1593]: I20260302 13:32:13.934988 1593 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Mar 2 13:32:13.941293 update_engine[1593]: I20260302 13:32:13.935009 1593 omaha_request_action.cc:617] Omaha request response: Mar 2 13:32:13.941293 update_engine[1593]: E20260302 13:32:13.937340 1593 omaha_request_action.cc:636] Omaha request network transfer failed. Mar 2 13:32:13.941293 update_engine[1593]: I20260302 13:32:13.939250 1593 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Mar 2 13:32:13.941293 update_engine[1593]: I20260302 13:32:13.939270 1593 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 2 13:32:13.941293 update_engine[1593]: I20260302 13:32:13.939280 1593 update_attempter.cc:306] Processing Done. Mar 2 13:32:14.308981 update_engine[1593]: E20260302 13:32:14.008271 1593 update_attempter.cc:619] Update failed. Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.010969 1593 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.010987 1593 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.010999 1593 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.011288 1593 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.037017 1593 omaha_request_action.cc:271] Posting an Omaha request to disabled Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.037261 1593 omaha_request_action.cc:272] Request: Mar 2 13:32:14.308981 update_engine[1593]: Mar 2 13:32:14.308981 update_engine[1593]: Mar 2 13:32:14.308981 update_engine[1593]: Mar 2 13:32:14.308981 update_engine[1593]: Mar 2 13:32:14.308981 update_engine[1593]: Mar 2 13:32:14.308981 update_engine[1593]: Mar 2 13:32:14.308981 update_engine[1593]: I20260302 13:32:14.037274 1593 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Mar 2 13:32:14.339135 update_engine[1593]: I20260302 13:32:14.333266 1593 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Mar 2 13:32:14.344120 update_engine[1593]: I20260302 13:32:14.341928 1593 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Mar 2 13:32:14.344220 locksmithd[1653]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Mar 2 13:32:14.942183 locksmithd[1653]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Mar 2 13:32:15.022018 update_engine[1593]: E20260302 13:32:14.431958 1593 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled (Domain name not found) Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.433362 1593 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.433387 1593 omaha_request_action.cc:617] Omaha request response: Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.433403 1593 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.433413 1593 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.433423 1593 update_attempter.cc:306] Processing Done. Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.446844 1593 update_attempter.cc:310] Error event sent. Mar 2 13:32:15.022018 update_engine[1593]: I20260302 13:32:14.449990 1593 update_check_scheduler.cc:74] Next update check in 43m7s Mar 2 13:32:15.995925 kubelet[2286]: E0302 13:32:15.935067 2286 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:32:16.205806 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:32:16.214466 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:32:16.227733 systemd[1]: kubelet.service: Consumed 2.042s CPU time, 112.3M memory peak. Mar 2 13:32:26.218539 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Mar 2 13:32:26.251255 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:32:32.580111 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:32:32.714778 (kubelet)[2344]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:32:34.257492 kubelet[2344]: E0302 13:32:34.254372 2344 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:32:34.283785 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:32:34.286397 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:32:34.290482 systemd[1]: kubelet.service: Consumed 1.905s CPU time, 110.4M memory peak. Mar 2 13:32:44.377449 containerd[1614]: time="2026-03-02T13:32:44.373522551Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:44.437978 containerd[1614]: time="2026-03-02T13:32:44.381444280Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=22377210" Mar 2 13:32:44.488793 containerd[1614]: time="2026-03-02T13:32:44.471528236Z" level=info msg="ImageCreate event name:\"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:44.605203 containerd[1614]: time="2026-03-02T13:32:44.597450285Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:44.693781 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Mar 2 13:32:44.717238 containerd[1614]: time="2026-03-02T13:32:44.715478500Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"22384805\" in 40.855352504s" Mar 2 13:32:44.717238 containerd[1614]: time="2026-03-02T13:32:44.715801809Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\"" Mar 2 13:32:44.793508 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:32:44.951709 containerd[1614]: time="2026-03-02T13:32:44.942319928Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Mar 2 13:32:47.392428 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2125196779.mount: Deactivated successfully. Mar 2 13:32:47.786501 containerd[1614]: time="2026-03-02T13:32:47.748137736Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:47.900259 containerd[1614]: time="2026-03-02T13:32:47.891791966Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=316649" Mar 2 13:32:47.962927 containerd[1614]: time="2026-03-02T13:32:47.961883667Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:48.019700 containerd[1614]: time="2026-03-02T13:32:48.019058710Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:32:48.021457 containerd[1614]: time="2026-03-02T13:32:48.020840116Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 3.078443123s" Mar 2 13:32:48.021457 containerd[1614]: time="2026-03-02T13:32:48.020957829Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Mar 2 13:32:48.051782 containerd[1614]: time="2026-03-02T13:32:48.048078713Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\"" Mar 2 13:32:48.613994 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:32:48.706780 (kubelet)[2367]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:32:49.853179 kubelet[2367]: E0302 13:32:49.853048 2367 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:32:49.886026 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:32:49.950997 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:32:49.959484 systemd[1]: kubelet.service: Consumed 1.485s CPU time, 110.5M memory peak. Mar 2 13:32:52.579062 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3082468215.mount: Deactivated successfully. Mar 2 13:33:01.014218 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Mar 2 13:33:01.047166 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:33:06.855157 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:33:07.424920 (kubelet)[2441]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:33:10.727300 kubelet[2441]: E0302 13:33:10.702525 2441 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:33:10.731522 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:33:10.748000 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:33:10.781121 systemd[1]: kubelet.service: Consumed 3.297s CPU time, 108.8M memory peak. Mar 2 13:33:20.563296 containerd[1614]: time="2026-03-02T13:33:20.558770537Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.5-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:33:20.577111 containerd[1614]: time="2026-03-02T13:33:20.577055913Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.5-0: active requests=0, bytes read=22849743" Mar 2 13:33:20.592954 containerd[1614]: time="2026-03-02T13:33:20.585668871Z" level=info msg="ImageCreate event name:\"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:33:20.605142 containerd[1614]: time="2026-03-02T13:33:20.604229636Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:33:20.613667 containerd[1614]: time="2026-03-02T13:33:20.612810429Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.5-0\" with image id \"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\", repo tag \"registry.k8s.io/etcd:3.6.5-0\", repo digest \"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\", size \"22871747\" in 32.564671161s" Mar 2 13:33:20.615765 containerd[1614]: time="2026-03-02T13:33:20.612964449Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\" returns image reference \"sha256:a3e246e9556e93d71e2850085ba581b376c76a9187b4b8a01c120f86579ef2b1\"" Mar 2 13:33:21.220264 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. Mar 2 13:33:21.290070 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:33:22.271029 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:33:22.327094 (kubelet)[2494]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:33:24.008993 kubelet[2494]: E0302 13:33:24.008054 2494 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:33:24.015306 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:33:24.017335 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:33:24.020164 systemd[1]: kubelet.service: Consumed 949ms CPU time, 108.9M memory peak. Mar 2 13:33:33.657441 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:33:33.660874 systemd[1]: kubelet.service: Consumed 949ms CPU time, 108.9M memory peak. Mar 2 13:33:33.673231 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:33:33.781378 systemd[1]: Reload requested from client PID 2513 ('systemctl') (unit session-8.scope)... Mar 2 13:33:33.781470 systemd[1]: Reloading... Mar 2 13:33:34.195181 zram_generator::config[2562]: No configuration found. Mar 2 13:33:35.138913 systemd[1]: Reloading finished in 1356 ms. Mar 2 13:33:35.464031 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:33:35.486927 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:33:35.505272 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 13:33:35.507761 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:33:35.507895 systemd[1]: kubelet.service: Consumed 315ms CPU time, 98.4M memory peak. Mar 2 13:33:35.528016 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:33:36.421348 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:33:36.460487 (kubelet)[2610]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 13:33:36.836784 kubelet[2610]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 13:33:36.836784 kubelet[2610]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 13:33:36.836784 kubelet[2610]: I0302 13:33:36.833291 2610 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 13:33:45.096742 kubelet[2610]: I0302 13:33:45.095258 2610 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 2 13:33:45.096742 kubelet[2610]: I0302 13:33:45.095371 2610 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 13:33:45.098959 kubelet[2610]: I0302 13:33:45.098783 2610 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 2 13:33:45.098959 kubelet[2610]: I0302 13:33:45.098808 2610 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 13:33:45.103729 kubelet[2610]: I0302 13:33:45.102681 2610 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 13:33:45.208866 kubelet[2610]: I0302 13:33:45.206123 2610 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 13:33:45.208866 kubelet[2610]: E0302 13:33:45.206932 2610 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.31:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:33:45.408941 kubelet[2610]: I0302 13:33:45.401017 2610 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 13:33:45.889113 kubelet[2610]: I0302 13:33:45.887762 2610 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 2 13:33:45.908343 kubelet[2610]: I0302 13:33:45.899342 2610 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 13:33:45.908343 kubelet[2610]: I0302 13:33:45.902216 2610 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 13:33:45.908343 kubelet[2610]: I0302 13:33:45.906866 2610 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 13:33:45.908343 kubelet[2610]: I0302 13:33:45.906888 2610 container_manager_linux.go:306] "Creating device plugin manager" Mar 2 13:33:45.913990 kubelet[2610]: I0302 13:33:45.911100 2610 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 2 13:33:45.933802 kubelet[2610]: I0302 13:33:45.928186 2610 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:33:45.933802 kubelet[2610]: I0302 13:33:45.929317 2610 kubelet.go:475] "Attempting to sync node with API server" Mar 2 13:33:45.933802 kubelet[2610]: I0302 13:33:45.929347 2610 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 13:33:45.933802 kubelet[2610]: I0302 13:33:45.929709 2610 kubelet.go:387] "Adding apiserver pod source" Mar 2 13:33:45.933802 kubelet[2610]: I0302 13:33:45.929733 2610 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 13:33:45.942075 kubelet[2610]: E0302 13:33:45.937372 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.31:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:33:45.942075 kubelet[2610]: E0302 13:33:45.938366 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.31:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:33:45.962230 kubelet[2610]: I0302 13:33:45.960977 2610 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Mar 2 13:33:45.974191 kubelet[2610]: I0302 13:33:45.970146 2610 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 13:33:45.974191 kubelet[2610]: I0302 13:33:45.970188 2610 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 2 13:33:45.974191 kubelet[2610]: W0302 13:33:45.970305 2610 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 2 13:33:46.010305 kubelet[2610]: I0302 13:33:46.010164 2610 server.go:1262] "Started kubelet" Mar 2 13:33:46.014415 kubelet[2610]: I0302 13:33:46.011246 2610 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 13:33:46.014415 kubelet[2610]: I0302 13:33:46.011904 2610 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 2 13:33:46.031131 kubelet[2610]: I0302 13:33:46.031109 2610 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 13:33:46.031273 kubelet[2610]: I0302 13:33:46.013788 2610 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 13:33:46.055056 kubelet[2610]: I0302 13:33:46.048287 2610 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 13:33:46.055056 kubelet[2610]: I0302 13:33:46.051059 2610 server.go:310] "Adding debug handlers to kubelet server" Mar 2 13:33:46.071228 kubelet[2610]: I0302 13:33:46.068800 2610 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 2 13:33:46.089835 kubelet[2610]: E0302 13:33:46.071296 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.089835 kubelet[2610]: I0302 13:33:46.085320 2610 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 13:33:46.100929 kubelet[2610]: E0302 13:33:46.091049 2610 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.31:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.31:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1899098c2d77154f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,LastTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 13:33:46.107962 kubelet[2610]: I0302 13:33:46.104120 2610 factory.go:223] Registration of the systemd container factory successfully Mar 2 13:33:46.107962 kubelet[2610]: I0302 13:33:46.104319 2610 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 13:33:46.107962 kubelet[2610]: I0302 13:33:46.104792 2610 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 2 13:33:46.107962 kubelet[2610]: I0302 13:33:46.104854 2610 reconciler.go:29] "Reconciler: start to sync state" Mar 2 13:33:46.107962 kubelet[2610]: E0302 13:33:46.105311 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.31:6443: connect: connection refused" interval="200ms" Mar 2 13:33:46.113167 kubelet[2610]: E0302 13:33:46.108841 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.31:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:33:46.113167 kubelet[2610]: E0302 13:33:46.108983 2610 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 13:33:46.139824 kubelet[2610]: I0302 13:33:46.135065 2610 factory.go:223] Registration of the containerd container factory successfully Mar 2 13:33:46.199881 kubelet[2610]: E0302 13:33:46.197815 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.459520 kubelet[2610]: E0302 13:33:46.382705 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.488765 kubelet[2610]: E0302 13:33:46.488245 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.31:6443: connect: connection refused" interval="400ms" Mar 2 13:33:46.489340 kubelet[2610]: E0302 13:33:46.489227 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.593108 kubelet[2610]: E0302 13:33:46.592386 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.701076 kubelet[2610]: E0302 13:33:46.695097 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.736984 kubelet[2610]: I0302 13:33:46.735973 2610 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 13:33:46.736984 kubelet[2610]: I0302 13:33:46.736000 2610 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 13:33:46.736984 kubelet[2610]: I0302 13:33:46.736027 2610 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:33:46.800318 kubelet[2610]: I0302 13:33:46.799835 2610 policy_none.go:49] "None policy: Start" Mar 2 13:33:46.808427 kubelet[2610]: I0302 13:33:46.807415 2610 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 2 13:33:46.808427 kubelet[2610]: E0302 13:33:46.802977 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.808427 kubelet[2610]: I0302 13:33:46.807539 2610 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 2 13:33:46.808427 kubelet[2610]: E0302 13:33:46.804225 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.31:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:33:46.840855 kubelet[2610]: I0302 13:33:46.840812 2610 policy_none.go:47] "Start" Mar 2 13:33:46.917192 kubelet[2610]: E0302 13:33:46.909152 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:46.917192 kubelet[2610]: E0302 13:33:46.915105 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.31:6443: connect: connection refused" interval="800ms" Mar 2 13:33:46.996313 kubelet[2610]: I0302 13:33:46.994866 2610 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 2 13:33:47.017954 kubelet[2610]: E0302 13:33:47.017896 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:47.030219 kubelet[2610]: I0302 13:33:47.029702 2610 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 2 13:33:47.039368 kubelet[2610]: I0302 13:33:47.038013 2610 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 2 13:33:47.041342 kubelet[2610]: I0302 13:33:47.041318 2610 kubelet.go:2428] "Starting kubelet main sync loop" Mar 2 13:33:47.041933 kubelet[2610]: E0302 13:33:47.041902 2610 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 13:33:47.042998 kubelet[2610]: E0302 13:33:47.042856 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.31:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:33:47.054707 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 2 13:33:47.136938 kubelet[2610]: E0302 13:33:47.131631 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:47.145019 kubelet[2610]: E0302 13:33:47.143873 2610 kubelet.go:2452] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 2 13:33:47.292363 kubelet[2610]: E0302 13:33:47.257330 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:47.292363 kubelet[2610]: E0302 13:33:47.277160 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.31:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:33:47.292363 kubelet[2610]: E0302 13:33:47.288138 2610 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.31:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:33:47.340789 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 2 13:33:47.346120 kubelet[2610]: E0302 13:33:47.344254 2610 kubelet.go:2452] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 2 13:33:47.366125 kubelet[2610]: E0302 13:33:47.365809 2610 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:33:47.420708 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 2 13:33:47.432901 kubelet[2610]: E0302 13:33:47.430408 2610 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 13:33:47.432901 kubelet[2610]: I0302 13:33:47.431168 2610 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 13:33:47.432901 kubelet[2610]: I0302 13:33:47.431186 2610 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 13:33:47.432901 kubelet[2610]: I0302 13:33:47.432007 2610 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 13:33:47.459656 kubelet[2610]: E0302 13:33:47.458917 2610 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 13:33:47.459656 kubelet[2610]: E0302 13:33:47.459051 2610 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:33:47.482975 kubelet[2610]: E0302 13:33:47.482917 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.31:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:33:47.624224 kubelet[2610]: I0302 13:33:47.622397 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:33:47.633275 kubelet[2610]: E0302 13:33:47.632203 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": dial tcp 10.0.0.31:6443: connect: connection refused" node="localhost" Mar 2 13:33:47.718084 kubelet[2610]: E0302 13:33:47.716447 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.31:6443: connect: connection refused" interval="1.6s" Mar 2 13:33:47.978361 kubelet[2610]: I0302 13:33:47.877455 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/08b2d708c40b0a38b3e0e2d2f01c0e0d-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"08b2d708c40b0a38b3e0e2d2f01c0e0d\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:33:47.978361 kubelet[2610]: I0302 13:33:47.881324 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/08b2d708c40b0a38b3e0e2d2f01c0e0d-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"08b2d708c40b0a38b3e0e2d2f01c0e0d\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:33:47.978361 kubelet[2610]: I0302 13:33:47.881729 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/08b2d708c40b0a38b3e0e2d2f01c0e0d-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"08b2d708c40b0a38b3e0e2d2f01c0e0d\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:33:48.019128 kubelet[2610]: I0302 13:33:48.019087 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:33:48.039651 kubelet[2610]: E0302 13:33:48.026410 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": dial tcp 10.0.0.31:6443: connect: connection refused" node="localhost" Mar 2 13:33:48.183744 systemd[1]: Created slice kubepods-burstable-pod08b2d708c40b0a38b3e0e2d2f01c0e0d.slice - libcontainer container kubepods-burstable-pod08b2d708c40b0a38b3e0e2d2f01c0e0d.slice. Mar 2 13:33:48.232992 kubelet[2610]: I0302 13:33:48.210050 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:33:48.232992 kubelet[2610]: I0302 13:33:48.222731 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:33:48.232992 kubelet[2610]: I0302 13:33:48.222859 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:33:48.232992 kubelet[2610]: I0302 13:33:48.222884 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:33:48.232992 kubelet[2610]: I0302 13:33:48.222906 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:33:48.234327 kubelet[2610]: I0302 13:33:48.223008 2610 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/89efda49e166906783d8d868d41ebb86-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"89efda49e166906783d8d868d41ebb86\") " pod="kube-system/kube-scheduler-localhost" Mar 2 13:33:48.245424 kubelet[2610]: E0302 13:33:48.244840 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:48.261768 kubelet[2610]: E0302 13:33:48.259138 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:48.263215 containerd[1614]: time="2026-03-02T13:33:48.262693049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:08b2d708c40b0a38b3e0e2d2f01c0e0d,Namespace:kube-system,Attempt:0,}" Mar 2 13:33:48.269286 systemd[1]: Created slice kubepods-burstable-poddb0989cdb653dfec284dd4f35625e9e7.slice - libcontainer container kubepods-burstable-poddb0989cdb653dfec284dd4f35625e9e7.slice. Mar 2 13:33:48.310850 kubelet[2610]: E0302 13:33:48.307915 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:48.326342 systemd[1]: Created slice kubepods-burstable-pod89efda49e166906783d8d868d41ebb86.slice - libcontainer container kubepods-burstable-pod89efda49e166906783d8d868d41ebb86.slice. Mar 2 13:33:48.352160 kubelet[2610]: E0302 13:33:48.352124 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:48.431854 kubelet[2610]: I0302 13:33:48.430802 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:33:48.437997 kubelet[2610]: E0302 13:33:48.437811 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": dial tcp 10.0.0.31:6443: connect: connection refused" node="localhost" Mar 2 13:33:48.522863 kubelet[2610]: E0302 13:33:48.521452 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:48.533192 containerd[1614]: time="2026-03-02T13:33:48.524452656Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:89efda49e166906783d8d868d41ebb86,Namespace:kube-system,Attempt:0,}" Mar 2 13:33:48.635932 kubelet[2610]: E0302 13:33:48.632989 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.31:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:33:48.647049 kubelet[2610]: E0302 13:33:48.646134 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:48.647416 containerd[1614]: time="2026-03-02T13:33:48.647351297Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:db0989cdb653dfec284dd4f35625e9e7,Namespace:kube-system,Attempt:0,}" Mar 2 13:33:49.048776 kubelet[2610]: E0302 13:33:49.046899 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.31:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:33:49.224202 kubelet[2610]: E0302 13:33:49.216431 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.31:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:33:49.262933 kubelet[2610]: I0302 13:33:49.261857 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:33:49.362864 kubelet[2610]: E0302 13:33:49.362364 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": dial tcp 10.0.0.31:6443: connect: connection refused" node="localhost" Mar 2 13:33:49.382399 kubelet[2610]: E0302 13:33:49.382056 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.31:6443: connect: connection refused" interval="3.2s" Mar 2 13:33:49.651845 kubelet[2610]: E0302 13:33:49.646186 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.31:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:33:50.338314 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3558741089.mount: Deactivated successfully. Mar 2 13:33:50.480731 containerd[1614]: time="2026-03-02T13:33:50.478526843Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:33:50.511209 containerd[1614]: time="2026-03-02T13:33:50.511072008Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=501" Mar 2 13:33:50.535508 containerd[1614]: time="2026-03-02T13:33:50.531877377Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:33:50.577104 containerd[1614]: time="2026-03-02T13:33:50.559891265Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:33:50.603228 containerd[1614]: time="2026-03-02T13:33:50.600083424Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 2 13:33:50.649006 containerd[1614]: time="2026-03-02T13:33:50.645812535Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:33:50.659052 containerd[1614]: time="2026-03-02T13:33:50.658963799Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:33:50.675155 containerd[1614]: time="2026-03-02T13:33:50.666464456Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 2.353021071s" Mar 2 13:33:50.685140 containerd[1614]: time="2026-03-02T13:33:50.685082529Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Mar 2 13:33:50.692776 containerd[1614]: time="2026-03-02T13:33:50.686300191Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 2.140906161s" Mar 2 13:33:50.710971 containerd[1614]: time="2026-03-02T13:33:50.707859810Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 2.016107249s" Mar 2 13:33:51.016800 containerd[1614]: time="2026-03-02T13:33:51.008366341Z" level=info msg="connecting to shim 99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8" address="unix:///run/containerd/s/3de32a7758329b3432bea8ac110bcc2ad4155ced6515008f6447fa859bd8e7bc" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:33:51.088914 kubelet[2610]: I0302 13:33:51.082848 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:33:51.088914 kubelet[2610]: E0302 13:33:51.084461 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": dial tcp 10.0.0.31:6443: connect: connection refused" node="localhost" Mar 2 13:33:51.099970 containerd[1614]: time="2026-03-02T13:33:51.099912536Z" level=info msg="connecting to shim 1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5" address="unix:///run/containerd/s/c52b81a14e47b43e9a85063dae38cd7161e4b4f17d376a4b5648542c95089668" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:33:51.182036 containerd[1614]: time="2026-03-02T13:33:51.181979670Z" level=info msg="connecting to shim 70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc" address="unix:///run/containerd/s/da5238a45576cb9842e131021ac2d307a0981c1e5236d4786a5972949dda1f50" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:33:51.457505 systemd[1]: Started cri-containerd-99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8.scope - libcontainer container 99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8. Mar 2 13:33:51.552369 kubelet[2610]: E0302 13:33:51.535924 2610 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.31:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:33:51.561221 systemd[1]: Started cri-containerd-1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5.scope - libcontainer container 1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5. Mar 2 13:33:51.633092 systemd[1]: Started cri-containerd-70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc.scope - libcontainer container 70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc. Mar 2 13:33:51.814294 kubelet[2610]: E0302 13:33:51.808043 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.31:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:33:52.231027 containerd[1614]: time="2026-03-02T13:33:52.230953604Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:08b2d708c40b0a38b3e0e2d2f01c0e0d,Namespace:kube-system,Attempt:0,} returns sandbox id \"1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5\"" Mar 2 13:33:52.261963 kubelet[2610]: E0302 13:33:52.261269 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:52.439011 containerd[1614]: time="2026-03-02T13:33:52.438107475Z" level=info msg="CreateContainer within sandbox \"1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 2 13:33:52.442933 containerd[1614]: time="2026-03-02T13:33:52.442417883Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:db0989cdb653dfec284dd4f35625e9e7,Namespace:kube-system,Attempt:0,} returns sandbox id \"70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc\"" Mar 2 13:33:52.448000 kubelet[2610]: E0302 13:33:52.447462 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:52.473217 containerd[1614]: time="2026-03-02T13:33:52.470196657Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:89efda49e166906783d8d868d41ebb86,Namespace:kube-system,Attempt:0,} returns sandbox id \"99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8\"" Mar 2 13:33:52.488191 kubelet[2610]: E0302 13:33:52.480061 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:52.513858 containerd[1614]: time="2026-03-02T13:33:52.512452777Z" level=info msg="CreateContainer within sandbox \"70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 2 13:33:52.561830 containerd[1614]: time="2026-03-02T13:33:52.550025013Z" level=info msg="CreateContainer within sandbox \"99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 2 13:33:52.590987 kubelet[2610]: E0302 13:33:52.586501 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.31:6443: connect: connection refused" interval="6.4s" Mar 2 13:33:52.588133 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount950778631.mount: Deactivated successfully. Mar 2 13:33:52.630095 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1390143028.mount: Deactivated successfully. Mar 2 13:33:52.634438 containerd[1614]: time="2026-03-02T13:33:52.634307257Z" level=info msg="Container 5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:33:52.744998 containerd[1614]: time="2026-03-02T13:33:52.739060365Z" level=info msg="Container fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:33:52.749229 containerd[1614]: time="2026-03-02T13:33:52.749046140Z" level=info msg="Container 886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:33:52.760368 containerd[1614]: time="2026-03-02T13:33:52.757346552Z" level=info msg="CreateContainer within sandbox \"1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f\"" Mar 2 13:33:52.760368 containerd[1614]: time="2026-03-02T13:33:52.758473433Z" level=info msg="StartContainer for \"5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f\"" Mar 2 13:33:52.771063 containerd[1614]: time="2026-03-02T13:33:52.770142856Z" level=info msg="connecting to shim 5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f" address="unix:///run/containerd/s/c52b81a14e47b43e9a85063dae38cd7161e4b4f17d376a4b5648542c95089668" protocol=ttrpc version=3 Mar 2 13:33:52.810841 containerd[1614]: time="2026-03-02T13:33:52.810325917Z" level=info msg="CreateContainer within sandbox \"70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c\"" Mar 2 13:33:52.828300 containerd[1614]: time="2026-03-02T13:33:52.827987558Z" level=info msg="StartContainer for \"fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c\"" Mar 2 13:33:52.834290 containerd[1614]: time="2026-03-02T13:33:52.831528403Z" level=info msg="connecting to shim fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c" address="unix:///run/containerd/s/da5238a45576cb9842e131021ac2d307a0981c1e5236d4786a5972949dda1f50" protocol=ttrpc version=3 Mar 2 13:33:52.852250 containerd[1614]: time="2026-03-02T13:33:52.852202262Z" level=info msg="CreateContainer within sandbox \"99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623\"" Mar 2 13:33:52.854303 containerd[1614]: time="2026-03-02T13:33:52.854266598Z" level=info msg="StartContainer for \"886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623\"" Mar 2 13:33:52.856535 containerd[1614]: time="2026-03-02T13:33:52.856501905Z" level=info msg="connecting to shim 886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623" address="unix:///run/containerd/s/3de32a7758329b3432bea8ac110bcc2ad4155ced6515008f6447fa859bd8e7bc" protocol=ttrpc version=3 Mar 2 13:33:52.897975 systemd[1]: Started cri-containerd-5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f.scope - libcontainer container 5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f. Mar 2 13:33:52.985048 systemd[1]: Started cri-containerd-fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c.scope - libcontainer container fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c. Mar 2 13:33:53.124956 systemd[1]: Started cri-containerd-886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623.scope - libcontainer container 886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623. Mar 2 13:33:53.841104 kubelet[2610]: E0302 13:33:53.840125 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.31:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:33:53.856416 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3814798614.mount: Deactivated successfully. Mar 2 13:33:53.905940 containerd[1614]: time="2026-03-02T13:33:53.905392727Z" level=info msg="StartContainer for \"5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f\" returns successfully" Mar 2 13:33:54.162852 kubelet[2610]: E0302 13:33:54.161490 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.31:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:33:54.186246 containerd[1614]: time="2026-03-02T13:33:54.186037197Z" level=info msg="StartContainer for \"fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c\" returns successfully" Mar 2 13:33:54.231191 containerd[1614]: time="2026-03-02T13:33:54.229856118Z" level=info msg="StartContainer for \"886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623\" returns successfully" Mar 2 13:33:54.309839 kubelet[2610]: I0302 13:33:54.301967 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:33:54.309839 kubelet[2610]: E0302 13:33:54.302468 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": dial tcp 10.0.0.31:6443: connect: connection refused" node="localhost" Mar 2 13:33:55.153074 kubelet[2610]: E0302 13:33:55.152841 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:55.153074 kubelet[2610]: E0302 13:33:55.153047 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:55.194055 kubelet[2610]: E0302 13:33:55.193121 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:55.194055 kubelet[2610]: E0302 13:33:55.193305 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:55.210876 kubelet[2610]: E0302 13:33:55.209103 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:55.210876 kubelet[2610]: E0302 13:33:55.209345 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:55.290639 kubelet[2610]: E0302 13:33:55.283782 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.31:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.31:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:33:56.224062 kubelet[2610]: E0302 13:33:56.220318 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:56.233475 kubelet[2610]: E0302 13:33:56.230341 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:56.257887 kubelet[2610]: E0302 13:33:56.257440 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:56.263453 kubelet[2610]: E0302 13:33:56.262229 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:56.263453 kubelet[2610]: E0302 13:33:56.263088 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:56.270054 kubelet[2610]: E0302 13:33:56.270023 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:57.243943 kubelet[2610]: E0302 13:33:57.243406 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:57.258252 kubelet[2610]: E0302 13:33:57.249523 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:57.277213 kubelet[2610]: E0302 13:33:57.276307 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:33:57.286431 kubelet[2610]: E0302 13:33:57.286068 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:33:57.468327 kubelet[2610]: E0302 13:33:57.468156 2610 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:34:01.061119 kubelet[2610]: I0302 13:34:01.048029 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:34:01.096518 kubelet[2610]: E0302 13:34:01.075463 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:34:01.096518 kubelet[2610]: E0302 13:34:01.086314 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:03.499333 kubelet[2610]: E0302 13:34:03.498334 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:34:03.499333 kubelet[2610]: E0302 13:34:03.501390 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:04.250327 kubelet[2610]: E0302 13:34:04.250267 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:34:04.296363 kubelet[2610]: E0302 13:34:04.295001 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:05.449464 kubelet[2610]: E0302 13:34:05.447969 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:34:05.449464 kubelet[2610]: E0302 13:34:05.448331 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:05.449464 kubelet[2610]: E0302 13:34:05.447520 2610 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.31:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.1899098c2d77154f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,LastTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 13:34:06.620982 kubelet[2610]: E0302 13:34:06.620079 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.31:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:34:07.517120 kubelet[2610]: E0302 13:34:07.516285 2610 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:34:09.010768 kubelet[2610]: E0302 13:34:09.009429 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: TLS handshake timeout" interval="7s" Mar 2 13:34:10.548479 kubelet[2610]: E0302 13:34:10.547235 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.31:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:34:11.082483 kubelet[2610]: E0302 13:34:10.935858 2610 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.31:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:34:11.283462 kubelet[2610]: E0302 13:34:11.202353 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Mar 2 13:34:14.329999 kubelet[2610]: E0302 13:34:14.329197 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.31:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:34:15.341055 kubelet[2610]: E0302 13:34:15.197532 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.31:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:34:17.592345 kubelet[2610]: E0302 13:34:17.565425 2610 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:34:18.519067 kubelet[2610]: I0302 13:34:18.488386 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:34:23.500144 kubelet[2610]: E0302 13:34:23.498074 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.31:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:34:25.456966 kubelet[2610]: E0302 13:34:25.455253 2610 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.31:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.1899098c2d77154f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,LastTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 13:34:25.546946 kubelet[2610]: E0302 13:34:25.543901 2610 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:34:25.546946 kubelet[2610]: E0302 13:34:25.544173 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:26.939493 kubelet[2610]: E0302 13:34:26.655312 2610 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="7s" Mar 2 13:34:27.904034 kubelet[2610]: E0302 13:34:27.822969 2610 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:34:28.526964 kubelet[2610]: E0302 13:34:28.526073 2610 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.31:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Mar 2 13:34:36.403055 kubelet[2610]: I0302 13:34:36.402926 2610 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:34:37.203256 kubelet[2610]: E0302 13:34:37.203170 2610 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.31:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:34:37.234255 kubelet[2610]: E0302 13:34:37.223068 2610 certificate_manager.go:461] "Reached backoff limit, still unable to rotate certs" err="timed out waiting for the condition" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:34:37.902742 kubelet[2610]: E0302 13:34:37.902075 2610 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:34:38.027443 kubelet[2610]: E0302 13:34:38.027291 2610 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.31:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:34:39.648436 kubelet[2610]: I0302 13:34:39.646296 2610 apiserver.go:52] "Watching apiserver" Mar 2 13:34:39.786965 kubelet[2610]: E0302 13:34:39.784181 2610 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 2 13:34:39.813023 kubelet[2610]: I0302 13:34:39.807807 2610 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 2 13:34:40.358389 kubelet[2610]: I0302 13:34:40.358116 2610 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 2 13:34:40.464239 kubelet[2610]: I0302 13:34:40.433295 2610 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 13:34:41.448925 kubelet[2610]: E0302 13:34:41.447245 2610 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.1899098c2d77154f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,LastTimestamp:2026-03-02 13:33:46.009851215 +0000 UTC m=+9.518588448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 13:34:41.726879 kubelet[2610]: I0302 13:34:41.725457 2610 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 13:34:41.728877 kubelet[2610]: E0302 13:34:41.727433 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:42.217146 kubelet[2610]: E0302 13:34:42.213456 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:42.217146 kubelet[2610]: I0302 13:34:42.214023 2610 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 13:34:43.029931 kubelet[2610]: E0302 13:34:43.015187 2610 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:34:46.591037 systemd[1692]: Created slice background.slice - User Background Tasks Slice. Mar 2 13:34:46.618389 systemd[1692]: Starting systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories... Mar 2 13:34:48.109990 systemd[1692]: Finished systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories. Mar 2 13:34:48.753932 kubelet[2610]: I0302 13:34:48.750706 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=7.750420517 podStartE2EDuration="7.750420517s" podCreationTimestamp="2026-03-02 13:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:34:48.568162401 +0000 UTC m=+72.076899634" watchObservedRunningTime="2026-03-02 13:34:48.750420517 +0000 UTC m=+72.259157750" Mar 2 13:34:48.753932 kubelet[2610]: I0302 13:34:48.750970 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=7.750954532 podStartE2EDuration="7.750954532s" podCreationTimestamp="2026-03-02 13:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:34:48.723470926 +0000 UTC m=+72.232208159" watchObservedRunningTime="2026-03-02 13:34:48.750954532 +0000 UTC m=+72.259691765" Mar 2 13:35:08.316258 systemd[1]: Reload requested from client PID 2909 ('systemctl') (unit session-8.scope)... Mar 2 13:35:08.316292 systemd[1]: Reloading... Mar 2 13:35:09.272865 zram_generator::config[2961]: No configuration found. Mar 2 13:35:10.566060 systemd[1]: Reloading finished in 2245 ms. Mar 2 13:35:10.734744 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:35:10.834441 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 13:35:10.835136 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:35:10.836966 systemd[1]: kubelet.service: Consumed 15.095s CPU time, 135M memory peak. Mar 2 13:35:10.852875 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:35:12.623050 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:35:12.715754 (kubelet)[3001]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 13:35:13.307939 kubelet[3001]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 13:35:13.307939 kubelet[3001]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 13:35:13.308888 kubelet[3001]: I0302 13:35:13.307920 3001 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 13:35:13.388879 kubelet[3001]: I0302 13:35:13.387130 3001 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 2 13:35:13.388879 kubelet[3001]: I0302 13:35:13.387338 3001 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 13:35:13.388879 kubelet[3001]: I0302 13:35:13.387384 3001 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 2 13:35:13.388879 kubelet[3001]: I0302 13:35:13.387401 3001 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 13:35:13.396743 kubelet[3001]: I0302 13:35:13.389174 3001 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 13:35:13.414480 kubelet[3001]: I0302 13:35:13.411952 3001 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 2 13:35:13.469917 kubelet[3001]: I0302 13:35:13.463459 3001 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 13:35:13.554124 kubelet[3001]: I0302 13:35:13.553968 3001 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 13:35:13.597179 kubelet[3001]: I0302 13:35:13.596924 3001 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 2 13:35:13.601133 kubelet[3001]: I0302 13:35:13.599223 3001 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 13:35:13.601133 kubelet[3001]: I0302 13:35:13.599420 3001 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 13:35:13.601133 kubelet[3001]: I0302 13:35:13.599758 3001 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 13:35:13.601133 kubelet[3001]: I0302 13:35:13.599772 3001 container_manager_linux.go:306] "Creating device plugin manager" Mar 2 13:35:13.601844 kubelet[3001]: I0302 13:35:13.599801 3001 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 2 13:35:13.601844 kubelet[3001]: I0302 13:35:13.600026 3001 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:35:13.602982 kubelet[3001]: I0302 13:35:13.602804 3001 kubelet.go:475] "Attempting to sync node with API server" Mar 2 13:35:13.603055 kubelet[3001]: I0302 13:35:13.602999 3001 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 13:35:13.603055 kubelet[3001]: I0302 13:35:13.603034 3001 kubelet.go:387] "Adding apiserver pod source" Mar 2 13:35:13.603055 kubelet[3001]: I0302 13:35:13.603053 3001 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 13:35:13.619823 kubelet[3001]: I0302 13:35:13.616989 3001 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Mar 2 13:35:13.644899 kubelet[3001]: I0302 13:35:13.641690 3001 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 13:35:13.644899 kubelet[3001]: I0302 13:35:13.641742 3001 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 2 13:35:13.707909 kubelet[3001]: I0302 13:35:13.706754 3001 server.go:1262] "Started kubelet" Mar 2 13:35:13.735198 kubelet[3001]: I0302 13:35:13.710128 3001 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 13:35:13.735198 kubelet[3001]: I0302 13:35:13.715969 3001 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 2 13:35:13.735198 kubelet[3001]: I0302 13:35:13.717723 3001 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 13:35:13.735198 kubelet[3001]: I0302 13:35:13.731939 3001 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 13:35:13.739745 kubelet[3001]: I0302 13:35:13.735810 3001 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 13:35:13.739745 kubelet[3001]: I0302 13:35:13.736154 3001 server.go:310] "Adding debug handlers to kubelet server" Mar 2 13:35:13.797133 kubelet[3001]: I0302 13:35:13.793933 3001 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 13:35:13.806372 kubelet[3001]: I0302 13:35:13.806068 3001 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 2 13:35:13.818805 kubelet[3001]: I0302 13:35:13.816427 3001 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 2 13:35:13.818805 kubelet[3001]: I0302 13:35:13.817902 3001 reconciler.go:29] "Reconciler: start to sync state" Mar 2 13:35:13.833974 kubelet[3001]: E0302 13:35:13.830085 3001 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 13:35:13.886117 kubelet[3001]: I0302 13:35:13.873744 3001 factory.go:223] Registration of the systemd container factory successfully Mar 2 13:35:13.886117 kubelet[3001]: I0302 13:35:13.873960 3001 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 13:35:13.896122 kubelet[3001]: I0302 13:35:13.895407 3001 factory.go:223] Registration of the containerd container factory successfully Mar 2 13:35:14.104793 kubelet[3001]: I0302 13:35:14.104473 3001 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 2 13:35:14.158878 kubelet[3001]: I0302 13:35:14.148028 3001 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 2 13:35:14.158878 kubelet[3001]: I0302 13:35:14.148068 3001 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 2 13:35:14.158878 kubelet[3001]: I0302 13:35:14.148102 3001 kubelet.go:2428] "Starting kubelet main sync loop" Mar 2 13:35:14.158878 kubelet[3001]: E0302 13:35:14.148171 3001 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 13:35:14.255395 kubelet[3001]: E0302 13:35:14.252089 3001 kubelet.go:2452] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 2 13:35:14.291799 kubelet[3001]: I0302 13:35:14.291115 3001 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 13:35:14.291799 kubelet[3001]: I0302 13:35:14.291221 3001 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 13:35:14.291799 kubelet[3001]: I0302 13:35:14.291254 3001 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:35:14.304642 kubelet[3001]: I0302 13:35:14.304153 3001 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 2 13:35:14.304642 kubelet[3001]: I0302 13:35:14.304377 3001 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 2 13:35:14.304642 kubelet[3001]: I0302 13:35:14.304408 3001 policy_none.go:49] "None policy: Start" Mar 2 13:35:14.304642 kubelet[3001]: I0302 13:35:14.304427 3001 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 2 13:35:14.304642 kubelet[3001]: I0302 13:35:14.304447 3001 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 2 13:35:14.322454 kubelet[3001]: I0302 13:35:14.320855 3001 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Mar 2 13:35:14.322454 kubelet[3001]: I0302 13:35:14.320979 3001 policy_none.go:47] "Start" Mar 2 13:35:14.404470 kubelet[3001]: E0302 13:35:14.400133 3001 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 13:35:14.404470 kubelet[3001]: I0302 13:35:14.402527 3001 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 13:35:14.404470 kubelet[3001]: I0302 13:35:14.402819 3001 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 13:35:14.404470 kubelet[3001]: I0302 13:35:14.404102 3001 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 13:35:14.483006 kubelet[3001]: I0302 13:35:14.470136 3001 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 13:35:14.483006 kubelet[3001]: I0302 13:35:14.474097 3001 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.483006 kubelet[3001]: I0302 13:35:14.478925 3001 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 13:35:14.489849 kubelet[3001]: I0302 13:35:14.486853 3001 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 2 13:35:14.499122 kubelet[3001]: E0302 13:35:14.491407 3001 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 13:35:14.529022 containerd[1614]: time="2026-03-02T13:35:14.522137174Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 2 13:35:14.529908 kubelet[3001]: I0302 13:35:14.523456 3001 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 2 13:35:14.581821 kubelet[3001]: I0302 13:35:14.580427 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.581821 kubelet[3001]: I0302 13:35:14.580831 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.581821 kubelet[3001]: I0302 13:35:14.580871 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/89efda49e166906783d8d868d41ebb86-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"89efda49e166906783d8d868d41ebb86\") " pod="kube-system/kube-scheduler-localhost" Mar 2 13:35:14.581821 kubelet[3001]: I0302 13:35:14.580896 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/08b2d708c40b0a38b3e0e2d2f01c0e0d-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"08b2d708c40b0a38b3e0e2d2f01c0e0d\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:35:14.581821 kubelet[3001]: I0302 13:35:14.580921 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/08b2d708c40b0a38b3e0e2d2f01c0e0d-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"08b2d708c40b0a38b3e0e2d2f01c0e0d\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:35:14.582244 kubelet[3001]: I0302 13:35:14.580942 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/08b2d708c40b0a38b3e0e2d2f01c0e0d-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"08b2d708c40b0a38b3e0e2d2f01c0e0d\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:35:14.582244 kubelet[3001]: I0302 13:35:14.580962 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.582244 kubelet[3001]: I0302 13:35:14.580987 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.582244 kubelet[3001]: I0302 13:35:14.581009 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/db0989cdb653dfec284dd4f35625e9e7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"db0989cdb653dfec284dd4f35625e9e7\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.601231 kubelet[3001]: E0302 13:35:14.597218 3001 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 2 13:35:14.601231 kubelet[3001]: E0302 13:35:14.597435 3001 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Mar 2 13:35:14.601231 kubelet[3001]: E0302 13:35:14.597724 3001 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Mar 2 13:35:14.620484 kubelet[3001]: I0302 13:35:14.619924 3001 apiserver.go:52] "Watching apiserver" Mar 2 13:35:14.719788 kubelet[3001]: I0302 13:35:14.717020 3001 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 2 13:35:14.763877 kubelet[3001]: I0302 13:35:14.755994 3001 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:35:14.858854 kubelet[3001]: I0302 13:35:14.858395 3001 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Mar 2 13:35:14.866081 kubelet[3001]: I0302 13:35:14.865797 3001 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 2 13:35:14.900995 kubelet[3001]: E0302 13:35:14.897972 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:14.900995 kubelet[3001]: E0302 13:35:14.898507 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:14.911164 kubelet[3001]: E0302 13:35:14.910910 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:15.266037 kubelet[3001]: E0302 13:35:15.263056 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:15.268042 kubelet[3001]: E0302 13:35:15.267109 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:15.274736 kubelet[3001]: E0302 13:35:15.268979 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:16.372208 kubelet[3001]: E0302 13:35:16.364018 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:16.766492 kubelet[3001]: E0302 13:35:16.749740 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:18.004520 kubelet[3001]: E0302 13:35:18.003983 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:19.850915 kubelet[3001]: E0302 13:35:19.833187 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:21.155959 kubelet[3001]: E0302 13:35:21.152866 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:21.703517 systemd[1]: Created slice kubepods-besteffort-pod42ea34f5_9530_44fe_aba8_7a7baa57399a.slice - libcontainer container kubepods-besteffort-pod42ea34f5_9530_44fe_aba8_7a7baa57399a.slice. Mar 2 13:35:22.723755 kubelet[3001]: I0302 13:35:22.723125 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/42ea34f5-9530-44fe-aba8-7a7baa57399a-xtables-lock\") pod \"kube-proxy-d4dng\" (UID: \"42ea34f5-9530-44fe-aba8-7a7baa57399a\") " pod="kube-system/kube-proxy-d4dng" Mar 2 13:35:22.908265 kubelet[3001]: I0302 13:35:22.747928 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbmm9\" (UniqueName: \"kubernetes.io/projected/42ea34f5-9530-44fe-aba8-7a7baa57399a-kube-api-access-vbmm9\") pod \"kube-proxy-d4dng\" (UID: \"42ea34f5-9530-44fe-aba8-7a7baa57399a\") " pod="kube-system/kube-proxy-d4dng" Mar 2 13:35:22.908265 kubelet[3001]: I0302 13:35:22.911291 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/42ea34f5-9530-44fe-aba8-7a7baa57399a-kube-proxy\") pod \"kube-proxy-d4dng\" (UID: \"42ea34f5-9530-44fe-aba8-7a7baa57399a\") " pod="kube-system/kube-proxy-d4dng" Mar 2 13:35:22.908265 kubelet[3001]: I0302 13:35:22.911778 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/42ea34f5-9530-44fe-aba8-7a7baa57399a-lib-modules\") pod \"kube-proxy-d4dng\" (UID: \"42ea34f5-9530-44fe-aba8-7a7baa57399a\") " pod="kube-system/kube-proxy-d4dng" Mar 2 13:35:24.750113 kubelet[3001]: E0302 13:35:24.738756 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:24.806380 kubelet[3001]: E0302 13:35:24.802112 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:24.859082 containerd[1614]: time="2026-03-02T13:35:24.849422281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-d4dng,Uid:42ea34f5-9530-44fe-aba8-7a7baa57399a,Namespace:kube-system,Attempt:0,}" Mar 2 13:35:25.701973 containerd[1614]: time="2026-03-02T13:35:25.701410015Z" level=info msg="connecting to shim 0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae" address="unix:///run/containerd/s/77aa43df85ef072f3e1134dcace991a732a39d85d5844ae74f8cc9e4c17b1870" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:35:26.352734 kubelet[3001]: E0302 13:35:26.336481 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:27.721750 kubelet[3001]: E0302 13:35:27.717776 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:28.210347 systemd[1]: Started cri-containerd-0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae.scope - libcontainer container 0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae. Mar 2 13:35:30.696470 containerd[1614]: time="2026-03-02T13:35:30.696156198Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-d4dng,Uid:42ea34f5-9530-44fe-aba8-7a7baa57399a,Namespace:kube-system,Attempt:0,} returns sandbox id \"0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae\"" Mar 2 13:35:30.705432 kubelet[3001]: E0302 13:35:30.705389 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:30.813922 kubelet[3001]: I0302 13:35:30.812267 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/31fa5815-be59-48c4-ab50-a26efe852a61-cni-plugin\") pod \"kube-flannel-ds-jlpj6\" (UID: \"31fa5815-be59-48c4-ab50-a26efe852a61\") " pod="kube-flannel/kube-flannel-ds-jlpj6" Mar 2 13:35:30.813922 kubelet[3001]: I0302 13:35:30.813059 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/31fa5815-be59-48c4-ab50-a26efe852a61-cni\") pod \"kube-flannel-ds-jlpj6\" (UID: \"31fa5815-be59-48c4-ab50-a26efe852a61\") " pod="kube-flannel/kube-flannel-ds-jlpj6" Mar 2 13:35:30.813922 kubelet[3001]: I0302 13:35:30.813174 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97sks\" (UniqueName: \"kubernetes.io/projected/31fa5815-be59-48c4-ab50-a26efe852a61-kube-api-access-97sks\") pod \"kube-flannel-ds-jlpj6\" (UID: \"31fa5815-be59-48c4-ab50-a26efe852a61\") " pod="kube-flannel/kube-flannel-ds-jlpj6" Mar 2 13:35:30.813922 kubelet[3001]: I0302 13:35:30.813420 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/31fa5815-be59-48c4-ab50-a26efe852a61-run\") pod \"kube-flannel-ds-jlpj6\" (UID: \"31fa5815-be59-48c4-ab50-a26efe852a61\") " pod="kube-flannel/kube-flannel-ds-jlpj6" Mar 2 13:35:30.859945 systemd[1]: Created slice kubepods-burstable-pod31fa5815_be59_48c4_ab50_a26efe852a61.slice - libcontainer container kubepods-burstable-pod31fa5815_be59_48c4_ab50_a26efe852a61.slice. Mar 2 13:35:30.876880 kubelet[3001]: I0302 13:35:30.813454 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/31fa5815-be59-48c4-ab50-a26efe852a61-flannel-cfg\") pod \"kube-flannel-ds-jlpj6\" (UID: \"31fa5815-be59-48c4-ab50-a26efe852a61\") " pod="kube-flannel/kube-flannel-ds-jlpj6" Mar 2 13:35:30.932192 kubelet[3001]: I0302 13:35:30.930389 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/31fa5815-be59-48c4-ab50-a26efe852a61-xtables-lock\") pod \"kube-flannel-ds-jlpj6\" (UID: \"31fa5815-be59-48c4-ab50-a26efe852a61\") " pod="kube-flannel/kube-flannel-ds-jlpj6" Mar 2 13:35:31.667850 containerd[1614]: time="2026-03-02T13:35:31.652497498Z" level=info msg="CreateContainer within sandbox \"0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 2 13:35:32.365949 containerd[1614]: time="2026-03-02T13:35:32.365211114Z" level=info msg="Container 70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:35:32.375275 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3603861300.mount: Deactivated successfully. Mar 2 13:35:32.468923 kubelet[3001]: E0302 13:35:32.465351 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:32.494426 containerd[1614]: time="2026-03-02T13:35:32.478327929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-jlpj6,Uid:31fa5815-be59-48c4-ab50-a26efe852a61,Namespace:kube-flannel,Attempt:0,}" Mar 2 13:35:32.895525 containerd[1614]: time="2026-03-02T13:35:32.893456411Z" level=info msg="CreateContainer within sandbox \"0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27\"" Mar 2 13:35:32.915873 containerd[1614]: time="2026-03-02T13:35:32.915168406Z" level=info msg="StartContainer for \"70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27\"" Mar 2 13:35:32.947478 containerd[1614]: time="2026-03-02T13:35:32.947415951Z" level=info msg="connecting to shim 70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27" address="unix:///run/containerd/s/77aa43df85ef072f3e1134dcace991a732a39d85d5844ae74f8cc9e4c17b1870" protocol=ttrpc version=3 Mar 2 13:35:33.266139 sudo[1796]: pam_unix(sudo:session): session closed for user root Mar 2 13:35:33.283825 sshd[1795]: Connection closed by 10.0.0.1 port 58202 Mar 2 13:35:33.375247 sshd-session[1791]: pam_unix(sshd:session): session closed for user core Mar 2 13:35:33.473205 systemd[1]: sshd@6-10.0.0.31:22-10.0.0.1:58202.service: Deactivated successfully. Mar 2 13:35:33.480526 containerd[1614]: time="2026-03-02T13:35:33.476134991Z" level=info msg="connecting to shim 12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876" address="unix:///run/containerd/s/3388e0d9174e8cda076272eba35e46c5041113f5157f5f3d8de6c58cbe370b43" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:35:33.520451 systemd[1]: session-8.scope: Deactivated successfully. Mar 2 13:35:33.521157 systemd[1]: session-8.scope: Consumed 22.382s CPU time, 239.6M memory peak. Mar 2 13:35:33.547427 systemd-logind[1587]: Session 8 logged out. Waiting for processes to exit. Mar 2 13:35:33.549956 systemd-logind[1587]: Removed session 8. Mar 2 13:35:33.711436 systemd[1]: Started cri-containerd-70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27.scope - libcontainer container 70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27. Mar 2 13:35:34.332243 systemd[1]: Started cri-containerd-12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876.scope - libcontainer container 12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876. Mar 2 13:35:34.917160 containerd[1614]: time="2026-03-02T13:35:34.911181813Z" level=info msg="StartContainer for \"70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27\" returns successfully" Mar 2 13:35:35.523122 containerd[1614]: time="2026-03-02T13:35:35.522368643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-jlpj6,Uid:31fa5815-be59-48c4-ab50-a26efe852a61,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\"" Mar 2 13:35:35.528311 kubelet[3001]: E0302 13:35:35.526017 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:35.531899 containerd[1614]: time="2026-03-02T13:35:35.531770721Z" level=info msg="PullImage \"ghcr.io/flannel-io/flannel-cni-plugin:v1.6.2-flannel1\"" Mar 2 13:35:35.599909 kubelet[3001]: E0302 13:35:35.598439 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:36.689497 kubelet[3001]: E0302 13:35:36.689063 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:41.258095 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2629548276.mount: Deactivated successfully. Mar 2 13:35:41.786302 containerd[1614]: time="2026-03-02T13:35:41.783220989Z" level=info msg="ImageCreate event name:\"ghcr.io/flannel-io/flannel-cni-plugin:v1.6.2-flannel1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:35:41.801264 containerd[1614]: time="2026-03-02T13:35:41.799413248Z" level=info msg="stop pulling image ghcr.io/flannel-io/flannel-cni-plugin:v1.6.2-flannel1: active requests=0, bytes read=3641610" Mar 2 13:35:41.812092 containerd[1614]: time="2026-03-02T13:35:41.811484112Z" level=info msg="ImageCreate event name:\"sha256:55ce2385d9d8c6f720091c177fbe885a21c9dc07c9e480bfb4d94b3001f58182\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:35:41.864971 containerd[1614]: time="2026-03-02T13:35:41.863263839Z" level=info msg="ImageCreate event name:\"ghcr.io/flannel-io/flannel-cni-plugin@sha256:f1812994f0edbcb5bb5ccb63be2147ba6ad10e1faaa7ca9fcdad4f441739d84f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:35:41.878001 containerd[1614]: time="2026-03-02T13:35:41.869527090Z" level=info msg="Pulled image \"ghcr.io/flannel-io/flannel-cni-plugin:v1.6.2-flannel1\" with image id \"sha256:55ce2385d9d8c6f720091c177fbe885a21c9dc07c9e480bfb4d94b3001f58182\", repo tag \"ghcr.io/flannel-io/flannel-cni-plugin:v1.6.2-flannel1\", repo digest \"ghcr.io/flannel-io/flannel-cni-plugin@sha256:f1812994f0edbcb5bb5ccb63be2147ba6ad10e1faaa7ca9fcdad4f441739d84f\", size \"4856838\" in 6.337703982s" Mar 2 13:35:41.878001 containerd[1614]: time="2026-03-02T13:35:41.875275022Z" level=info msg="PullImage \"ghcr.io/flannel-io/flannel-cni-plugin:v1.6.2-flannel1\" returns image reference \"sha256:55ce2385d9d8c6f720091c177fbe885a21c9dc07c9e480bfb4d94b3001f58182\"" Mar 2 13:35:41.981105 containerd[1614]: time="2026-03-02T13:35:41.979107913Z" level=info msg="CreateContainer within sandbox \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Mar 2 13:35:42.193946 containerd[1614]: time="2026-03-02T13:35:42.182356859Z" level=info msg="Container d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:35:42.315223 containerd[1614]: time="2026-03-02T13:35:42.308436179Z" level=info msg="CreateContainer within sandbox \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff\"" Mar 2 13:35:42.316923 containerd[1614]: time="2026-03-02T13:35:42.316801724Z" level=info msg="StartContainer for \"d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff\"" Mar 2 13:35:42.332909 containerd[1614]: time="2026-03-02T13:35:42.332487088Z" level=info msg="connecting to shim d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff" address="unix:///run/containerd/s/3388e0d9174e8cda076272eba35e46c5041113f5157f5f3d8de6c58cbe370b43" protocol=ttrpc version=3 Mar 2 13:35:42.853249 systemd[1]: Started cri-containerd-d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff.scope - libcontainer container d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff. Mar 2 13:35:43.834113 systemd[1]: cri-containerd-d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff.scope: Deactivated successfully. Mar 2 13:35:43.943390 containerd[1614]: time="2026-03-02T13:35:43.943254347Z" level=info msg="received container exit event container_id:\"d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff\" id:\"d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff\" pid:3324 exited_at:{seconds:1772458543 nanos:937980537}" Mar 2 13:35:44.042288 containerd[1614]: time="2026-03-02T13:35:44.034699201Z" level=info msg="StartContainer for \"d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff\" returns successfully" Mar 2 13:35:44.733078 kubelet[3001]: E0302 13:35:44.726023 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:45.237448 kubelet[3001]: I0302 13:35:45.230118 3001 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-d4dng" podStartSLOduration=32.230096892 podStartE2EDuration="32.230096892s" podCreationTimestamp="2026-03-02 13:35:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:35:35.723976797 +0000 UTC m=+22.919569061" watchObservedRunningTime="2026-03-02 13:35:45.230096892 +0000 UTC m=+32.425689146" Mar 2 13:35:45.322997 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff-rootfs.mount: Deactivated successfully. Mar 2 13:35:46.307501 kubelet[3001]: E0302 13:35:46.301735 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:35:46.356123 containerd[1614]: time="2026-03-02T13:35:46.324700576Z" level=info msg="PullImage \"ghcr.io/flannel-io/flannel:v0.26.7\"" Mar 2 13:35:59.307055 kubelet[3001]: E0302 13:35:59.226953 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.042s" Mar 2 13:36:06.586966 kubelet[3001]: E0302 13:36:06.578022 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.429s" Mar 2 13:36:20.887860 containerd[1614]: time="2026-03-02T13:36:20.866984346Z" level=info msg="ImageCreate event name:\"ghcr.io/flannel-io/flannel:v0.26.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:36:20.906448 containerd[1614]: time="2026-03-02T13:36:20.901862679Z" level=info msg="stop pulling image ghcr.io/flannel-io/flannel:v0.26.7: active requests=0, bytes read=29341321" Mar 2 13:36:20.940871 containerd[1614]: time="2026-03-02T13:36:20.936269363Z" level=info msg="ImageCreate event name:\"sha256:965b9dd4aa4c1b6b68a4c54a166692b4645b6e6f8a5937d8dc17736cb63f515e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:36:21.111336 containerd[1614]: time="2026-03-02T13:36:21.106518419Z" level=info msg="ImageCreate event name:\"ghcr.io/flannel-io/flannel@sha256:7f471907fa940f944867270de4ed78121b8b4c5d564e17f940dc787cb16dea82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:36:21.146267 containerd[1614]: time="2026-03-02T13:36:21.143892175Z" level=info msg="Pulled image \"ghcr.io/flannel-io/flannel:v0.26.7\" with image id \"sha256:965b9dd4aa4c1b6b68a4c54a166692b4645b6e6f8a5937d8dc17736cb63f515e\", repo tag \"ghcr.io/flannel-io/flannel:v0.26.7\", repo digest \"ghcr.io/flannel-io/flannel@sha256:7f471907fa940f944867270de4ed78121b8b4c5d564e17f940dc787cb16dea82\", size \"32996046\" in 34.819142848s" Mar 2 13:36:21.146267 containerd[1614]: time="2026-03-02T13:36:21.144077815Z" level=info msg="PullImage \"ghcr.io/flannel-io/flannel:v0.26.7\" returns image reference \"sha256:965b9dd4aa4c1b6b68a4c54a166692b4645b6e6f8a5937d8dc17736cb63f515e\"" Mar 2 13:36:21.467164 containerd[1614]: time="2026-03-02T13:36:21.461098510Z" level=info msg="CreateContainer within sandbox \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 2 13:36:21.764886 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1507587147.mount: Deactivated successfully. Mar 2 13:36:21.796824 containerd[1614]: time="2026-03-02T13:36:21.793187997Z" level=info msg="Container 8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:36:21.924886 containerd[1614]: time="2026-03-02T13:36:21.924516936Z" level=info msg="CreateContainer within sandbox \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69\"" Mar 2 13:36:21.949258 containerd[1614]: time="2026-03-02T13:36:21.935036772Z" level=info msg="StartContainer for \"8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69\"" Mar 2 13:36:22.036799 containerd[1614]: time="2026-03-02T13:36:22.025278384Z" level=info msg="connecting to shim 8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69" address="unix:///run/containerd/s/3388e0d9174e8cda076272eba35e46c5041113f5157f5f3d8de6c58cbe370b43" protocol=ttrpc version=3 Mar 2 13:36:22.879234 systemd[1]: Started cri-containerd-8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69.scope - libcontainer container 8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69. Mar 2 13:36:24.830062 systemd[1]: cri-containerd-8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69.scope: Deactivated successfully. Mar 2 13:36:25.001424 kubelet[3001]: I0302 13:36:24.999337 3001 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Mar 2 13:36:25.154001 containerd[1614]: time="2026-03-02T13:36:25.112468072Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31fa5815_be59_48c4_ab50_a26efe852a61.slice/cri-containerd-8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69.scope/memory.events\": no such file or directory" Mar 2 13:36:25.210038 containerd[1614]: time="2026-03-02T13:36:25.208301714Z" level=info msg="received container exit event container_id:\"8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69\" id:\"8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69\" pid:3436 exited_at:{seconds:1772458585 nanos:111914277}" Mar 2 13:36:25.244774 containerd[1614]: time="2026-03-02T13:36:25.243254924Z" level=info msg="StartContainer for \"8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69\" returns successfully" Mar 2 13:36:26.282386 kubelet[3001]: E0302 13:36:26.282209 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:26.362703 kubelet[3001]: I0302 13:36:26.362098 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2pgg\" (UniqueName: \"kubernetes.io/projected/b6bb16d1-3aa0-40ca-9fd6-883920917fda-kube-api-access-c2pgg\") pod \"coredns-66bc5c9577-2kt5r\" (UID: \"b6bb16d1-3aa0-40ca-9fd6-883920917fda\") " pod="kube-system/coredns-66bc5c9577-2kt5r" Mar 2 13:36:26.362703 kubelet[3001]: I0302 13:36:26.362512 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6bb16d1-3aa0-40ca-9fd6-883920917fda-config-volume\") pod \"coredns-66bc5c9577-2kt5r\" (UID: \"b6bb16d1-3aa0-40ca-9fd6-883920917fda\") " pod="kube-system/coredns-66bc5c9577-2kt5r" Mar 2 13:36:26.391467 systemd[1]: Created slice kubepods-burstable-podb6bb16d1_3aa0_40ca_9fd6_883920917fda.slice - libcontainer container kubepods-burstable-podb6bb16d1_3aa0_40ca_9fd6_883920917fda.slice. Mar 2 13:36:26.471947 systemd[1]: Created slice kubepods-burstable-pod026e42c9_1716_4494_9461_8f8d83a0e4b6.slice - libcontainer container kubepods-burstable-pod026e42c9_1716_4494_9461_8f8d83a0e4b6.slice. Mar 2 13:36:26.478324 kubelet[3001]: I0302 13:36:26.478288 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfdwx\" (UniqueName: \"kubernetes.io/projected/026e42c9-1716-4494-9461-8f8d83a0e4b6-kube-api-access-pfdwx\") pod \"coredns-66bc5c9577-mp7pf\" (UID: \"026e42c9-1716-4494-9461-8f8d83a0e4b6\") " pod="kube-system/coredns-66bc5c9577-mp7pf" Mar 2 13:36:26.479324 kubelet[3001]: I0302 13:36:26.479296 3001 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/026e42c9-1716-4494-9461-8f8d83a0e4b6-config-volume\") pod \"coredns-66bc5c9577-mp7pf\" (UID: \"026e42c9-1716-4494-9461-8f8d83a0e4b6\") " pod="kube-system/coredns-66bc5c9577-mp7pf" Mar 2 13:36:26.530801 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69-rootfs.mount: Deactivated successfully. Mar 2 13:36:26.818343 kubelet[3001]: E0302 13:36:26.814368 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:26.917115 kubelet[3001]: E0302 13:36:26.912200 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:26.931968 containerd[1614]: time="2026-03-02T13:36:26.931374513Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mp7pf,Uid:026e42c9-1716-4494-9461-8f8d83a0e4b6,Namespace:kube-system,Attempt:0,}" Mar 2 13:36:27.159370 kubelet[3001]: E0302 13:36:27.083152 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:27.159861 containerd[1614]: time="2026-03-02T13:36:27.091165285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2kt5r,Uid:b6bb16d1-3aa0-40ca-9fd6-883920917fda,Namespace:kube-system,Attempt:0,}" Mar 2 13:36:28.023342 kubelet[3001]: E0302 13:36:28.023240 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:28.034329 systemd[1]: run-netns-cni\x2d0d1b33ab\x2d208b\x2d10d6\x2dd0e4\x2db48d38a9d269.mount: Deactivated successfully. Mar 2 13:36:28.037061 systemd[1]: run-netns-cni\x2df9a8308e\x2dcd09\x2d32b5\x2d7a2e\x2d649a2e6ba863.mount: Deactivated successfully. Mar 2 13:36:28.058955 containerd[1614]: time="2026-03-02T13:36:28.054014665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2kt5r,Uid:b6bb16d1-3aa0-40ca-9fd6-883920917fda,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"67fec9700b357e07c0675b4c9e9748b72e121e622a56bc1338b9d1e5b3f9393c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Mar 2 13:36:28.065122 kubelet[3001]: E0302 13:36:28.060536 3001 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"67fec9700b357e07c0675b4c9e9748b72e121e622a56bc1338b9d1e5b3f9393c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Mar 2 13:36:28.065122 kubelet[3001]: E0302 13:36:28.060857 3001 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"67fec9700b357e07c0675b4c9e9748b72e121e622a56bc1338b9d1e5b3f9393c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-66bc5c9577-2kt5r" Mar 2 13:36:28.065122 kubelet[3001]: E0302 13:36:28.060891 3001 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"67fec9700b357e07c0675b4c9e9748b72e121e622a56bc1338b9d1e5b3f9393c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-66bc5c9577-2kt5r" Mar 2 13:36:28.065122 kubelet[3001]: E0302 13:36:28.060949 3001 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-2kt5r_kube-system(b6bb16d1-3aa0-40ca-9fd6-883920917fda)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-2kt5r_kube-system(b6bb16d1-3aa0-40ca-9fd6-883920917fda)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"67fec9700b357e07c0675b4c9e9748b72e121e622a56bc1338b9d1e5b3f9393c\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-66bc5c9577-2kt5r" podUID="b6bb16d1-3aa0-40ca-9fd6-883920917fda" Mar 2 13:36:28.090682 containerd[1614]: time="2026-03-02T13:36:28.090058708Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mp7pf,Uid:026e42c9-1716-4494-9461-8f8d83a0e4b6,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1714cb59cb3379d1d5160fff42f404483bd8b14a5322ea1b87afd67165f79834\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Mar 2 13:36:28.090840 kubelet[3001]: E0302 13:36:28.090357 3001 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1714cb59cb3379d1d5160fff42f404483bd8b14a5322ea1b87afd67165f79834\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Mar 2 13:36:28.098868 kubelet[3001]: E0302 13:36:28.095375 3001 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1714cb59cb3379d1d5160fff42f404483bd8b14a5322ea1b87afd67165f79834\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-66bc5c9577-mp7pf" Mar 2 13:36:28.098868 kubelet[3001]: E0302 13:36:28.095534 3001 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1714cb59cb3379d1d5160fff42f404483bd8b14a5322ea1b87afd67165f79834\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-66bc5c9577-mp7pf" Mar 2 13:36:28.098868 kubelet[3001]: E0302 13:36:28.095850 3001 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-mp7pf_kube-system(026e42c9-1716-4494-9461-8f8d83a0e4b6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-mp7pf_kube-system(026e42c9-1716-4494-9461-8f8d83a0e4b6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1714cb59cb3379d1d5160fff42f404483bd8b14a5322ea1b87afd67165f79834\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-66bc5c9577-mp7pf" podUID="026e42c9-1716-4494-9461-8f8d83a0e4b6" Mar 2 13:36:28.163126 containerd[1614]: time="2026-03-02T13:36:28.162196975Z" level=info msg="CreateContainer within sandbox \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Mar 2 13:36:28.369847 containerd[1614]: time="2026-03-02T13:36:28.360793114Z" level=info msg="Container 23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:36:28.445941 containerd[1614]: time="2026-03-02T13:36:28.444856469Z" level=info msg="CreateContainer within sandbox \"12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0\"" Mar 2 13:36:28.457762 containerd[1614]: time="2026-03-02T13:36:28.450909701Z" level=info msg="StartContainer for \"23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0\"" Mar 2 13:36:28.474944 containerd[1614]: time="2026-03-02T13:36:28.473169886Z" level=info msg="connecting to shim 23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0" address="unix:///run/containerd/s/3388e0d9174e8cda076272eba35e46c5041113f5157f5f3d8de6c58cbe370b43" protocol=ttrpc version=3 Mar 2 13:36:29.808110 systemd[1]: Started cri-containerd-23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0.scope - libcontainer container 23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0. Mar 2 13:36:30.458715 containerd[1614]: time="2026-03-02T13:36:30.456964891Z" level=info msg="StartContainer for \"23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0\" returns successfully" Mar 2 13:36:30.813720 kubelet[3001]: E0302 13:36:30.812086 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:32.469901 systemd-networkd[1516]: flannel.1: Link UP Mar 2 13:36:32.469917 systemd-networkd[1516]: flannel.1: Gained carrier Mar 2 13:36:34.240497 systemd-networkd[1516]: flannel.1: Gained IPv6LL Mar 2 13:36:36.158746 kubelet[3001]: E0302 13:36:36.158325 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:37.240271 kubelet[3001]: E0302 13:36:37.239799 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:37.416394 kubelet[3001]: E0302 13:36:37.333834 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:39.436451 kubelet[3001]: E0302 13:36:39.435716 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.222s" Mar 2 13:36:39.828418 kubelet[3001]: E0302 13:36:39.827826 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:39.943487 kubelet[3001]: E0302 13:36:39.933476 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:40.128018 containerd[1614]: time="2026-03-02T13:36:40.121876042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mp7pf,Uid:026e42c9-1716-4494-9461-8f8d83a0e4b6,Namespace:kube-system,Attempt:0,}" Mar 2 13:36:42.562438 kubelet[3001]: E0302 13:36:42.553198 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.404s" Mar 2 13:36:43.032829 kubelet[3001]: E0302 13:36:43.029443 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:43.037078 containerd[1614]: time="2026-03-02T13:36:43.036245255Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2kt5r,Uid:b6bb16d1-3aa0-40ca-9fd6-883920917fda,Namespace:kube-system,Attempt:0,}" Mar 2 13:36:45.625458 kubelet[3001]: E0302 13:36:45.561321 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.301s" Mar 2 13:36:45.857061 systemd-networkd[1516]: cni0: Link UP Mar 2 13:36:45.857074 systemd-networkd[1516]: cni0: Gained carrier Mar 2 13:36:45.886320 systemd-networkd[1516]: cni0: Lost carrier Mar 2 13:36:46.256157 systemd-networkd[1516]: vethdd0c1b7f: Link UP Mar 2 13:36:46.302706 kernel: cni0: port 1(vethdd0c1b7f) entered blocking state Mar 2 13:36:46.302872 kernel: cni0: port 1(vethdd0c1b7f) entered disabled state Mar 2 13:36:46.303139 kernel: vethdd0c1b7f: entered allmulticast mode Mar 2 13:36:46.427302 kernel: vethdd0c1b7f: entered promiscuous mode Mar 2 13:36:46.694860 systemd-networkd[1516]: veth0d1f40b5: Link UP Mar 2 13:36:46.841952 kernel: cni0: port 2(veth0d1f40b5) entered blocking state Mar 2 13:36:46.845836 kernel: cni0: port 2(veth0d1f40b5) entered disabled state Mar 2 13:36:47.182759 kernel: veth0d1f40b5: entered allmulticast mode Mar 2 13:36:47.282739 kernel: veth0d1f40b5: entered promiscuous mode Mar 2 13:36:47.355196 kernel: cni0: port 1(vethdd0c1b7f) entered blocking state Mar 2 13:36:47.355422 kernel: cni0: port 1(vethdd0c1b7f) entered forwarding state Mar 2 13:36:47.353896 systemd-networkd[1516]: vethdd0c1b7f: Gained carrier Mar 2 13:36:47.356079 systemd-networkd[1516]: cni0: Gained carrier Mar 2 13:36:47.666505 containerd[1614]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil), MTU:0, AdvMSS:0, Priority:0, Table:(*int)(nil), Scope:(*int)(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc000102950), "name":"cbr0", "type":"bridge"} Mar 2 13:36:47.666505 containerd[1614]: delegateAdd: netconf sent to delegate plugin: Mar 2 13:36:47.754203 systemd-networkd[1516]: cni0: Gained IPv6LL Mar 2 13:36:48.203356 kernel: cni0: port 2(veth0d1f40b5) entered blocking state Mar 2 13:36:48.205496 kernel: cni0: port 2(veth0d1f40b5) entered forwarding state Mar 2 13:36:48.203177 systemd-networkd[1516]: veth0d1f40b5: Gained carrier Mar 2 13:36:48.221189 containerd[1614]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"} Mar 2 13:36:48.221189 containerd[1614]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil), MTU:0, AdvMSS:0, Priority:0, Table:(*int)(nil), Scope:(*int)(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc00009a950), "name":"cbr0", "type":"bridge"} Mar 2 13:36:48.221189 containerd[1614]: delegateAdd: netconf sent to delegate plugin: Mar 2 13:36:49.217487 systemd-networkd[1516]: vethdd0c1b7f: Gained IPv6LL Mar 2 13:36:49.757764 containerd[1614]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2026-03-02T13:36:49.750875277Z" level=info msg="connecting to shim c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80" address="unix:///run/containerd/s/f1774323d9f4c3cf86666f6f4582ff273aaa23b08272894e853d24333e7eedec" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:36:49.926375 containerd[1614]: time="2026-03-02T13:36:49.926279444Z" level=info msg="connecting to shim 96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3" address="unix:///run/containerd/s/2b2e8c5d4ad99c5e589420f8dcd355270b510b6bf13c34a7669b88718f06b504" namespace=k8s.io protocol=ttrpc version=3 Mar 2 13:36:50.124748 systemd-networkd[1516]: veth0d1f40b5: Gained IPv6LL Mar 2 13:36:51.404728 kubelet[3001]: E0302 13:36:51.393776 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.243s" Mar 2 13:36:52.643304 systemd[1]: Started cri-containerd-c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80.scope - libcontainer container c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80. Mar 2 13:36:53.035254 systemd[1]: Started cri-containerd-96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3.scope - libcontainer container 96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3. Mar 2 13:36:53.297432 systemd-resolved[1293]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 13:36:53.556071 systemd-resolved[1293]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 13:36:54.439307 containerd[1614]: time="2026-03-02T13:36:54.421864994Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mp7pf,Uid:026e42c9-1716-4494-9461-8f8d83a0e4b6,Namespace:kube-system,Attempt:0,} returns sandbox id \"c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80\"" Mar 2 13:36:54.467165 kubelet[3001]: E0302 13:36:54.456481 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:54.750018 containerd[1614]: time="2026-03-02T13:36:54.747141325Z" level=info msg="CreateContainer within sandbox \"c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 13:36:55.190329 containerd[1614]: time="2026-03-02T13:36:55.120854592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-2kt5r,Uid:b6bb16d1-3aa0-40ca-9fd6-883920917fda,Namespace:kube-system,Attempt:0,} returns sandbox id \"96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3\"" Mar 2 13:36:55.289395 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2122446479.mount: Deactivated successfully. Mar 2 13:36:55.348531 kubelet[3001]: E0302 13:36:55.348490 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:36:55.420245 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount48033379.mount: Deactivated successfully. Mar 2 13:36:55.489663 containerd[1614]: time="2026-03-02T13:36:55.479808496Z" level=info msg="Container f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:36:55.581857 containerd[1614]: time="2026-03-02T13:36:55.558462341Z" level=info msg="CreateContainer within sandbox \"96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 13:36:55.801347 containerd[1614]: time="2026-03-02T13:36:55.797983984Z" level=info msg="CreateContainer within sandbox \"c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe\"" Mar 2 13:36:55.820971 containerd[1614]: time="2026-03-02T13:36:55.820518755Z" level=info msg="StartContainer for \"f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe\"" Mar 2 13:36:55.909960 containerd[1614]: time="2026-03-02T13:36:55.905031395Z" level=info msg="connecting to shim f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe" address="unix:///run/containerd/s/f1774323d9f4c3cf86666f6f4582ff273aaa23b08272894e853d24333e7eedec" protocol=ttrpc version=3 Mar 2 13:36:56.017753 containerd[1614]: time="2026-03-02T13:36:56.016784247Z" level=info msg="Container 8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe: CDI devices from CRI Config.CDIDevices: []" Mar 2 13:36:56.152267 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount745481043.mount: Deactivated successfully. Mar 2 13:36:56.306170 containerd[1614]: time="2026-03-02T13:36:56.302384442Z" level=info msg="CreateContainer within sandbox \"96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe\"" Mar 2 13:36:56.339946 containerd[1614]: time="2026-03-02T13:36:56.339165934Z" level=info msg="StartContainer for \"8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe\"" Mar 2 13:36:56.415936 containerd[1614]: time="2026-03-02T13:36:56.401204211Z" level=info msg="connecting to shim 8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe" address="unix:///run/containerd/s/2b2e8c5d4ad99c5e589420f8dcd355270b510b6bf13c34a7669b88718f06b504" protocol=ttrpc version=3 Mar 2 13:36:57.414382 systemd[1]: Started cri-containerd-f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe.scope - libcontainer container f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe. Mar 2 13:36:58.007299 systemd[1]: Started cri-containerd-8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe.scope - libcontainer container 8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe. Mar 2 13:36:59.215441 containerd[1614]: time="2026-03-02T13:36:59.215392284Z" level=info msg="StartContainer for \"f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe\" returns successfully" Mar 2 13:36:59.343113 containerd[1614]: time="2026-03-02T13:36:59.343018115Z" level=info msg="StartContainer for \"8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe\" returns successfully" Mar 2 13:37:00.025094 kubelet[3001]: E0302 13:37:00.025048 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:00.067871 kubelet[3001]: E0302 13:37:00.055193 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:00.395938 kubelet[3001]: I0302 13:37:00.389008 3001 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-jlpj6" podStartSLOduration=44.762230499 podStartE2EDuration="1m30.388968225s" podCreationTimestamp="2026-03-02 13:35:30 +0000 UTC" firstStartedPulling="2026-03-02 13:35:35.531221227 +0000 UTC m=+22.726813471" lastFinishedPulling="2026-03-02 13:36:21.157958953 +0000 UTC m=+68.353551197" observedRunningTime="2026-03-02 13:36:30.963470547 +0000 UTC m=+78.159062811" watchObservedRunningTime="2026-03-02 13:37:00.388968225 +0000 UTC m=+107.584560479" Mar 2 13:37:00.617873 kubelet[3001]: I0302 13:37:00.612970 3001 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-2kt5r" podStartSLOduration=107.612950043 podStartE2EDuration="1m47.612950043s" podCreationTimestamp="2026-03-02 13:35:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:37:00.420150787 +0000 UTC m=+107.615743041" watchObservedRunningTime="2026-03-02 13:37:00.612950043 +0000 UTC m=+107.808542287" Mar 2 13:37:00.617873 kubelet[3001]: I0302 13:37:00.613084 3001 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-mp7pf" podStartSLOduration=107.613078266 podStartE2EDuration="1m47.613078266s" podCreationTimestamp="2026-03-02 13:35:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:37:00.605365013 +0000 UTC m=+107.800957257" watchObservedRunningTime="2026-03-02 13:37:00.613078266 +0000 UTC m=+107.808670510" Mar 2 13:37:01.099918 kubelet[3001]: E0302 13:37:01.093012 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:01.099918 kubelet[3001]: E0302 13:37:01.097987 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:02.100976 kubelet[3001]: E0302 13:37:02.099497 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:02.122107 kubelet[3001]: E0302 13:37:02.120828 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:37.153948 kubelet[3001]: E0302 13:37:37.151442 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:48.161745 kubelet[3001]: E0302 13:37:48.161197 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:49.206454 kubelet[3001]: E0302 13:37:49.174963 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:37:52.175445 kubelet[3001]: E0302 13:37:52.166010 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:03.444980 kubelet[3001]: E0302 13:38:03.444849 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:17.150779 kubelet[3001]: E0302 13:38:17.150433 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:27.185109 kubelet[3001]: E0302 13:38:27.165494 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:46.174729 kubelet[3001]: E0302 13:38:46.172347 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:52.204227 kubelet[3001]: E0302 13:38:52.204168 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:52.223248 kubelet[3001]: E0302 13:38:52.206039 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:38:52.244984 containerd[1614]: time="2026-03-02T13:38:52.232825990Z" level=info msg="container event discarded" container=1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5 type=CONTAINER_CREATED_EVENT Mar 2 13:38:52.254850 containerd[1614]: time="2026-03-02T13:38:52.247458464Z" level=info msg="container event discarded" container=1932058ed63ef0eb15a96f16a634c32958b94685127ed3aa6e58717fbad525d5 type=CONTAINER_STARTED_EVENT Mar 2 13:38:52.475473 containerd[1614]: time="2026-03-02T13:38:52.464203993Z" level=info msg="container event discarded" container=70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc type=CONTAINER_CREATED_EVENT Mar 2 13:38:52.475473 containerd[1614]: time="2026-03-02T13:38:52.464259295Z" level=info msg="container event discarded" container=70a8a1a789c9b49b20ad600a8c890e8b103b11d64ec59e581a4ad12eb897ddfc type=CONTAINER_STARTED_EVENT Mar 2 13:38:52.490731 containerd[1614]: time="2026-03-02T13:38:52.482745521Z" level=info msg="container event discarded" container=99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8 type=CONTAINER_CREATED_EVENT Mar 2 13:38:52.490731 containerd[1614]: time="2026-03-02T13:38:52.482799591Z" level=info msg="container event discarded" container=99083723406b940a307c4bd8c1af48e77c4241f9a89d6d0afb338054815f76f8 type=CONTAINER_STARTED_EVENT Mar 2 13:38:52.778097 containerd[1614]: time="2026-03-02T13:38:52.771175954Z" level=info msg="container event discarded" container=5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f type=CONTAINER_CREATED_EVENT Mar 2 13:38:52.895068 containerd[1614]: time="2026-03-02T13:38:52.883206939Z" level=info msg="container event discarded" container=fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c type=CONTAINER_CREATED_EVENT Mar 2 13:38:52.956861 containerd[1614]: time="2026-03-02T13:38:52.954430401Z" level=info msg="container event discarded" container=886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623 type=CONTAINER_CREATED_EVENT Mar 2 13:38:53.459535 containerd[1614]: time="2026-03-02T13:38:53.459406879Z" level=info msg="container event discarded" container=5bbc5a14cd4bf0a1cc4f9773db2014e91bc096748eb1ed5eb92bbdbf86fd7a9f type=CONTAINER_STARTED_EVENT Mar 2 13:38:54.151822 containerd[1614]: time="2026-03-02T13:38:54.149736971Z" level=info msg="container event discarded" container=fcb083085bf90f0da9b1c19ab4c6e0823d393ce442349ee1eb87bf6149c2d17c type=CONTAINER_STARTED_EVENT Mar 2 13:38:54.220479 containerd[1614]: time="2026-03-02T13:38:54.217347532Z" level=info msg="container event discarded" container=886aa273f914fd56c7379f580280998972dcba943909645a26fd6f1be551b623 type=CONTAINER_STARTED_EVENT Mar 2 13:39:17.155709 kubelet[3001]: E0302 13:39:17.155353 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:39:18.171471 kubelet[3001]: E0302 13:39:18.168335 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:39:37.153010 kubelet[3001]: E0302 13:39:37.149424 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:39:46.162343 kubelet[3001]: E0302 13:39:46.161969 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:39:52.215289 kubelet[3001]: E0302 13:39:52.212293 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:39:56.259125 kubelet[3001]: E0302 13:39:56.257889 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:40:14.240769 kubelet[3001]: E0302 13:40:14.240308 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:40:19.149350 kubelet[3001]: E0302 13:40:19.149301 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:40:27.155274 kubelet[3001]: E0302 13:40:27.151274 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:40:30.708940 containerd[1614]: time="2026-03-02T13:40:30.708853266Z" level=info msg="container event discarded" container=0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae type=CONTAINER_CREATED_EVENT Mar 2 13:40:30.710195 containerd[1614]: time="2026-03-02T13:40:30.710143056Z" level=info msg="container event discarded" container=0fd76d16232e299d9a9f2e287f872a73d740a728155351df47a4a387a070a6ae type=CONTAINER_STARTED_EVENT Mar 2 13:40:32.920995 containerd[1614]: time="2026-03-02T13:40:32.920901580Z" level=info msg="container event discarded" container=70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27 type=CONTAINER_CREATED_EVENT Mar 2 13:40:33.139935 systemd[1]: Started sshd@7-10.0.0.31:22-10.0.0.1:49602.service - OpenSSH per-connection server daemon (10.0.0.1:49602). Mar 2 13:40:33.187831 systemd[1]: Starting systemd-tmpfiles-clean.service - Cleanup of Temporary Directories... Mar 2 13:40:33.774402 systemd-tmpfiles[4739]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Mar 2 13:40:33.780812 systemd-tmpfiles[4739]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Mar 2 13:40:33.796284 systemd-tmpfiles[4739]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 2 13:40:33.827320 systemd-tmpfiles[4739]: ACLs are not supported, ignoring. Mar 2 13:40:33.828848 systemd-tmpfiles[4739]: ACLs are not supported, ignoring. Mar 2 13:40:33.925877 systemd-tmpfiles[4739]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 13:40:33.925902 systemd-tmpfiles[4739]: Skipping /boot Mar 2 13:40:33.954647 sshd[4738]: Accepted publickey for core from 10.0.0.1 port 49602 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:40:33.974149 sshd-session[4738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:40:34.001237 systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully. Mar 2 13:40:34.007015 systemd[1]: Finished systemd-tmpfiles-clean.service - Cleanup of Temporary Directories. Mar 2 13:40:34.078162 systemd-logind[1587]: New session 9 of user core. Mar 2 13:40:34.116141 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 2 13:40:34.890906 containerd[1614]: time="2026-03-02T13:40:34.890810826Z" level=info msg="container event discarded" container=70437cbec5261d77673462efa34fcf828b95c71f8d38e6ddd014ced582b5df27 type=CONTAINER_STARTED_EVENT Mar 2 13:40:35.539950 containerd[1614]: time="2026-03-02T13:40:35.539239552Z" level=info msg="container event discarded" container=12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876 type=CONTAINER_CREATED_EVENT Mar 2 13:40:35.539950 containerd[1614]: time="2026-03-02T13:40:35.539314441Z" level=info msg="container event discarded" container=12636c7b82e8d72d6680c1b7d00abf7916e3967cb93b9d9e389c2d8dea567876 type=CONTAINER_STARTED_EVENT Mar 2 13:40:36.221319 sshd[4745]: Connection closed by 10.0.0.1 port 49602 Mar 2 13:40:36.221023 sshd-session[4738]: pam_unix(sshd:session): session closed for user core Mar 2 13:40:36.357778 systemd[1]: sshd@7-10.0.0.31:22-10.0.0.1:49602.service: Deactivated successfully. Mar 2 13:40:36.443003 systemd[1]: session-9.scope: Deactivated successfully. Mar 2 13:40:36.490871 systemd-logind[1587]: Session 9 logged out. Waiting for processes to exit. Mar 2 13:40:36.515108 systemd-logind[1587]: Removed session 9. Mar 2 13:40:41.289375 systemd[1]: Started sshd@8-10.0.0.31:22-10.0.0.1:41950.service - OpenSSH per-connection server daemon (10.0.0.1:41950). Mar 2 13:40:42.319904 containerd[1614]: time="2026-03-02T13:40:42.314046369Z" level=info msg="container event discarded" container=d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff type=CONTAINER_CREATED_EVENT Mar 2 13:40:42.523530 sshd[4802]: Accepted publickey for core from 10.0.0.1 port 41950 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:40:42.536030 sshd-session[4802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:40:42.591533 systemd-logind[1587]: New session 10 of user core. Mar 2 13:40:42.659059 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 2 13:40:43.923948 containerd[1614]: time="2026-03-02T13:40:43.913152962Z" level=info msg="container event discarded" container=d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff type=CONTAINER_STARTED_EVENT Mar 2 13:40:44.371757 sshd[4806]: Connection closed by 10.0.0.1 port 41950 Mar 2 13:40:44.352916 sshd-session[4802]: pam_unix(sshd:session): session closed for user core Mar 2 13:40:44.511368 systemd[1]: sshd@8-10.0.0.31:22-10.0.0.1:41950.service: Deactivated successfully. Mar 2 13:40:44.585125 systemd[1]: session-10.scope: Deactivated successfully. Mar 2 13:40:44.633812 systemd-logind[1587]: Session 10 logged out. Waiting for processes to exit. Mar 2 13:40:44.688365 systemd-logind[1587]: Removed session 10. Mar 2 13:40:46.103869 containerd[1614]: time="2026-03-02T13:40:46.092532155Z" level=info msg="container event discarded" container=d88f93955214ee23e7963fddcbc7ca2a79c60ac463e67802f5d06dcba5f329ff type=CONTAINER_STOPPED_EVENT Mar 2 13:40:49.408044 systemd[1]: Started sshd@9-10.0.0.31:22-10.0.0.1:41960.service - OpenSSH per-connection server daemon (10.0.0.1:41960). Mar 2 13:40:50.213909 sshd[4848]: Accepted publickey for core from 10.0.0.1 port 41960 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:40:50.231239 sshd-session[4848]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:40:50.326856 systemd-logind[1587]: New session 11 of user core. Mar 2 13:40:50.351330 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 2 13:40:51.166821 sshd[4853]: Connection closed by 10.0.0.1 port 41960 Mar 2 13:40:51.180274 sshd-session[4848]: pam_unix(sshd:session): session closed for user core Mar 2 13:40:51.217363 systemd[1]: sshd@9-10.0.0.31:22-10.0.0.1:41960.service: Deactivated successfully. Mar 2 13:40:51.256130 systemd[1]: session-11.scope: Deactivated successfully. Mar 2 13:40:51.267909 systemd-logind[1587]: Session 11 logged out. Waiting for processes to exit. Mar 2 13:40:51.272749 systemd-logind[1587]: Removed session 11. Mar 2 13:40:57.005309 systemd[1]: Started sshd@10-10.0.0.31:22-10.0.0.1:37322.service - OpenSSH per-connection server daemon (10.0.0.1:37322). Mar 2 13:41:01.299748 kubelet[3001]: E0302 13:41:01.280979 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.133s" Mar 2 13:41:02.228134 kubelet[3001]: E0302 13:41:02.228090 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:03.250811 sshd[4880]: Accepted publickey for core from 10.0.0.1 port 37322 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:03.265174 sshd-session[4880]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:03.329235 systemd-logind[1587]: New session 12 of user core. Mar 2 13:41:03.369262 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 2 13:41:04.213836 kubelet[3001]: E0302 13:41:04.208831 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:05.210995 sshd[4892]: Connection closed by 10.0.0.1 port 37322 Mar 2 13:41:05.217972 sshd-session[4880]: pam_unix(sshd:session): session closed for user core Mar 2 13:41:05.295084 systemd[1]: sshd@10-10.0.0.31:22-10.0.0.1:37322.service: Deactivated successfully. Mar 2 13:41:05.328149 systemd[1]: session-12.scope: Deactivated successfully. Mar 2 13:41:05.357299 systemd-logind[1587]: Session 12 logged out. Waiting for processes to exit. Mar 2 13:41:05.391364 systemd-logind[1587]: Removed session 12. Mar 2 13:41:09.162155 kubelet[3001]: E0302 13:41:09.152397 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:10.338124 systemd[1]: Started sshd@11-10.0.0.31:22-10.0.0.1:50552.service - OpenSSH per-connection server daemon (10.0.0.1:50552). Mar 2 13:41:11.033137 sshd[4929]: Accepted publickey for core from 10.0.0.1 port 50552 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:11.059097 sshd-session[4929]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:11.187395 kubelet[3001]: E0302 13:41:11.166124 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:11.377392 systemd-logind[1587]: New session 13 of user core. Mar 2 13:41:11.426948 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 2 13:41:13.328396 sshd[4935]: Connection closed by 10.0.0.1 port 50552 Mar 2 13:41:13.328702 sshd-session[4929]: pam_unix(sshd:session): session closed for user core Mar 2 13:41:13.388100 systemd[1]: sshd@11-10.0.0.31:22-10.0.0.1:50552.service: Deactivated successfully. Mar 2 13:41:13.409928 systemd[1]: session-13.scope: Deactivated successfully. Mar 2 13:41:13.453183 systemd-logind[1587]: Session 13 logged out. Waiting for processes to exit. Mar 2 13:41:13.483014 systemd-logind[1587]: Removed session 13. Mar 2 13:41:18.446215 systemd[1]: Started sshd@12-10.0.0.31:22-10.0.0.1:50560.service - OpenSSH per-connection server daemon (10.0.0.1:50560). Mar 2 13:41:19.049962 sshd[4988]: Accepted publickey for core from 10.0.0.1 port 50560 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:19.086979 sshd-session[4988]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:19.165940 systemd-logind[1587]: New session 14 of user core. Mar 2 13:41:19.196042 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 2 13:41:20.528981 sshd[4995]: Connection closed by 10.0.0.1 port 50560 Mar 2 13:41:20.529845 sshd-session[4988]: pam_unix(sshd:session): session closed for user core Mar 2 13:41:20.585205 systemd[1]: sshd@12-10.0.0.31:22-10.0.0.1:50560.service: Deactivated successfully. Mar 2 13:41:20.611141 systemd[1]: session-14.scope: Deactivated successfully. Mar 2 13:41:20.634868 systemd-logind[1587]: Session 14 logged out. Waiting for processes to exit. Mar 2 13:41:20.651858 systemd-logind[1587]: Removed session 14. Mar 2 13:41:21.935829 containerd[1614]: time="2026-03-02T13:41:21.933918175Z" level=info msg="container event discarded" container=8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69 type=CONTAINER_CREATED_EVENT Mar 2 13:41:25.253132 containerd[1614]: time="2026-03-02T13:41:25.252997761Z" level=info msg="container event discarded" container=8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69 type=CONTAINER_STARTED_EVENT Mar 2 13:41:25.654232 systemd[1]: Started sshd@13-10.0.0.31:22-10.0.0.1:53794.service - OpenSSH per-connection server daemon (10.0.0.1:53794). Mar 2 13:41:26.075040 sshd[5031]: Accepted publickey for core from 10.0.0.1 port 53794 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:26.094715 sshd-session[5031]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:26.173719 systemd-logind[1587]: New session 15 of user core. Mar 2 13:41:26.211016 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 2 13:41:26.800194 sshd[5035]: Connection closed by 10.0.0.1 port 53794 Mar 2 13:41:26.806900 sshd-session[5031]: pam_unix(sshd:session): session closed for user core Mar 2 13:41:26.823825 systemd[1]: sshd@13-10.0.0.31:22-10.0.0.1:53794.service: Deactivated successfully. Mar 2 13:41:26.843398 containerd[1614]: time="2026-03-02T13:41:26.823906798Z" level=info msg="container event discarded" container=8f84d82052364b7fd66c1b628356a7902c546de7dc5f689664e75ddabf7e0a69 type=CONTAINER_STOPPED_EVENT Mar 2 13:41:26.839220 systemd[1]: session-15.scope: Deactivated successfully. Mar 2 13:41:26.853801 systemd-logind[1587]: Session 15 logged out. Waiting for processes to exit. Mar 2 13:41:26.865266 systemd-logind[1587]: Removed session 15. Mar 2 13:41:28.454975 containerd[1614]: time="2026-03-02T13:41:28.453318342Z" level=info msg="container event discarded" container=23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0 type=CONTAINER_CREATED_EVENT Mar 2 13:41:29.152233 kubelet[3001]: E0302 13:41:29.152189 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:30.457404 containerd[1614]: time="2026-03-02T13:41:30.452723999Z" level=info msg="container event discarded" container=23ba5e22347ffd481a1ed6f60b853c5a954879e7cb8637fa42587c3c9fe7c5f0 type=CONTAINER_STARTED_EVENT Mar 2 13:41:36.901391 systemd[1]: Started sshd@14-10.0.0.31:22-10.0.0.1:47046.service - OpenSSH per-connection server daemon (10.0.0.1:47046). Mar 2 13:41:37.451172 kubelet[3001]: E0302 13:41:37.449964 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="5.088s" Mar 2 13:41:38.084506 sshd[5073]: Accepted publickey for core from 10.0.0.1 port 47046 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:38.107694 sshd-session[5073]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:38.159798 systemd-logind[1587]: New session 16 of user core. Mar 2 13:41:38.170367 kubelet[3001]: E0302 13:41:38.164261 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:38.226102 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 2 13:41:39.436160 sshd[5096]: Connection closed by 10.0.0.1 port 47046 Mar 2 13:41:39.440079 sshd-session[5073]: pam_unix(sshd:session): session closed for user core Mar 2 13:41:39.525313 systemd[1]: sshd@14-10.0.0.31:22-10.0.0.1:47046.service: Deactivated successfully. Mar 2 13:41:39.570215 systemd[1]: session-16.scope: Deactivated successfully. Mar 2 13:41:39.605281 systemd-logind[1587]: Session 16 logged out. Waiting for processes to exit. Mar 2 13:41:39.631535 systemd-logind[1587]: Removed session 16. Mar 2 13:41:44.501258 systemd[1]: Started sshd@15-10.0.0.31:22-10.0.0.1:46198.service - OpenSSH per-connection server daemon (10.0.0.1:46198). Mar 2 13:41:44.901198 sshd[5131]: Accepted publickey for core from 10.0.0.1 port 46198 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:44.912146 sshd-session[5131]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:44.967139 systemd-logind[1587]: New session 17 of user core. Mar 2 13:41:45.021911 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 2 13:41:46.059088 sshd[5135]: Connection closed by 10.0.0.1 port 46198 Mar 2 13:41:46.058358 sshd-session[5131]: pam_unix(sshd:session): session closed for user core Mar 2 13:41:46.114697 systemd[1]: sshd@15-10.0.0.31:22-10.0.0.1:46198.service: Deactivated successfully. Mar 2 13:41:46.143365 systemd[1]: session-17.scope: Deactivated successfully. Mar 2 13:41:46.153301 kubelet[3001]: E0302 13:41:46.152813 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:41:46.197942 systemd-logind[1587]: Session 17 logged out. Waiting for processes to exit. Mar 2 13:41:46.199993 systemd-logind[1587]: Removed session 17. Mar 2 13:41:51.217121 systemd[1]: Started sshd@16-10.0.0.31:22-10.0.0.1:46006.service - OpenSSH per-connection server daemon (10.0.0.1:46006). Mar 2 13:41:51.883377 sshd[5169]: Accepted publickey for core from 10.0.0.1 port 46006 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:41:51.912089 sshd-session[5169]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:41:52.092107 systemd-logind[1587]: New session 18 of user core. Mar 2 13:41:52.142065 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 2 13:41:54.532106 containerd[1614]: time="2026-03-02T13:41:54.520174697Z" level=info msg="container event discarded" container=c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80 type=CONTAINER_CREATED_EVENT Mar 2 13:41:54.904381 containerd[1614]: time="2026-03-02T13:41:54.742957018Z" level=info msg="container event discarded" container=c2444af1a6dcb5d221e1dc5e1d8fdf42577a37534d4db7bad141aed137c0bb80 type=CONTAINER_STARTED_EVENT Mar 2 13:41:56.399305 containerd[1614]: time="2026-03-02T13:41:56.184829706Z" level=info msg="container event discarded" container=96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3 type=CONTAINER_CREATED_EVENT Mar 2 13:41:56.399305 containerd[1614]: time="2026-03-02T13:41:56.185428453Z" level=info msg="container event discarded" container=96525ae5dcfee9707031bbb61f43b33b4819835d60bd69d2c0487ee0fb0fece3 type=CONTAINER_STARTED_EVENT Mar 2 13:41:58.797395 containerd[1614]: time="2026-03-02T13:41:58.488867005Z" level=info msg="container event discarded" container=f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe type=CONTAINER_CREATED_EVENT Mar 2 13:41:59.063818 containerd[1614]: time="2026-03-02T13:41:59.045206524Z" level=info msg="container event discarded" container=8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe type=CONTAINER_CREATED_EVENT Mar 2 13:41:59.131238 containerd[1614]: time="2026-03-02T13:41:59.108427346Z" level=info msg="container event discarded" container=f7a9f49b866f7dfd6af7c5808febd151c1ba5903306c7235ff88dc5e3bd0bafe type=CONTAINER_STARTED_EVENT Mar 2 13:41:59.201992 kubelet[3001]: E0302 13:41:59.197680 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.666s" Mar 2 13:41:59.314326 containerd[1614]: time="2026-03-02T13:41:59.298376604Z" level=info msg="container event discarded" container=8dfcdd7d02f99ec5693fbf5e8824b6370b2b6ac709ce65e9cb9c0b1f5aecbefe type=CONTAINER_STARTED_EVENT Mar 2 13:42:00.009784 sshd[5173]: Connection closed by 10.0.0.1 port 46006 Mar 2 13:42:00.021296 sshd-session[5169]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:00.109693 systemd[1]: sshd@16-10.0.0.31:22-10.0.0.1:46006.service: Deactivated successfully. Mar 2 13:42:00.185751 systemd[1]: session-18.scope: Deactivated successfully. Mar 2 13:42:00.198412 systemd[1]: session-18.scope: Consumed 1.037s CPU time, 14.6M memory peak. Mar 2 13:42:00.211399 systemd-logind[1587]: Session 18 logged out. Waiting for processes to exit. Mar 2 13:42:00.239210 systemd-logind[1587]: Removed session 18. Mar 2 13:42:05.161452 systemd[1]: Started sshd@17-10.0.0.31:22-10.0.0.1:45470.service - OpenSSH per-connection server daemon (10.0.0.1:45470). Mar 2 13:42:05.795943 sshd[5219]: Accepted publickey for core from 10.0.0.1 port 45470 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:05.813417 sshd-session[5219]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:05.944457 systemd-logind[1587]: New session 19 of user core. Mar 2 13:42:05.974912 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 2 13:42:06.159992 kubelet[3001]: E0302 13:42:06.159849 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:42:07.209207 sshd[5232]: Connection closed by 10.0.0.1 port 45470 Mar 2 13:42:07.206380 sshd-session[5219]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:07.299807 systemd[1]: sshd@17-10.0.0.31:22-10.0.0.1:45470.service: Deactivated successfully. Mar 2 13:42:07.331191 systemd[1]: session-19.scope: Deactivated successfully. Mar 2 13:42:07.364905 systemd-logind[1587]: Session 19 logged out. Waiting for processes to exit. Mar 2 13:42:07.391340 systemd-logind[1587]: Removed session 19. Mar 2 13:42:12.311766 systemd[1]: Started sshd@18-10.0.0.31:22-10.0.0.1:58450.service - OpenSSH per-connection server daemon (10.0.0.1:58450). Mar 2 13:42:12.836164 sshd[5268]: Accepted publickey for core from 10.0.0.1 port 58450 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:12.866464 sshd-session[5268]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:12.921142 systemd-logind[1587]: New session 20 of user core. Mar 2 13:42:12.953362 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 2 13:42:13.153184 kubelet[3001]: E0302 13:42:13.150191 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:42:13.972034 sshd[5272]: Connection closed by 10.0.0.1 port 58450 Mar 2 13:42:13.993279 sshd-session[5268]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:14.116206 systemd[1]: sshd@18-10.0.0.31:22-10.0.0.1:58450.service: Deactivated successfully. Mar 2 13:42:14.142937 systemd[1]: session-20.scope: Deactivated successfully. Mar 2 13:42:14.184847 systemd-logind[1587]: Session 20 logged out. Waiting for processes to exit. Mar 2 13:42:14.211186 systemd-logind[1587]: Removed session 20. Mar 2 13:42:19.198882 systemd[1]: Started sshd@19-10.0.0.31:22-10.0.0.1:58458.service - OpenSSH per-connection server daemon (10.0.0.1:58458). Mar 2 13:42:19.836709 sshd[5308]: Accepted publickey for core from 10.0.0.1 port 58458 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:19.858821 sshd-session[5308]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:19.934781 systemd-logind[1587]: New session 21 of user core. Mar 2 13:42:19.963995 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 2 13:42:21.029197 sshd[5312]: Connection closed by 10.0.0.1 port 58458 Mar 2 13:42:21.042179 sshd-session[5308]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:21.095865 systemd[1]: sshd@19-10.0.0.31:22-10.0.0.1:58458.service: Deactivated successfully. Mar 2 13:42:21.096248 systemd-logind[1587]: Session 21 logged out. Waiting for processes to exit. Mar 2 13:42:21.122388 systemd[1]: session-21.scope: Deactivated successfully. Mar 2 13:42:21.182927 systemd-logind[1587]: Removed session 21. Mar 2 13:42:26.150064 systemd[1]: Started sshd@20-10.0.0.31:22-10.0.0.1:57470.service - OpenSSH per-connection server daemon (10.0.0.1:57470). Mar 2 13:42:27.217052 sshd[5352]: Accepted publickey for core from 10.0.0.1 port 57470 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:27.229327 sshd-session[5352]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:27.349832 systemd-logind[1587]: New session 22 of user core. Mar 2 13:42:27.427984 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 2 13:42:28.431939 sshd[5367]: Connection closed by 10.0.0.1 port 57470 Mar 2 13:42:28.435042 sshd-session[5352]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:28.502292 systemd[1]: sshd@20-10.0.0.31:22-10.0.0.1:57470.service: Deactivated successfully. Mar 2 13:42:28.557782 systemd[1]: session-22.scope: Deactivated successfully. Mar 2 13:42:28.590536 systemd-logind[1587]: Session 22 logged out. Waiting for processes to exit. Mar 2 13:42:28.593170 systemd-logind[1587]: Removed session 22. Mar 2 13:42:29.326291 kubelet[3001]: E0302 13:42:29.290395 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:42:32.155103 kubelet[3001]: E0302 13:42:32.150876 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:42:33.572119 systemd[1]: Started sshd@21-10.0.0.31:22-10.0.0.1:43160.service - OpenSSH per-connection server daemon (10.0.0.1:43160). Mar 2 13:42:34.297115 sshd[5404]: Accepted publickey for core from 10.0.0.1 port 43160 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:34.331302 sshd-session[5404]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:34.454880 systemd-logind[1587]: New session 23 of user core. Mar 2 13:42:34.608405 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 2 13:42:35.683043 sshd[5408]: Connection closed by 10.0.0.1 port 43160 Mar 2 13:42:35.680292 sshd-session[5404]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:35.790134 systemd[1]: Started sshd@22-10.0.0.31:22-10.0.0.1:43174.service - OpenSSH per-connection server daemon (10.0.0.1:43174). Mar 2 13:42:35.794419 systemd[1]: sshd@21-10.0.0.31:22-10.0.0.1:43160.service: Deactivated successfully. Mar 2 13:42:35.817450 systemd[1]: session-23.scope: Deactivated successfully. Mar 2 13:42:35.873251 systemd-logind[1587]: Session 23 logged out. Waiting for processes to exit. Mar 2 13:42:35.934305 systemd-logind[1587]: Removed session 23. Mar 2 13:42:36.668300 sshd[5422]: Accepted publickey for core from 10.0.0.1 port 43174 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:36.671263 sshd-session[5422]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:36.708008 systemd-logind[1587]: New session 24 of user core. Mar 2 13:42:36.750419 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 2 13:42:39.179164 sshd[5434]: Connection closed by 10.0.0.1 port 43174 Mar 2 13:42:39.173427 sshd-session[5422]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:39.289403 systemd[1]: sshd@22-10.0.0.31:22-10.0.0.1:43174.service: Deactivated successfully. Mar 2 13:42:39.335338 systemd[1]: session-24.scope: Deactivated successfully. Mar 2 13:42:39.363033 systemd-logind[1587]: Session 24 logged out. Waiting for processes to exit. Mar 2 13:42:39.466098 systemd[1]: Started sshd@23-10.0.0.31:22-10.0.0.1:43190.service - OpenSSH per-connection server daemon (10.0.0.1:43190). Mar 2 13:42:39.485070 systemd-logind[1587]: Removed session 24. Mar 2 13:42:40.203952 sshd[5464]: Accepted publickey for core from 10.0.0.1 port 43190 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:40.225005 sshd-session[5464]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:40.455476 systemd-logind[1587]: New session 25 of user core. Mar 2 13:42:40.523489 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 2 13:42:46.507951 kubelet[3001]: E0302 13:42:46.507437 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:42:46.666928 kubelet[3001]: E0302 13:42:46.662273 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.703s" Mar 2 13:42:46.863499 sshd[5468]: Connection closed by 10.0.0.1 port 43190 Mar 2 13:42:46.925995 systemd-logind[1587]: Session 25 logged out. Waiting for processes to exit. Mar 2 13:42:46.878236 sshd-session[5464]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:46.940340 systemd[1]: sshd@23-10.0.0.31:22-10.0.0.1:43190.service: Deactivated successfully. Mar 2 13:42:46.996441 systemd[1]: session-25.scope: Deactivated successfully. Mar 2 13:42:47.051256 systemd-logind[1587]: Removed session 25. Mar 2 13:42:52.108413 systemd[1]: Started sshd@24-10.0.0.31:22-10.0.0.1:35406.service - OpenSSH per-connection server daemon (10.0.0.1:35406). Mar 2 13:42:53.909528 sshd[5517]: Accepted publickey for core from 10.0.0.1 port 35406 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:42:53.945405 sshd-session[5517]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:42:54.088213 systemd-logind[1587]: New session 26 of user core. Mar 2 13:42:54.126181 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 2 13:42:55.162888 kubelet[3001]: E0302 13:42:55.156318 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:42:55.918319 sshd[5535]: Connection closed by 10.0.0.1 port 35406 Mar 2 13:42:55.919120 sshd-session[5517]: pam_unix(sshd:session): session closed for user core Mar 2 13:42:56.001295 systemd[1]: sshd@24-10.0.0.31:22-10.0.0.1:35406.service: Deactivated successfully. Mar 2 13:42:56.128522 systemd[1]: session-26.scope: Deactivated successfully. Mar 2 13:42:56.237032 systemd-logind[1587]: Session 26 logged out. Waiting for processes to exit. Mar 2 13:42:56.358488 systemd-logind[1587]: Removed session 26. Mar 2 13:43:01.048118 systemd[1]: Started sshd@25-10.0.0.31:22-10.0.0.1:40202.service - OpenSSH per-connection server daemon (10.0.0.1:40202). Mar 2 13:43:02.147494 sshd[5568]: Accepted publickey for core from 10.0.0.1 port 40202 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:02.186406 sshd-session[5568]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:02.324981 systemd-logind[1587]: New session 27 of user core. Mar 2 13:43:02.441414 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 2 13:43:05.488104 sshd[5575]: Connection closed by 10.0.0.1 port 40202 Mar 2 13:43:05.500091 sshd-session[5568]: pam_unix(sshd:session): session closed for user core Mar 2 13:43:05.625231 systemd[1]: sshd@25-10.0.0.31:22-10.0.0.1:40202.service: Deactivated successfully. Mar 2 13:43:05.673309 systemd[1]: session-27.scope: Deactivated successfully. Mar 2 13:43:05.702127 systemd-logind[1587]: Session 27 logged out. Waiting for processes to exit. Mar 2 13:43:05.739374 systemd-logind[1587]: Removed session 27. Mar 2 13:43:07.166296 kubelet[3001]: E0302 13:43:07.155497 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:43:10.609113 systemd[1]: Started sshd@26-10.0.0.31:22-10.0.0.1:51676.service - OpenSSH per-connection server daemon (10.0.0.1:51676). Mar 2 13:43:11.895362 sshd[5614]: Accepted publickey for core from 10.0.0.1 port 51676 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:11.926381 sshd-session[5614]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:12.088083 systemd-logind[1587]: New session 28 of user core. Mar 2 13:43:12.113474 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 2 13:43:14.357150 sshd[5638]: Connection closed by 10.0.0.1 port 51676 Mar 2 13:43:14.364124 sshd-session[5614]: pam_unix(sshd:session): session closed for user core Mar 2 13:43:14.519533 systemd[1]: sshd@26-10.0.0.31:22-10.0.0.1:51676.service: Deactivated successfully. Mar 2 13:43:14.523947 systemd-logind[1587]: Session 28 logged out. Waiting for processes to exit. Mar 2 13:43:14.549522 systemd[1]: session-28.scope: Deactivated successfully. Mar 2 13:43:14.605112 systemd-logind[1587]: Removed session 28. Mar 2 13:43:19.179847 kubelet[3001]: E0302 13:43:19.177940 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:43:19.487452 systemd[1]: Started sshd@27-10.0.0.31:22-10.0.0.1:51678.service - OpenSSH per-connection server daemon (10.0.0.1:51678). Mar 2 13:43:20.624694 sshd[5682]: Accepted publickey for core from 10.0.0.1 port 51678 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:20.623382 sshd-session[5682]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:20.748002 systemd-logind[1587]: New session 29 of user core. Mar 2 13:43:20.963253 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 2 13:43:22.654068 sshd[5686]: Connection closed by 10.0.0.1 port 51678 Mar 2 13:43:22.652856 sshd-session[5682]: pam_unix(sshd:session): session closed for user core Mar 2 13:43:22.680692 systemd-logind[1587]: Session 29 logged out. Waiting for processes to exit. Mar 2 13:43:22.681236 systemd[1]: sshd@27-10.0.0.31:22-10.0.0.1:51678.service: Deactivated successfully. Mar 2 13:43:22.690042 systemd[1]: session-29.scope: Deactivated successfully. Mar 2 13:43:22.698398 systemd-logind[1587]: Removed session 29. Mar 2 13:43:27.839101 systemd[1]: Started sshd@28-10.0.0.31:22-10.0.0.1:45732.service - OpenSSH per-connection server daemon (10.0.0.1:45732). Mar 2 13:43:28.555474 sshd[5723]: Accepted publickey for core from 10.0.0.1 port 45732 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:28.578241 sshd-session[5723]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:28.724513 systemd-logind[1587]: New session 30 of user core. Mar 2 13:43:28.779439 systemd[1]: Started session-30.scope - Session 30 of User core. Mar 2 13:43:29.185893 kubelet[3001]: E0302 13:43:29.183391 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:43:29.974493 sshd[5741]: Connection closed by 10.0.0.1 port 45732 Mar 2 13:43:29.977494 sshd-session[5723]: pam_unix(sshd:session): session closed for user core Mar 2 13:43:30.038491 systemd-logind[1587]: Session 30 logged out. Waiting for processes to exit. Mar 2 13:43:30.043802 systemd[1]: sshd@28-10.0.0.31:22-10.0.0.1:45732.service: Deactivated successfully. Mar 2 13:43:30.066973 systemd[1]: session-30.scope: Deactivated successfully. Mar 2 13:43:30.085361 systemd-logind[1587]: Removed session 30. Mar 2 13:43:35.093782 systemd[1]: Started sshd@29-10.0.0.31:22-10.0.0.1:32814.service - OpenSSH per-connection server daemon (10.0.0.1:32814). Mar 2 13:43:35.733710 sshd[5776]: Accepted publickey for core from 10.0.0.1 port 32814 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:35.765831 sshd-session[5776]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:35.886169 systemd-logind[1587]: New session 31 of user core. Mar 2 13:43:35.939420 systemd[1]: Started session-31.scope - Session 31 of User core. Mar 2 13:43:36.161285 kubelet[3001]: E0302 13:43:36.153350 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:43:37.102758 sshd[5780]: Connection closed by 10.0.0.1 port 32814 Mar 2 13:43:37.104791 sshd-session[5776]: pam_unix(sshd:session): session closed for user core Mar 2 13:43:37.146828 systemd[1]: sshd@29-10.0.0.31:22-10.0.0.1:32814.service: Deactivated successfully. Mar 2 13:43:37.163499 systemd[1]: session-31.scope: Deactivated successfully. Mar 2 13:43:37.204001 systemd-logind[1587]: Session 31 logged out. Waiting for processes to exit. Mar 2 13:43:37.226777 systemd-logind[1587]: Removed session 31. Mar 2 13:43:42.260721 systemd[1]: Started sshd@30-10.0.0.31:22-10.0.0.1:60286.service - OpenSSH per-connection server daemon (10.0.0.1:60286). Mar 2 13:43:47.785711 kubelet[3001]: E0302 13:43:47.761433 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:43:48.527340 sshd[5815]: Accepted publickey for core from 10.0.0.1 port 60286 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:48.552835 sshd-session[5815]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:48.617882 systemd-logind[1587]: New session 32 of user core. Mar 2 13:43:48.639263 systemd[1]: Started session-32.scope - Session 32 of User core. Mar 2 13:43:49.627176 sshd[5839]: Connection closed by 10.0.0.1 port 60286 Mar 2 13:43:49.633305 sshd-session[5815]: pam_unix(sshd:session): session closed for user core Mar 2 13:43:49.670161 systemd[1]: sshd@30-10.0.0.31:22-10.0.0.1:60286.service: Deactivated successfully. Mar 2 13:43:49.686275 systemd[1]: session-32.scope: Deactivated successfully. Mar 2 13:43:49.707412 systemd-logind[1587]: Session 32 logged out. Waiting for processes to exit. Mar 2 13:43:49.709892 systemd-logind[1587]: Removed session 32. Mar 2 13:43:55.155173 systemd[1]: Started sshd@31-10.0.0.31:22-10.0.0.1:36620.service - OpenSSH per-connection server daemon (10.0.0.1:36620). Mar 2 13:43:56.606459 sshd[5873]: Accepted publickey for core from 10.0.0.1 port 36620 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:43:56.614362 sshd-session[5873]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:43:56.693235 systemd-logind[1587]: New session 33 of user core. Mar 2 13:43:56.715495 systemd[1]: Started session-33.scope - Session 33 of User core. Mar 2 13:43:58.800246 kubelet[3001]: E0302 13:43:58.761220 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:44:00.342041 sshd[5877]: Connection closed by 10.0.0.1 port 36620 Mar 2 13:44:00.342374 sshd-session[5873]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:00.389122 systemd[1]: sshd@31-10.0.0.31:22-10.0.0.1:36620.service: Deactivated successfully. Mar 2 13:44:00.407780 systemd[1]: session-33.scope: Deactivated successfully. Mar 2 13:44:00.420827 systemd-logind[1587]: Session 33 logged out. Waiting for processes to exit. Mar 2 13:44:00.453535 systemd-logind[1587]: Removed session 33. Mar 2 13:44:05.453275 systemd[1]: Started sshd@32-10.0.0.31:22-10.0.0.1:58096.service - OpenSSH per-connection server daemon (10.0.0.1:58096). Mar 2 13:44:05.717921 sshd[5930]: Accepted publickey for core from 10.0.0.1 port 58096 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:05.737480 sshd-session[5930]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:05.802295 systemd-logind[1587]: New session 34 of user core. Mar 2 13:44:05.824790 systemd[1]: Started session-34.scope - Session 34 of User core. Mar 2 13:44:06.772496 sshd[5934]: Connection closed by 10.0.0.1 port 58096 Mar 2 13:44:06.780136 sshd-session[5930]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:06.806857 systemd[1]: sshd@32-10.0.0.31:22-10.0.0.1:58096.service: Deactivated successfully. Mar 2 13:44:06.807435 systemd-logind[1587]: Session 34 logged out. Waiting for processes to exit. Mar 2 13:44:06.815370 systemd[1]: session-34.scope: Deactivated successfully. Mar 2 13:44:06.839090 systemd-logind[1587]: Removed session 34. Mar 2 13:44:11.868461 systemd[1]: Started sshd@33-10.0.0.31:22-10.0.0.1:48720.service - OpenSSH per-connection server daemon (10.0.0.1:48720). Mar 2 13:44:12.371369 sshd[5970]: Accepted publickey for core from 10.0.0.1 port 48720 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:12.400723 sshd-session[5970]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:12.489860 systemd-logind[1587]: New session 35 of user core. Mar 2 13:44:12.496828 systemd[1]: Started session-35.scope - Session 35 of User core. Mar 2 13:44:13.754913 sshd[5974]: Connection closed by 10.0.0.1 port 48720 Mar 2 13:44:13.768417 sshd-session[5970]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:13.832465 systemd[1]: sshd@33-10.0.0.31:22-10.0.0.1:48720.service: Deactivated successfully. Mar 2 13:44:13.858901 systemd[1]: session-35.scope: Deactivated successfully. Mar 2 13:44:13.894775 systemd-logind[1587]: Session 35 logged out. Waiting for processes to exit. Mar 2 13:44:13.928270 systemd-logind[1587]: Removed session 35. Mar 2 13:44:18.913966 systemd[1]: Started sshd@34-10.0.0.31:22-10.0.0.1:48726.service - OpenSSH per-connection server daemon (10.0.0.1:48726). Mar 2 13:44:19.201472 kubelet[3001]: E0302 13:44:19.199986 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:44:19.584740 sshd[6010]: Accepted publickey for core from 10.0.0.1 port 48726 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:19.616397 sshd-session[6010]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:19.713780 systemd-logind[1587]: New session 36 of user core. Mar 2 13:44:19.790877 systemd[1]: Started session-36.scope - Session 36 of User core. Mar 2 13:44:21.642917 sshd[6017]: Connection closed by 10.0.0.1 port 48726 Mar 2 13:44:21.649267 sshd-session[6010]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:21.705505 systemd[1]: sshd@34-10.0.0.31:22-10.0.0.1:48726.service: Deactivated successfully. Mar 2 13:44:21.732346 systemd[1]: session-36.scope: Deactivated successfully. Mar 2 13:44:21.754947 systemd-logind[1587]: Session 36 logged out. Waiting for processes to exit. Mar 2 13:44:21.788709 systemd-logind[1587]: Removed session 36. Mar 2 13:44:26.761199 systemd[1]: Started sshd@35-10.0.0.31:22-10.0.0.1:59484.service - OpenSSH per-connection server daemon (10.0.0.1:59484). Mar 2 13:44:27.213938 sshd[6067]: Accepted publickey for core from 10.0.0.1 port 59484 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:27.236063 sshd-session[6067]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:27.322994 systemd-logind[1587]: New session 37 of user core. Mar 2 13:44:27.380238 systemd[1]: Started session-37.scope - Session 37 of User core. Mar 2 13:44:28.306266 sshd[6071]: Connection closed by 10.0.0.1 port 59484 Mar 2 13:44:28.314751 sshd-session[6067]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:28.359959 systemd[1]: sshd@35-10.0.0.31:22-10.0.0.1:59484.service: Deactivated successfully. Mar 2 13:44:28.378105 systemd[1]: session-37.scope: Deactivated successfully. Mar 2 13:44:28.400338 systemd-logind[1587]: Session 37 logged out. Waiting for processes to exit. Mar 2 13:44:28.419711 systemd-logind[1587]: Removed session 37. Mar 2 13:44:33.421010 systemd[1]: Started sshd@36-10.0.0.31:22-10.0.0.1:37274.service - OpenSSH per-connection server daemon (10.0.0.1:37274). Mar 2 13:44:33.898471 sshd[6104]: Accepted publickey for core from 10.0.0.1 port 37274 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:33.931510 sshd-session[6104]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:34.044091 systemd-logind[1587]: New session 38 of user core. Mar 2 13:44:34.173909 kubelet[3001]: E0302 13:44:34.170444 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:44:34.171945 systemd[1]: Started session-38.scope - Session 38 of User core. Mar 2 13:44:35.217303 sshd[6108]: Connection closed by 10.0.0.1 port 37274 Mar 2 13:44:35.220019 sshd-session[6104]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:35.263839 systemd[1]: sshd@36-10.0.0.31:22-10.0.0.1:37274.service: Deactivated successfully. Mar 2 13:44:35.316835 systemd[1]: session-38.scope: Deactivated successfully. Mar 2 13:44:35.331104 systemd-logind[1587]: Session 38 logged out. Waiting for processes to exit. Mar 2 13:44:35.351750 systemd-logind[1587]: Removed session 38. Mar 2 13:44:36.165001 kubelet[3001]: E0302 13:44:36.161991 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:44:40.332904 systemd[1]: Started sshd@37-10.0.0.31:22-10.0.0.1:52596.service - OpenSSH per-connection server daemon (10.0.0.1:52596). Mar 2 13:44:40.941159 sshd[6145]: Accepted publickey for core from 10.0.0.1 port 52596 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:40.976441 sshd-session[6145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:41.072062 systemd-logind[1587]: New session 39 of user core. Mar 2 13:44:41.130162 systemd[1]: Started session-39.scope - Session 39 of User core. Mar 2 13:44:42.421005 sshd[6155]: Connection closed by 10.0.0.1 port 52596 Mar 2 13:44:42.416966 sshd-session[6145]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:42.473111 systemd[1]: sshd@37-10.0.0.31:22-10.0.0.1:52596.service: Deactivated successfully. Mar 2 13:44:42.525920 systemd[1]: session-39.scope: Deactivated successfully. Mar 2 13:44:42.550798 systemd-logind[1587]: Session 39 logged out. Waiting for processes to exit. Mar 2 13:44:42.608071 systemd-logind[1587]: Removed session 39. Mar 2 13:44:47.552209 systemd[1]: Started sshd@38-10.0.0.31:22-10.0.0.1:52606.service - OpenSSH per-connection server daemon (10.0.0.1:52606). Mar 2 13:44:48.405984 sshd[6190]: Accepted publickey for core from 10.0.0.1 port 52606 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:48.434529 sshd-session[6190]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:48.596419 systemd-logind[1587]: New session 40 of user core. Mar 2 13:44:48.697892 systemd[1]: Started session-40.scope - Session 40 of User core. Mar 2 13:44:50.499817 sshd[6205]: Connection closed by 10.0.0.1 port 52606 Mar 2 13:44:50.504918 sshd-session[6190]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:50.540208 systemd[1]: sshd@38-10.0.0.31:22-10.0.0.1:52606.service: Deactivated successfully. Mar 2 13:44:50.560211 systemd[1]: session-40.scope: Deactivated successfully. Mar 2 13:44:50.565793 systemd-logind[1587]: Session 40 logged out. Waiting for processes to exit. Mar 2 13:44:50.624877 systemd-logind[1587]: Removed session 40. Mar 2 13:44:55.605175 systemd[1]: Started sshd@39-10.0.0.31:22-10.0.0.1:42480.service - OpenSSH per-connection server daemon (10.0.0.1:42480). Mar 2 13:44:56.424414 sshd[6242]: Accepted publickey for core from 10.0.0.1 port 42480 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:56.446188 sshd-session[6242]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:56.616138 systemd-logind[1587]: New session 41 of user core. Mar 2 13:44:56.657516 systemd[1]: Started session-41.scope - Session 41 of User core. Mar 2 13:44:58.470090 sshd[6252]: Connection closed by 10.0.0.1 port 42480 Mar 2 13:44:58.489264 sshd-session[6242]: pam_unix(sshd:session): session closed for user core Mar 2 13:44:58.617242 systemd[1]: sshd@39-10.0.0.31:22-10.0.0.1:42480.service: Deactivated successfully. Mar 2 13:44:58.655253 systemd[1]: session-41.scope: Deactivated successfully. Mar 2 13:44:58.697905 systemd-logind[1587]: Session 41 logged out. Waiting for processes to exit. Mar 2 13:44:58.761953 systemd[1]: Started sshd@40-10.0.0.31:22-10.0.0.1:42488.service - OpenSSH per-connection server daemon (10.0.0.1:42488). Mar 2 13:44:58.805843 systemd-logind[1587]: Removed session 41. Mar 2 13:44:59.161193 kubelet[3001]: E0302 13:44:59.161149 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:44:59.702491 sshd[6265]: Accepted publickey for core from 10.0.0.1 port 42488 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:44:59.745531 sshd-session[6265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:44:59.880831 systemd-logind[1587]: New session 42 of user core. Mar 2 13:44:59.940963 systemd[1]: Started session-42.scope - Session 42 of User core. Mar 2 13:45:02.196239 kubelet[3001]: E0302 13:45:02.186499 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:45:04.309498 sshd[6281]: Connection closed by 10.0.0.1 port 42488 Mar 2 13:45:04.308999 sshd-session[6265]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:04.406903 systemd[1]: Started sshd@41-10.0.0.31:22-10.0.0.1:43214.service - OpenSSH per-connection server daemon (10.0.0.1:43214). Mar 2 13:45:04.412266 systemd[1]: sshd@40-10.0.0.31:22-10.0.0.1:42488.service: Deactivated successfully. Mar 2 13:45:04.444951 systemd[1]: session-42.scope: Deactivated successfully. Mar 2 13:45:04.501826 systemd-logind[1587]: Session 42 logged out. Waiting for processes to exit. Mar 2 13:45:04.529282 systemd-logind[1587]: Removed session 42. Mar 2 13:45:05.098696 sshd[6298]: Accepted publickey for core from 10.0.0.1 port 43214 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:05.114194 sshd-session[6298]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:05.201887 systemd-logind[1587]: New session 43 of user core. Mar 2 13:45:05.237091 systemd[1]: Started session-43.scope - Session 43 of User core. Mar 2 13:45:12.163213 kubelet[3001]: E0302 13:45:12.163167 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:45:13.742156 sshd[6308]: Connection closed by 10.0.0.1 port 43214 Mar 2 13:45:13.749990 sshd-session[6298]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:13.937242 systemd[1]: Started sshd@42-10.0.0.31:22-10.0.0.1:51114.service - OpenSSH per-connection server daemon (10.0.0.1:51114). Mar 2 13:45:13.966209 systemd[1]: sshd@41-10.0.0.31:22-10.0.0.1:43214.service: Deactivated successfully. Mar 2 13:45:14.068921 systemd[1]: session-43.scope: Deactivated successfully. Mar 2 13:45:14.070152 systemd[1]: session-43.scope: Consumed 2.311s CPU time, 36.2M memory peak. Mar 2 13:45:14.304059 systemd-logind[1587]: Session 43 logged out. Waiting for processes to exit. Mar 2 13:45:14.336279 systemd-logind[1587]: Removed session 43. Mar 2 13:45:14.950343 sshd[6366]: Accepted publickey for core from 10.0.0.1 port 51114 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:14.961293 sshd-session[6366]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:15.110019 systemd-logind[1587]: New session 44 of user core. Mar 2 13:45:15.173121 systemd[1]: Started session-44.scope - Session 44 of User core. Mar 2 13:45:24.225172 kernel: sched: DL replenish lagged too much Mar 2 13:45:24.562766 kubelet[3001]: E0302 13:45:24.521334 3001 controller.go:195] "Failed to update lease" err="Put \"https://10.0.0.31:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 2 13:45:26.352731 kubelet[3001]: E0302 13:45:26.344892 3001 kubelet.go:2618] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.006s" Mar 2 13:45:26.490351 kubelet[3001]: E0302 13:45:26.489985 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:45:27.391688 kubelet[3001]: E0302 13:45:27.382991 3001 controller.go:195] "Failed to update lease" err="Operation cannot be fulfilled on leases.coordination.k8s.io \"localhost\": the object has been modified; please apply your changes to the latest version and try again" Mar 2 13:45:28.818805 sshd[6379]: Connection closed by 10.0.0.1 port 51114 Mar 2 13:45:28.823966 sshd-session[6366]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:28.865215 systemd[1]: sshd@42-10.0.0.31:22-10.0.0.1:51114.service: Deactivated successfully. Mar 2 13:45:28.877886 systemd[1]: session-44.scope: Deactivated successfully. Mar 2 13:45:28.898124 systemd[1]: Started sshd@43-10.0.0.31:22-10.0.0.1:51328.service - OpenSSH per-connection server daemon (10.0.0.1:51328). Mar 2 13:45:28.909839 systemd-logind[1587]: Session 44 logged out. Waiting for processes to exit. Mar 2 13:45:28.924324 systemd-logind[1587]: Removed session 44. Mar 2 13:45:29.317848 sshd[6410]: Accepted publickey for core from 10.0.0.1 port 51328 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:29.328425 sshd-session[6410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:29.399217 systemd-logind[1587]: New session 45 of user core. Mar 2 13:45:29.430991 systemd[1]: Started session-45.scope - Session 45 of User core. Mar 2 13:45:30.046771 sshd[6414]: Connection closed by 10.0.0.1 port 51328 Mar 2 13:45:30.046103 sshd-session[6410]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:30.114818 systemd-logind[1587]: Session 45 logged out. Waiting for processes to exit. Mar 2 13:45:30.118102 systemd[1]: sshd@43-10.0.0.31:22-10.0.0.1:51328.service: Deactivated successfully. Mar 2 13:45:30.141337 systemd[1]: session-45.scope: Deactivated successfully. Mar 2 13:45:30.179376 systemd-logind[1587]: Removed session 45. Mar 2 13:45:35.108723 systemd[1]: Started sshd@44-10.0.0.31:22-10.0.0.1:43274.service - OpenSSH per-connection server daemon (10.0.0.1:43274). Mar 2 13:45:35.516825 sshd[6448]: Accepted publickey for core from 10.0.0.1 port 43274 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:35.514865 sshd-session[6448]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:35.557028 systemd-logind[1587]: New session 46 of user core. Mar 2 13:45:35.572489 systemd[1]: Started session-46.scope - Session 46 of User core. Mar 2 13:45:36.179408 kubelet[3001]: E0302 13:45:36.169399 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:45:36.249965 sshd[6452]: Connection closed by 10.0.0.1 port 43274 Mar 2 13:45:36.249280 sshd-session[6448]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:36.288501 systemd[1]: sshd@44-10.0.0.31:22-10.0.0.1:43274.service: Deactivated successfully. Mar 2 13:45:36.301154 systemd[1]: session-46.scope: Deactivated successfully. Mar 2 13:45:36.330214 systemd-logind[1587]: Session 46 logged out. Waiting for processes to exit. Mar 2 13:45:36.332174 systemd-logind[1587]: Removed session 46. Mar 2 13:45:41.352217 systemd[1]: Started sshd@45-10.0.0.31:22-10.0.0.1:47992.service - OpenSSH per-connection server daemon (10.0.0.1:47992). Mar 2 13:45:41.979980 sshd[6488]: Accepted publickey for core from 10.0.0.1 port 47992 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:42.016404 sshd-session[6488]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:42.111095 systemd-logind[1587]: New session 47 of user core. Mar 2 13:45:42.153780 systemd[1]: Started session-47.scope - Session 47 of User core. Mar 2 13:45:43.267041 sshd[6493]: Connection closed by 10.0.0.1 port 47992 Mar 2 13:45:43.266360 sshd-session[6488]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:43.320342 systemd[1]: sshd@45-10.0.0.31:22-10.0.0.1:47992.service: Deactivated successfully. Mar 2 13:45:43.339936 systemd[1]: session-47.scope: Deactivated successfully. Mar 2 13:45:43.351401 systemd-logind[1587]: Session 47 logged out. Waiting for processes to exit. Mar 2 13:45:43.353766 systemd-logind[1587]: Removed session 47. Mar 2 13:45:45.151950 kubelet[3001]: E0302 13:45:45.149514 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:45:48.352408 systemd[1]: Started sshd@46-10.0.0.31:22-10.0.0.1:48008.service - OpenSSH per-connection server daemon (10.0.0.1:48008). Mar 2 13:45:48.802462 sshd[6531]: Accepted publickey for core from 10.0.0.1 port 48008 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:48.835351 sshd-session[6531]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:48.936167 systemd-logind[1587]: New session 48 of user core. Mar 2 13:45:48.969179 systemd[1]: Started session-48.scope - Session 48 of User core. Mar 2 13:45:49.828923 sshd[6538]: Connection closed by 10.0.0.1 port 48008 Mar 2 13:45:49.831172 sshd-session[6531]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:49.869239 systemd[1]: sshd@46-10.0.0.31:22-10.0.0.1:48008.service: Deactivated successfully. Mar 2 13:45:49.882997 systemd[1]: session-48.scope: Deactivated successfully. Mar 2 13:45:49.911306 systemd-logind[1587]: Session 48 logged out. Waiting for processes to exit. Mar 2 13:45:49.927101 systemd-logind[1587]: Removed session 48. Mar 2 13:45:52.151052 kubelet[3001]: E0302 13:45:52.150100 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:45:54.889464 systemd[1]: Started sshd@47-10.0.0.31:22-10.0.0.1:56496.service - OpenSSH per-connection server daemon (10.0.0.1:56496). Mar 2 13:45:57.082094 sshd[6583]: Accepted publickey for core from 10.0.0.1 port 56496 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:45:57.098297 sshd-session[6583]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:45:57.147851 systemd-logind[1587]: New session 49 of user core. Mar 2 13:45:57.193102 systemd[1]: Started session-49.scope - Session 49 of User core. Mar 2 13:45:57.710849 sshd[6587]: Connection closed by 10.0.0.1 port 56496 Mar 2 13:45:57.711013 sshd-session[6583]: pam_unix(sshd:session): session closed for user core Mar 2 13:45:57.743248 systemd[1]: sshd@47-10.0.0.31:22-10.0.0.1:56496.service: Deactivated successfully. Mar 2 13:45:57.754843 systemd[1]: session-49.scope: Deactivated successfully. Mar 2 13:45:57.773908 systemd-logind[1587]: Session 49 logged out. Waiting for processes to exit. Mar 2 13:45:57.795270 systemd-logind[1587]: Removed session 49. Mar 2 13:46:02.864151 systemd[1]: Started sshd@48-10.0.0.31:22-10.0.0.1:50338.service - OpenSSH per-connection server daemon (10.0.0.1:50338). Mar 2 13:46:03.906905 sshd[6626]: Accepted publickey for core from 10.0.0.1 port 50338 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:03.915958 sshd-session[6626]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:03.992489 systemd-logind[1587]: New session 50 of user core. Mar 2 13:46:04.020520 systemd[1]: Started session-50.scope - Session 50 of User core. Mar 2 13:46:04.785725 sshd[6631]: Connection closed by 10.0.0.1 port 50338 Mar 2 13:46:04.798138 sshd-session[6626]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:04.832511 systemd[1]: sshd@48-10.0.0.31:22-10.0.0.1:50338.service: Deactivated successfully. Mar 2 13:46:04.887997 systemd[1]: session-50.scope: Deactivated successfully. Mar 2 13:46:04.900120 systemd-logind[1587]: Session 50 logged out. Waiting for processes to exit. Mar 2 13:46:04.919432 systemd-logind[1587]: Removed session 50. Mar 2 13:46:05.150941 kubelet[3001]: E0302 13:46:05.150300 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:46:11.710364 systemd[1]: Started sshd@49-10.0.0.31:22-10.0.0.1:50348.service - OpenSSH per-connection server daemon (10.0.0.1:50348). Mar 2 13:46:12.416416 sshd[6668]: Accepted publickey for core from 10.0.0.1 port 50348 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:12.455207 sshd-session[6668]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:12.559258 systemd-logind[1587]: New session 51 of user core. Mar 2 13:46:12.614276 systemd[1]: Started session-51.scope - Session 51 of User core. Mar 2 13:46:13.562754 kubelet[3001]: E0302 13:46:13.562021 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:46:15.160601 sshd[6684]: Connection closed by 10.0.0.1 port 50348 Mar 2 13:46:15.167047 sshd-session[6668]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:15.219351 systemd[1]: sshd@49-10.0.0.31:22-10.0.0.1:50348.service: Deactivated successfully. Mar 2 13:46:15.259465 systemd[1]: session-51.scope: Deactivated successfully. Mar 2 13:46:15.278170 systemd-logind[1587]: Session 51 logged out. Waiting for processes to exit. Mar 2 13:46:15.287056 systemd-logind[1587]: Removed session 51. Mar 2 13:46:20.246403 systemd[1]: Started sshd@50-10.0.0.31:22-10.0.0.1:57874.service - OpenSSH per-connection server daemon (10.0.0.1:57874). Mar 2 13:46:20.715454 sshd[6726]: Accepted publickey for core from 10.0.0.1 port 57874 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:20.735179 sshd-session[6726]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:20.816793 systemd-logind[1587]: New session 52 of user core. Mar 2 13:46:20.845176 systemd[1]: Started session-52.scope - Session 52 of User core. Mar 2 13:46:21.872958 sshd[6730]: Connection closed by 10.0.0.1 port 57874 Mar 2 13:46:21.881451 sshd-session[6726]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:21.946057 systemd[1]: sshd@50-10.0.0.31:22-10.0.0.1:57874.service: Deactivated successfully. Mar 2 13:46:21.961048 systemd[1]: session-52.scope: Deactivated successfully. Mar 2 13:46:21.998229 systemd-logind[1587]: Session 52 logged out. Waiting for processes to exit. Mar 2 13:46:22.028533 systemd-logind[1587]: Removed session 52. Mar 2 13:46:27.028864 systemd[1]: Started sshd@51-10.0.0.31:22-10.0.0.1:57890.service - OpenSSH per-connection server daemon (10.0.0.1:57890). Mar 2 13:46:27.712139 sshd[6763]: Accepted publickey for core from 10.0.0.1 port 57890 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:27.740373 sshd-session[6763]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:27.802346 systemd-logind[1587]: New session 53 of user core. Mar 2 13:46:27.848744 systemd[1]: Started session-53.scope - Session 53 of User core. Mar 2 13:46:28.933886 sshd[6767]: Connection closed by 10.0.0.1 port 57890 Mar 2 13:46:28.936282 sshd-session[6763]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:28.983349 systemd-logind[1587]: Session 53 logged out. Waiting for processes to exit. Mar 2 13:46:28.991169 systemd[1]: sshd@51-10.0.0.31:22-10.0.0.1:57890.service: Deactivated successfully. Mar 2 13:46:29.007794 systemd[1]: session-53.scope: Deactivated successfully. Mar 2 13:46:29.026442 systemd-logind[1587]: Removed session 53. Mar 2 13:46:30.159277 kubelet[3001]: E0302 13:46:30.153764 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:46:34.075461 systemd[1]: Started sshd@52-10.0.0.31:22-10.0.0.1:57202.service - OpenSSH per-connection server daemon (10.0.0.1:57202). Mar 2 13:46:35.430518 sshd[6814]: Accepted publickey for core from 10.0.0.1 port 57202 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:35.445865 sshd-session[6814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:35.598512 systemd-logind[1587]: New session 54 of user core. Mar 2 13:46:35.650379 systemd[1]: Started session-54.scope - Session 54 of User core. Mar 2 13:46:36.431881 sshd[6824]: Connection closed by 10.0.0.1 port 57202 Mar 2 13:46:36.442745 sshd-session[6814]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:36.470454 systemd[1]: sshd@52-10.0.0.31:22-10.0.0.1:57202.service: Deactivated successfully. Mar 2 13:46:36.484259 systemd[1]: session-54.scope: Deactivated successfully. Mar 2 13:46:36.495533 systemd-logind[1587]: Session 54 logged out. Waiting for processes to exit. Mar 2 13:46:36.501907 systemd-logind[1587]: Removed session 54. Mar 2 13:46:41.568814 systemd[1]: Started sshd@53-10.0.0.31:22-10.0.0.1:55162.service - OpenSSH per-connection server daemon (10.0.0.1:55162). Mar 2 13:46:42.001811 sshd[6860]: Accepted publickey for core from 10.0.0.1 port 55162 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:42.021849 sshd-session[6860]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:42.107522 systemd-logind[1587]: New session 55 of user core. Mar 2 13:46:42.142341 systemd[1]: Started session-55.scope - Session 55 of User core. Mar 2 13:46:42.898456 sshd[6864]: Connection closed by 10.0.0.1 port 55162 Mar 2 13:46:42.897883 sshd-session[6860]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:42.941143 systemd[1]: sshd@53-10.0.0.31:22-10.0.0.1:55162.service: Deactivated successfully. Mar 2 13:46:42.979883 systemd[1]: session-55.scope: Deactivated successfully. Mar 2 13:46:43.011385 systemd-logind[1587]: Session 55 logged out. Waiting for processes to exit. Mar 2 13:46:43.022492 systemd-logind[1587]: Removed session 55. Mar 2 13:46:48.023444 systemd[1]: Started sshd@54-10.0.0.31:22-10.0.0.1:55174.service - OpenSSH per-connection server daemon (10.0.0.1:55174). Mar 2 13:46:48.735804 sshd[6898]: Accepted publickey for core from 10.0.0.1 port 55174 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:48.739205 sshd-session[6898]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:48.836536 systemd-logind[1587]: New session 56 of user core. Mar 2 13:46:48.874267 systemd[1]: Started session-56.scope - Session 56 of User core. Mar 2 13:46:50.400522 sshd[6902]: Connection closed by 10.0.0.1 port 55174 Mar 2 13:46:50.404997 sshd-session[6898]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:50.465240 systemd[1]: sshd@54-10.0.0.31:22-10.0.0.1:55174.service: Deactivated successfully. Mar 2 13:46:50.466459 systemd-logind[1587]: Session 56 logged out. Waiting for processes to exit. Mar 2 13:46:50.515970 systemd[1]: session-56.scope: Deactivated successfully. Mar 2 13:46:50.566207 systemd-logind[1587]: Removed session 56. Mar 2 13:46:53.152819 kubelet[3001]: E0302 13:46:53.151910 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:46:55.532468 systemd[1]: Started sshd@55-10.0.0.31:22-10.0.0.1:39612.service - OpenSSH per-connection server daemon (10.0.0.1:39612). Mar 2 13:46:56.420940 sshd[6937]: Accepted publickey for core from 10.0.0.1 port 39612 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:46:56.430931 sshd-session[6937]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:46:56.568411 systemd-logind[1587]: New session 57 of user core. Mar 2 13:46:56.631035 systemd[1]: Started session-57.scope - Session 57 of User core. Mar 2 13:46:57.203824 kubelet[3001]: E0302 13:46:57.160912 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:46:57.203824 kubelet[3001]: E0302 13:46:57.163307 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:46:58.319359 sshd[6959]: Connection closed by 10.0.0.1 port 39612 Mar 2 13:46:58.328255 sshd-session[6937]: pam_unix(sshd:session): session closed for user core Mar 2 13:46:58.393933 systemd[1]: sshd@55-10.0.0.31:22-10.0.0.1:39612.service: Deactivated successfully. Mar 2 13:46:58.422922 systemd[1]: session-57.scope: Deactivated successfully. Mar 2 13:46:58.433920 systemd-logind[1587]: Session 57 logged out. Waiting for processes to exit. Mar 2 13:46:58.502303 systemd-logind[1587]: Removed session 57. Mar 2 13:47:03.688836 systemd[1]: Started sshd@56-10.0.0.31:22-10.0.0.1:33188.service - OpenSSH per-connection server daemon (10.0.0.1:33188). Mar 2 13:47:04.647363 sshd[6993]: Accepted publickey for core from 10.0.0.1 port 33188 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:04.663074 sshd-session[6993]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:04.853065 systemd-logind[1587]: New session 58 of user core. Mar 2 13:47:04.948128 systemd[1]: Started session-58.scope - Session 58 of User core. Mar 2 13:47:06.419031 sshd[6997]: Connection closed by 10.0.0.1 port 33188 Mar 2 13:47:06.426046 sshd-session[6993]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:06.510869 systemd[1]: sshd@56-10.0.0.31:22-10.0.0.1:33188.service: Deactivated successfully. Mar 2 13:47:06.531776 systemd-logind[1587]: Session 58 logged out. Waiting for processes to exit. Mar 2 13:47:06.558749 systemd[1]: session-58.scope: Deactivated successfully. Mar 2 13:47:06.602740 systemd-logind[1587]: Removed session 58. Mar 2 13:47:09.225953 kubelet[3001]: E0302 13:47:09.224150 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:47:11.553160 systemd[1]: Started sshd@57-10.0.0.31:22-10.0.0.1:47734.service - OpenSSH per-connection server daemon (10.0.0.1:47734). Mar 2 13:47:11.984984 sshd[7038]: Accepted publickey for core from 10.0.0.1 port 47734 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:12.000120 sshd-session[7038]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:12.045781 systemd-logind[1587]: New session 59 of user core. Mar 2 13:47:12.083114 systemd[1]: Started session-59.scope - Session 59 of User core. Mar 2 13:47:12.992754 sshd[7043]: Connection closed by 10.0.0.1 port 47734 Mar 2 13:47:13.000121 sshd-session[7038]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:13.051767 systemd[1]: sshd@57-10.0.0.31:22-10.0.0.1:47734.service: Deactivated successfully. Mar 2 13:47:13.087889 systemd[1]: session-59.scope: Deactivated successfully. Mar 2 13:47:13.116914 systemd-logind[1587]: Session 59 logged out. Waiting for processes to exit. Mar 2 13:47:13.127046 systemd-logind[1587]: Removed session 59. Mar 2 13:47:17.169584 kubelet[3001]: E0302 13:47:17.157171 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:47:18.178446 systemd[1]: Started sshd@58-10.0.0.31:22-10.0.0.1:47748.service - OpenSSH per-connection server daemon (10.0.0.1:47748). Mar 2 13:47:18.610979 sshd[7088]: Accepted publickey for core from 10.0.0.1 port 47748 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:18.627946 sshd-session[7088]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:18.717833 systemd-logind[1587]: New session 60 of user core. Mar 2 13:47:18.770456 systemd[1]: Started session-60.scope - Session 60 of User core. Mar 2 13:47:19.810930 sshd[7098]: Connection closed by 10.0.0.1 port 47748 Mar 2 13:47:19.810984 sshd-session[7088]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:19.889952 systemd[1]: sshd@58-10.0.0.31:22-10.0.0.1:47748.service: Deactivated successfully. Mar 2 13:47:19.929131 systemd[1]: session-60.scope: Deactivated successfully. Mar 2 13:47:19.991993 systemd-logind[1587]: Session 60 logged out. Waiting for processes to exit. Mar 2 13:47:20.059956 systemd-logind[1587]: Removed session 60. Mar 2 13:47:24.159396 kubelet[3001]: E0302 13:47:24.156404 3001 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:47:24.866190 systemd[1]: Started sshd@59-10.0.0.31:22-10.0.0.1:52414.service - OpenSSH per-connection server daemon (10.0.0.1:52414). Mar 2 13:47:25.167787 sshd[7134]: Accepted publickey for core from 10.0.0.1 port 52414 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:25.178009 sshd-session[7134]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:25.214796 systemd-logind[1587]: New session 61 of user core. Mar 2 13:47:25.236052 systemd[1]: Started session-61.scope - Session 61 of User core. Mar 2 13:47:25.691432 sshd[7139]: Connection closed by 10.0.0.1 port 52414 Mar 2 13:47:25.692984 sshd-session[7134]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:25.733072 systemd[1]: sshd@59-10.0.0.31:22-10.0.0.1:52414.service: Deactivated successfully. Mar 2 13:47:25.755946 systemd[1]: session-61.scope: Deactivated successfully. Mar 2 13:47:25.798239 systemd-logind[1587]: Session 61 logged out. Waiting for processes to exit. Mar 2 13:47:25.832874 systemd-logind[1587]: Removed session 61. Mar 2 13:47:30.723995 systemd[1]: Started sshd@60-10.0.0.31:22-10.0.0.1:43702.service - OpenSSH per-connection server daemon (10.0.0.1:43702). Mar 2 13:47:31.040757 sshd[7173]: Accepted publickey for core from 10.0.0.1 port 43702 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:31.050438 sshd-session[7173]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:31.081529 systemd-logind[1587]: New session 62 of user core. Mar 2 13:47:31.097966 systemd[1]: Started session-62.scope - Session 62 of User core. Mar 2 13:47:31.566322 sshd[7177]: Connection closed by 10.0.0.1 port 43702 Mar 2 13:47:31.567049 sshd-session[7173]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:31.607102 systemd[1]: sshd@60-10.0.0.31:22-10.0.0.1:43702.service: Deactivated successfully. Mar 2 13:47:31.618060 systemd[1]: session-62.scope: Deactivated successfully. Mar 2 13:47:31.635436 systemd-logind[1587]: Session 62 logged out. Waiting for processes to exit. Mar 2 13:47:31.639303 systemd-logind[1587]: Removed session 62. Mar 2 13:47:36.629169 systemd[1]: Started sshd@61-10.0.0.31:22-10.0.0.1:43718.service - OpenSSH per-connection server daemon (10.0.0.1:43718). Mar 2 13:47:36.998341 sshd[7210]: Accepted publickey for core from 10.0.0.1 port 43718 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:37.006174 sshd-session[7210]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:37.040719 systemd-logind[1587]: New session 63 of user core. Mar 2 13:47:37.054738 systemd[1]: Started session-63.scope - Session 63 of User core. Mar 2 13:47:37.555851 sshd[7214]: Connection closed by 10.0.0.1 port 43718 Mar 2 13:47:37.558940 sshd-session[7210]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:37.588822 systemd-logind[1587]: Session 63 logged out. Waiting for processes to exit. Mar 2 13:47:37.601974 systemd[1]: sshd@61-10.0.0.31:22-10.0.0.1:43718.service: Deactivated successfully. Mar 2 13:47:37.611525 systemd[1]: session-63.scope: Deactivated successfully. Mar 2 13:47:37.617941 systemd-logind[1587]: Removed session 63. Mar 2 13:47:42.645315 systemd[1]: Started sshd@62-10.0.0.31:22-10.0.0.1:32946.service - OpenSSH per-connection server daemon (10.0.0.1:32946). Mar 2 13:47:42.990349 sshd[7256]: Accepted publickey for core from 10.0.0.1 port 32946 ssh2: RSA SHA256:4o3SX1Y9Osa46bn9P83Ew6CNhhKj/2fpc+TFeqtGVQs Mar 2 13:47:42.995102 sshd-session[7256]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:47:43.052094 systemd-logind[1587]: New session 64 of user core. Mar 2 13:47:43.074229 systemd[1]: Started session-64.scope - Session 64 of User core. Mar 2 13:47:43.563893 sshd[7260]: Connection closed by 10.0.0.1 port 32946 Mar 2 13:47:43.565344 sshd-session[7256]: pam_unix(sshd:session): session closed for user core Mar 2 13:47:43.615959 systemd[1]: sshd@62-10.0.0.31:22-10.0.0.1:32946.service: Deactivated successfully. Mar 2 13:47:43.632118 systemd[1]: session-64.scope: Deactivated successfully. Mar 2 13:47:43.646713 systemd-logind[1587]: Session 64 logged out. Waiting for processes to exit. Mar 2 13:47:43.655844 systemd-logind[1587]: Removed session 64.