Mar 6 01:32:01.373009 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Mar 5 23:31:42 -00 2026 Mar 6 01:32:01.373029 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a6bcd99e714cc2f1b95dc0d61d9d762252de26a434f12074c16f59200c97ba9c Mar 6 01:32:01.373041 kernel: BIOS-provided physical RAM map: Mar 6 01:32:01.373047 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Mar 6 01:32:01.373053 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Mar 6 01:32:01.373058 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Mar 6 01:32:01.373065 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Mar 6 01:32:01.373071 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Mar 6 01:32:01.373077 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Mar 6 01:32:01.373085 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Mar 6 01:32:01.373091 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 6 01:32:01.373097 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Mar 6 01:32:01.373132 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 6 01:32:01.373139 kernel: NX (Execute Disable) protection: active Mar 6 01:32:01.373146 kernel: APIC: Static calls initialized Mar 6 01:32:01.373177 kernel: SMBIOS 2.8 present. Mar 6 01:32:01.373183 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Mar 6 01:32:01.373190 kernel: Hypervisor detected: KVM Mar 6 01:32:01.373196 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 6 01:32:01.373202 kernel: kvm-clock: using sched offset of 8482115001 cycles Mar 6 01:32:01.373208 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 6 01:32:01.373215 kernel: tsc: Detected 2445.424 MHz processor Mar 6 01:32:01.373221 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 6 01:32:01.373227 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 6 01:32:01.373237 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Mar 6 01:32:01.373243 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Mar 6 01:32:01.373250 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 6 01:32:01.373256 kernel: Using GB pages for direct mapping Mar 6 01:32:01.373262 kernel: ACPI: Early table checksum verification disabled Mar 6 01:32:01.373268 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Mar 6 01:32:01.373274 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373280 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373286 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373295 kernel: ACPI: FACS 0x000000009CFE0000 000040 Mar 6 01:32:01.373302 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373308 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373314 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373320 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 6 01:32:01.373326 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Mar 6 01:32:01.373333 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Mar 6 01:32:01.373343 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Mar 6 01:32:01.373353 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Mar 6 01:32:01.373359 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Mar 6 01:32:01.373365 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Mar 6 01:32:01.373372 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Mar 6 01:32:01.373378 kernel: No NUMA configuration found Mar 6 01:32:01.373385 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Mar 6 01:32:01.373391 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Mar 6 01:32:01.373401 kernel: Zone ranges: Mar 6 01:32:01.373407 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 6 01:32:01.373413 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Mar 6 01:32:01.373420 kernel: Normal empty Mar 6 01:32:01.373426 kernel: Movable zone start for each node Mar 6 01:32:01.373432 kernel: Early memory node ranges Mar 6 01:32:01.373439 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Mar 6 01:32:01.373445 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Mar 6 01:32:01.373452 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Mar 6 01:32:01.373461 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 6 01:32:01.373488 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Mar 6 01:32:01.373495 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Mar 6 01:32:01.373502 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 6 01:32:01.373579 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 6 01:32:01.373588 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 6 01:32:01.373594 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 6 01:32:01.373601 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 6 01:32:01.373607 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 6 01:32:01.373618 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 6 01:32:01.373625 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 6 01:32:01.373631 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 6 01:32:01.373637 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 6 01:32:01.373644 kernel: TSC deadline timer available Mar 6 01:32:01.373650 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Mar 6 01:32:01.373657 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 6 01:32:01.373663 kernel: kvm-guest: KVM setup pv remote TLB flush Mar 6 01:32:01.373693 kernel: kvm-guest: setup PV sched yield Mar 6 01:32:01.373704 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Mar 6 01:32:01.373711 kernel: Booting paravirtualized kernel on KVM Mar 6 01:32:01.373717 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 6 01:32:01.373724 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Mar 6 01:32:01.373730 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u524288 Mar 6 01:32:01.373737 kernel: pcpu-alloc: s196328 r8192 d28952 u524288 alloc=1*2097152 Mar 6 01:32:01.373743 kernel: pcpu-alloc: [0] 0 1 2 3 Mar 6 01:32:01.373749 kernel: kvm-guest: PV spinlocks enabled Mar 6 01:32:01.373756 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 6 01:32:01.373766 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a6bcd99e714cc2f1b95dc0d61d9d762252de26a434f12074c16f59200c97ba9c Mar 6 01:32:01.373773 kernel: random: crng init done Mar 6 01:32:01.373780 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 6 01:32:01.373786 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 6 01:32:01.373793 kernel: Fallback order for Node 0: 0 Mar 6 01:32:01.373799 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Mar 6 01:32:01.373805 kernel: Policy zone: DMA32 Mar 6 01:32:01.373812 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 6 01:32:01.373822 kernel: Memory: 2434608K/2571752K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42892K init, 2304K bss, 136884K reserved, 0K cma-reserved) Mar 6 01:32:01.373828 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 6 01:32:01.373835 kernel: ftrace: allocating 37996 entries in 149 pages Mar 6 01:32:01.373841 kernel: ftrace: allocated 149 pages with 4 groups Mar 6 01:32:01.373847 kernel: Dynamic Preempt: voluntary Mar 6 01:32:01.373854 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 6 01:32:01.373861 kernel: rcu: RCU event tracing is enabled. Mar 6 01:32:01.373868 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 6 01:32:01.373874 kernel: Trampoline variant of Tasks RCU enabled. Mar 6 01:32:01.373884 kernel: Rude variant of Tasks RCU enabled. Mar 6 01:32:01.373890 kernel: Tracing variant of Tasks RCU enabled. Mar 6 01:32:01.373897 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 6 01:32:01.373903 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 6 01:32:01.373930 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Mar 6 01:32:01.373937 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 6 01:32:01.373943 kernel: Console: colour VGA+ 80x25 Mar 6 01:32:01.373950 kernel: printk: console [ttyS0] enabled Mar 6 01:32:01.373956 kernel: ACPI: Core revision 20230628 Mar 6 01:32:01.373963 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 6 01:32:01.373973 kernel: APIC: Switch to symmetric I/O mode setup Mar 6 01:32:01.373979 kernel: x2apic enabled Mar 6 01:32:01.373986 kernel: APIC: Switched APIC routing to: physical x2apic Mar 6 01:32:01.373992 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Mar 6 01:32:01.373999 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Mar 6 01:32:01.374005 kernel: kvm-guest: setup PV IPIs Mar 6 01:32:01.374012 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 6 01:32:01.374030 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Mar 6 01:32:01.374037 kernel: Calibrating delay loop (skipped) preset value.. 4890.84 BogoMIPS (lpj=2445424) Mar 6 01:32:01.374044 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 6 01:32:01.374051 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 6 01:32:01.374060 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 6 01:32:01.374067 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 6 01:32:01.374074 kernel: Spectre V2 : Mitigation: Retpolines Mar 6 01:32:01.374081 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Mar 6 01:32:01.374087 kernel: Speculative Store Bypass: Vulnerable Mar 6 01:32:01.374097 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Mar 6 01:32:01.374123 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Mar 6 01:32:01.374131 kernel: active return thunk: srso_alias_return_thunk Mar 6 01:32:01.374137 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Mar 6 01:32:01.374144 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Mar 6 01:32:01.374151 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Mar 6 01:32:01.374158 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 6 01:32:01.374165 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 6 01:32:01.374171 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 6 01:32:01.374182 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 6 01:32:01.374188 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 6 01:32:01.374195 kernel: Freeing SMP alternatives memory: 32K Mar 6 01:32:01.374202 kernel: pid_max: default: 32768 minimum: 301 Mar 6 01:32:01.374209 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 6 01:32:01.374216 kernel: landlock: Up and running. Mar 6 01:32:01.374222 kernel: SELinux: Initializing. Mar 6 01:32:01.374229 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 6 01:32:01.374236 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 6 01:32:01.374245 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Mar 6 01:32:01.374252 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 6 01:32:01.374259 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 6 01:32:01.374266 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 6 01:32:01.374273 kernel: Performance Events: PMU not available due to virtualization, using software events only. Mar 6 01:32:01.374280 kernel: signal: max sigframe size: 1776 Mar 6 01:32:01.374304 kernel: rcu: Hierarchical SRCU implementation. Mar 6 01:32:01.374311 kernel: rcu: Max phase no-delay instances is 400. Mar 6 01:32:01.374321 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 6 01:32:01.374328 kernel: smp: Bringing up secondary CPUs ... Mar 6 01:32:01.374335 kernel: smpboot: x86: Booting SMP configuration: Mar 6 01:32:01.374341 kernel: .... node #0, CPUs: #1 #2 #3 Mar 6 01:32:01.374348 kernel: smp: Brought up 1 node, 4 CPUs Mar 6 01:32:01.374355 kernel: smpboot: Max logical packages: 1 Mar 6 01:32:01.374361 kernel: smpboot: Total of 4 processors activated (19563.39 BogoMIPS) Mar 6 01:32:01.374368 kernel: devtmpfs: initialized Mar 6 01:32:01.374375 kernel: x86/mm: Memory block size: 128MB Mar 6 01:32:01.374381 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 6 01:32:01.374391 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 6 01:32:01.374398 kernel: pinctrl core: initialized pinctrl subsystem Mar 6 01:32:01.374405 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 6 01:32:01.374411 kernel: audit: initializing netlink subsys (disabled) Mar 6 01:32:01.374418 kernel: audit: type=2000 audit(1772760718.369:1): state=initialized audit_enabled=0 res=1 Mar 6 01:32:01.374425 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 6 01:32:01.374432 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 6 01:32:01.374438 kernel: cpuidle: using governor menu Mar 6 01:32:01.374448 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 6 01:32:01.374455 kernel: dca service started, version 1.12.1 Mar 6 01:32:01.374461 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Mar 6 01:32:01.374468 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Mar 6 01:32:01.374475 kernel: PCI: Using configuration type 1 for base access Mar 6 01:32:01.374482 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 6 01:32:01.374489 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 6 01:32:01.374495 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 6 01:32:01.374502 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 6 01:32:01.374582 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 6 01:32:01.374590 kernel: ACPI: Added _OSI(Module Device) Mar 6 01:32:01.374597 kernel: ACPI: Added _OSI(Processor Device) Mar 6 01:32:01.374604 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 6 01:32:01.374610 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 6 01:32:01.374617 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 6 01:32:01.374624 kernel: ACPI: Interpreter enabled Mar 6 01:32:01.374631 kernel: ACPI: PM: (supports S0 S3 S5) Mar 6 01:32:01.374638 kernel: ACPI: Using IOAPIC for interrupt routing Mar 6 01:32:01.374649 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 6 01:32:01.374656 kernel: PCI: Using E820 reservations for host bridge windows Mar 6 01:32:01.374663 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 6 01:32:01.374669 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 6 01:32:01.375154 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 6 01:32:01.375329 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 6 01:32:01.375483 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 6 01:32:01.375493 kernel: PCI host bridge to bus 0000:00 Mar 6 01:32:01.375960 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 6 01:32:01.376214 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 6 01:32:01.376473 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 6 01:32:01.376782 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Mar 6 01:32:01.376969 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Mar 6 01:32:01.377133 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Mar 6 01:32:01.377280 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 6 01:32:01.377727 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Mar 6 01:32:01.377970 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Mar 6 01:32:01.378152 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Mar 6 01:32:01.378749 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Mar 6 01:32:01.378991 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Mar 6 01:32:01.379213 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 6 01:32:01.379451 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Mar 6 01:32:01.379757 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Mar 6 01:32:01.379916 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Mar 6 01:32:01.380063 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Mar 6 01:32:01.380335 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Mar 6 01:32:01.380487 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Mar 6 01:32:01.380765 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Mar 6 01:32:01.380927 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Mar 6 01:32:01.381148 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Mar 6 01:32:01.381301 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Mar 6 01:32:01.381446 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Mar 6 01:32:01.381702 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Mar 6 01:32:01.381889 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Mar 6 01:32:01.382160 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Mar 6 01:32:01.382320 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 6 01:32:01.382785 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Mar 6 01:32:01.382939 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Mar 6 01:32:01.383084 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Mar 6 01:32:01.383342 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Mar 6 01:32:01.383492 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Mar 6 01:32:01.383507 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 6 01:32:01.383585 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 6 01:32:01.383592 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 6 01:32:01.383599 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 6 01:32:01.383606 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 6 01:32:01.383613 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 6 01:32:01.383620 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 6 01:32:01.383627 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 6 01:32:01.383633 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 6 01:32:01.383644 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 6 01:32:01.383651 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 6 01:32:01.383658 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 6 01:32:01.383665 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 6 01:32:01.383671 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 6 01:32:01.383678 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 6 01:32:01.383685 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 6 01:32:01.383694 kernel: iommu: Default domain type: Translated Mar 6 01:32:01.383701 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 6 01:32:01.383711 kernel: PCI: Using ACPI for IRQ routing Mar 6 01:32:01.383718 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 6 01:32:01.383725 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Mar 6 01:32:01.383732 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Mar 6 01:32:01.383890 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 6 01:32:01.384036 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 6 01:32:01.384181 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 6 01:32:01.384190 kernel: vgaarb: loaded Mar 6 01:32:01.384202 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 6 01:32:01.384209 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 6 01:32:01.384216 kernel: clocksource: Switched to clocksource kvm-clock Mar 6 01:32:01.384223 kernel: VFS: Disk quotas dquot_6.6.0 Mar 6 01:32:01.384230 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 6 01:32:01.384237 kernel: pnp: PnP ACPI init Mar 6 01:32:01.384498 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Mar 6 01:32:01.384576 kernel: pnp: PnP ACPI: found 6 devices Mar 6 01:32:01.384585 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 6 01:32:01.384597 kernel: NET: Registered PF_INET protocol family Mar 6 01:32:01.384604 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 6 01:32:01.384610 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 6 01:32:01.384617 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 6 01:32:01.384624 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 6 01:32:01.384631 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 6 01:32:01.384638 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 6 01:32:01.384645 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 6 01:32:01.384655 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 6 01:32:01.384662 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 6 01:32:01.384669 kernel: NET: Registered PF_XDP protocol family Mar 6 01:32:01.384816 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 6 01:32:01.384953 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 6 01:32:01.385087 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 6 01:32:01.385221 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Mar 6 01:32:01.385413 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Mar 6 01:32:01.385707 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Mar 6 01:32:01.385726 kernel: PCI: CLS 0 bytes, default 64 Mar 6 01:32:01.385733 kernel: Initialise system trusted keyrings Mar 6 01:32:01.385741 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 6 01:32:01.385747 kernel: Key type asymmetric registered Mar 6 01:32:01.385754 kernel: Asymmetric key parser 'x509' registered Mar 6 01:32:01.385761 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 6 01:32:01.385768 kernel: io scheduler mq-deadline registered Mar 6 01:32:01.385774 kernel: io scheduler kyber registered Mar 6 01:32:01.385781 kernel: io scheduler bfq registered Mar 6 01:32:01.385791 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 6 01:32:01.385798 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 6 01:32:01.385805 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 6 01:32:01.385812 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 6 01:32:01.385819 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 6 01:32:01.385855 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 6 01:32:01.385863 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 6 01:32:01.385870 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 6 01:32:01.385877 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 6 01:32:01.386127 kernel: rtc_cmos 00:04: RTC can wake from S4 Mar 6 01:32:01.386141 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 6 01:32:01.386284 kernel: rtc_cmos 00:04: registered as rtc0 Mar 6 01:32:01.386504 kernel: rtc_cmos 00:04: setting system clock to 2026-03-06T01:32:00 UTC (1772760720) Mar 6 01:32:01.386721 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Mar 6 01:32:01.386731 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 6 01:32:01.386738 kernel: NET: Registered PF_INET6 protocol family Mar 6 01:32:01.386745 kernel: Segment Routing with IPv6 Mar 6 01:32:01.386757 kernel: In-situ OAM (IOAM) with IPv6 Mar 6 01:32:01.386764 kernel: NET: Registered PF_PACKET protocol family Mar 6 01:32:01.386771 kernel: Key type dns_resolver registered Mar 6 01:32:01.386777 kernel: IPI shorthand broadcast: enabled Mar 6 01:32:01.386784 kernel: sched_clock: Marking stable (2407016482, 430707076)->(3281151996, -443428438) Mar 6 01:32:01.386791 kernel: registered taskstats version 1 Mar 6 01:32:01.386798 kernel: Loading compiled-in X.509 certificates Mar 6 01:32:01.386805 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 6d88f6264570591a57b3c9c1e1c99fca6c68b8ca' Mar 6 01:32:01.386812 kernel: Key type .fscrypt registered Mar 6 01:32:01.386822 kernel: Key type fscrypt-provisioning registered Mar 6 01:32:01.386829 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 6 01:32:01.386836 kernel: ima: Allocated hash algorithm: sha1 Mar 6 01:32:01.386842 kernel: ima: No architecture policies found Mar 6 01:32:01.386849 kernel: clk: Disabling unused clocks Mar 6 01:32:01.386856 kernel: Freeing unused kernel image (initmem) memory: 42892K Mar 6 01:32:01.386863 kernel: Write protecting the kernel read-only data: 36864k Mar 6 01:32:01.386869 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Mar 6 01:32:01.386879 kernel: Run /init as init process Mar 6 01:32:01.386886 kernel: with arguments: Mar 6 01:32:01.386893 kernel: /init Mar 6 01:32:01.386899 kernel: with environment: Mar 6 01:32:01.386906 kernel: HOME=/ Mar 6 01:32:01.386913 kernel: TERM=linux Mar 6 01:32:01.386921 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 6 01:32:01.386930 systemd[1]: Detected virtualization kvm. Mar 6 01:32:01.386941 systemd[1]: Detected architecture x86-64. Mar 6 01:32:01.386947 systemd[1]: Running in initrd. Mar 6 01:32:01.386954 systemd[1]: No hostname configured, using default hostname. Mar 6 01:32:01.386961 systemd[1]: Hostname set to . Mar 6 01:32:01.386969 systemd[1]: Initializing machine ID from VM UUID. Mar 6 01:32:01.386976 systemd[1]: Queued start job for default target initrd.target. Mar 6 01:32:01.386983 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 6 01:32:01.386990 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 6 01:32:01.387002 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 6 01:32:01.387009 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 6 01:32:01.387016 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 6 01:32:01.387024 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 6 01:32:01.387032 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 6 01:32:01.387039 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 6 01:32:01.387047 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 6 01:32:01.387057 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 6 01:32:01.387064 systemd[1]: Reached target paths.target - Path Units. Mar 6 01:32:01.387071 systemd[1]: Reached target slices.target - Slice Units. Mar 6 01:32:01.387079 systemd[1]: Reached target swap.target - Swaps. Mar 6 01:32:01.387101 systemd[1]: Reached target timers.target - Timer Units. Mar 6 01:32:01.387111 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 6 01:32:01.387121 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 6 01:32:01.387129 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 6 01:32:01.387136 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 6 01:32:01.387144 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 6 01:32:01.387151 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 6 01:32:01.387158 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 6 01:32:01.387166 systemd[1]: Reached target sockets.target - Socket Units. Mar 6 01:32:01.387173 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 6 01:32:01.387181 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 6 01:32:01.387191 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 6 01:32:01.387198 systemd[1]: Starting systemd-fsck-usr.service... Mar 6 01:32:01.387206 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 6 01:32:01.387213 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 6 01:32:01.387220 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 01:32:01.387250 systemd-journald[195]: Collecting audit messages is disabled. Mar 6 01:32:01.387271 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 6 01:32:01.387279 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 6 01:32:01.387287 systemd[1]: Finished systemd-fsck-usr.service. Mar 6 01:32:01.387295 systemd-journald[195]: Journal started Mar 6 01:32:01.387313 systemd-journald[195]: Runtime Journal (/run/log/journal/57dfbeab1899465b9907bfed1f139555) is 6.0M, max 48.4M, 42.3M free. Mar 6 01:32:01.394601 systemd[1]: Started systemd-journald.service - Journal Service. Mar 6 01:32:01.395176 systemd-modules-load[196]: Inserted module 'overlay' Mar 6 01:32:01.558027 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 6 01:32:01.558090 kernel: Bridge firewalling registered Mar 6 01:32:01.428945 systemd-modules-load[196]: Inserted module 'br_netfilter' Mar 6 01:32:01.561655 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 6 01:32:01.569505 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 01:32:01.594030 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 6 01:32:01.595859 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 6 01:32:01.600762 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 6 01:32:01.604951 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 6 01:32:01.620710 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 6 01:32:01.633871 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 6 01:32:01.643113 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 6 01:32:01.648439 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 6 01:32:01.650989 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 6 01:32:01.667015 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 6 01:32:01.670729 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 6 01:32:01.682910 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 6 01:32:01.692100 systemd-resolved[227]: Positive Trust Anchors: Mar 6 01:32:01.692137 systemd-resolved[227]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 6 01:32:01.702593 dracut-cmdline[231]: dracut-dracut-053 Mar 6 01:32:01.702593 dracut-cmdline[231]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a6bcd99e714cc2f1b95dc0d61d9d762252de26a434f12074c16f59200c97ba9c Mar 6 01:32:01.692164 systemd-resolved[227]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 6 01:32:01.694934 systemd-resolved[227]: Defaulting to hostname 'linux'. Mar 6 01:32:01.697149 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 6 01:32:01.702623 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 6 01:32:01.804670 kernel: SCSI subsystem initialized Mar 6 01:32:01.815651 kernel: Loading iSCSI transport class v2.0-870. Mar 6 01:32:01.830638 kernel: iscsi: registered transport (tcp) Mar 6 01:32:01.855442 kernel: iscsi: registered transport (qla4xxx) Mar 6 01:32:01.855583 kernel: QLogic iSCSI HBA Driver Mar 6 01:32:01.922271 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 6 01:32:01.936774 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 6 01:32:01.977837 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 6 01:32:01.977876 kernel: device-mapper: uevent: version 1.0.3 Mar 6 01:32:01.981113 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 6 01:32:02.030774 kernel: raid6: avx2x4 gen() 27249 MB/s Mar 6 01:32:02.049677 kernel: raid6: avx2x2 gen() 25088 MB/s Mar 6 01:32:02.069276 kernel: raid6: avx2x1 gen() 21379 MB/s Mar 6 01:32:02.069326 kernel: raid6: using algorithm avx2x4 gen() 27249 MB/s Mar 6 01:32:02.089631 kernel: raid6: .... xor() 4447 MB/s, rmw enabled Mar 6 01:32:02.089710 kernel: raid6: using avx2x2 recovery algorithm Mar 6 01:32:02.111662 kernel: xor: automatically using best checksumming function avx Mar 6 01:32:02.317742 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 6 01:32:02.337973 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 6 01:32:02.365847 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 6 01:32:02.387904 systemd-udevd[414]: Using default interface naming scheme 'v255'. Mar 6 01:32:02.393436 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 6 01:32:02.414833 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 6 01:32:02.434065 dracut-pre-trigger[427]: rd.md=0: removing MD RAID activation Mar 6 01:32:02.479055 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 6 01:32:02.504849 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 6 01:32:02.675117 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 6 01:32:02.685830 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 6 01:32:02.707704 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 6 01:32:02.717251 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 6 01:32:02.725920 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 6 01:32:02.729695 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 6 01:32:02.753832 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 6 01:32:02.768689 kernel: cryptd: max_cpu_qlen set to 1000 Mar 6 01:32:02.774640 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 6 01:32:02.785977 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Mar 6 01:32:02.790073 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 6 01:32:02.808192 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 6 01:32:02.793008 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 6 01:32:02.801919 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 6 01:32:02.826332 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 6 01:32:02.826357 kernel: GPT:9289727 != 19775487 Mar 6 01:32:02.826369 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 6 01:32:02.817916 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 6 01:32:02.837845 kernel: GPT:9289727 != 19775487 Mar 6 01:32:02.837863 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 6 01:32:02.837874 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 01:32:02.818125 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 01:32:02.848739 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 01:32:02.868694 kernel: AVX2 version of gcm_enc/dec engaged. Mar 6 01:32:02.868744 kernel: libata version 3.00 loaded. Mar 6 01:32:02.868762 kernel: AES CTR mode by8 optimization enabled Mar 6 01:32:02.873919 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 01:32:02.893680 kernel: ahci 0000:00:1f.2: version 3.0 Mar 6 01:32:02.897618 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 6 01:32:02.911951 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Mar 6 01:32:02.912259 kernel: BTRFS: device fsid eccec0b1-0068-4620-ab61-f332f16460fa devid 1 transid 35 /dev/vda3 scanned by (udev-worker) (477) Mar 6 01:32:02.912278 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 6 01:32:02.923906 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (475) Mar 6 01:32:02.923953 kernel: scsi host0: ahci Mar 6 01:32:02.927635 kernel: scsi host1: ahci Mar 6 01:32:02.930709 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 6 01:32:03.157432 kernel: scsi host2: ahci Mar 6 01:32:03.157955 kernel: scsi host3: ahci Mar 6 01:32:03.158197 kernel: scsi host4: ahci Mar 6 01:32:03.158385 kernel: scsi host5: ahci Mar 6 01:32:03.158681 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Mar 6 01:32:03.158704 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Mar 6 01:32:03.158723 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Mar 6 01:32:03.158740 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Mar 6 01:32:03.158751 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Mar 6 01:32:03.158761 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Mar 6 01:32:02.975599 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 6 01:32:03.161868 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 6 01:32:03.173249 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 6 01:32:03.185856 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 6 01:32:03.203845 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 6 01:32:03.208188 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 01:32:03.224277 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 01:32:03.214322 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 6 01:32:03.232709 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 01:32:03.232738 disk-uuid[554]: Primary Header is updated. Mar 6 01:32:03.232738 disk-uuid[554]: Secondary Entries is updated. Mar 6 01:32:03.232738 disk-uuid[554]: Secondary Header is updated. Mar 6 01:32:03.274689 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 01:32:03.274723 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 6 01:32:03.274741 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 6 01:32:03.274771 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 6 01:32:03.274787 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 6 01:32:03.274803 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 6 01:32:03.274820 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 6 01:32:03.280233 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 6 01:32:03.280263 kernel: ata3.00: applying bridge limits Mar 6 01:32:03.281265 kernel: ata3.00: configured for UDMA/100 Mar 6 01:32:03.288822 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 6 01:32:03.317378 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 6 01:32:03.369834 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 6 01:32:03.370191 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 6 01:32:03.388654 kernel: hrtimer: interrupt took 6278343 ns Mar 6 01:32:03.396805 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Mar 6 01:32:04.249790 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 6 01:32:04.251793 disk-uuid[557]: The operation has completed successfully. Mar 6 01:32:04.293373 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 6 01:32:04.293629 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 6 01:32:04.310922 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 6 01:32:04.318302 sh[597]: Success Mar 6 01:32:04.335690 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Mar 6 01:32:04.399666 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 6 01:32:04.419489 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 6 01:32:04.423743 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 6 01:32:04.452433 kernel: BTRFS info (device dm-0): first mount of filesystem eccec0b1-0068-4620-ab61-f332f16460fa Mar 6 01:32:04.452493 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 6 01:32:04.452879 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 6 01:32:04.461241 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 6 01:32:04.463780 kernel: BTRFS info (device dm-0): using free space tree Mar 6 01:32:04.476199 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 6 01:32:04.482806 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 6 01:32:04.494829 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 6 01:32:04.498686 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 6 01:32:04.519716 kernel: BTRFS info (device vda6): first mount of filesystem dcd455b6-671f-4d9f-a5ce-de07977c88a5 Mar 6 01:32:04.519748 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 6 01:32:04.519773 kernel: BTRFS info (device vda6): using free space tree Mar 6 01:32:04.525607 kernel: BTRFS info (device vda6): auto enabling async discard Mar 6 01:32:04.540116 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 6 01:32:04.552638 kernel: BTRFS info (device vda6): last unmount of filesystem dcd455b6-671f-4d9f-a5ce-de07977c88a5 Mar 6 01:32:04.563456 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 6 01:32:04.574790 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 6 01:32:04.779769 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 6 01:32:04.779950 ignition[697]: Ignition 2.19.0 Mar 6 01:32:04.779959 ignition[697]: Stage: fetch-offline Mar 6 01:32:04.792852 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 6 01:32:04.780076 ignition[697]: no configs at "/usr/lib/ignition/base.d" Mar 6 01:32:04.780090 ignition[697]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 01:32:04.780355 ignition[697]: parsed url from cmdline: "" Mar 6 01:32:04.780360 ignition[697]: no config URL provided Mar 6 01:32:04.780367 ignition[697]: reading system config file "/usr/lib/ignition/user.ign" Mar 6 01:32:04.780378 ignition[697]: no config at "/usr/lib/ignition/user.ign" Mar 6 01:32:04.780504 ignition[697]: op(1): [started] loading QEMU firmware config module Mar 6 01:32:04.780511 ignition[697]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 6 01:32:04.900713 ignition[697]: op(1): [finished] loading QEMU firmware config module Mar 6 01:32:04.939247 systemd-networkd[784]: lo: Link UP Mar 6 01:32:04.939299 systemd-networkd[784]: lo: Gained carrier Mar 6 01:32:04.943107 systemd-networkd[784]: Enumeration completed Mar 6 01:32:04.946104 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 6 01:32:04.946615 systemd-networkd[784]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 01:32:04.946622 systemd-networkd[784]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 6 01:32:04.948091 systemd-networkd[784]: eth0: Link UP Mar 6 01:32:04.948098 systemd-networkd[784]: eth0: Gained carrier Mar 6 01:32:04.948108 systemd-networkd[784]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 01:32:04.957995 systemd[1]: Reached target network.target - Network. Mar 6 01:32:04.992658 systemd-networkd[784]: eth0: DHCPv4 address 10.0.0.56/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 6 01:32:05.155412 ignition[697]: parsing config with SHA512: 3e6feefb452f59e47e083057002e4b5c6161cea9a5d7294e39de43d5bc9b8291aa3c693c2779342498edc556501715c9fef9e36826e1967d489ae45185fc400e Mar 6 01:32:05.187356 unknown[697]: fetched base config from "system" Mar 6 01:32:05.187378 unknown[697]: fetched user config from "qemu" Mar 6 01:32:05.191503 ignition[697]: fetch-offline: fetch-offline passed Mar 6 01:32:05.191926 ignition[697]: Ignition finished successfully Mar 6 01:32:05.194776 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 6 01:32:05.202012 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 6 01:32:05.214861 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 6 01:32:05.263902 ignition[789]: Ignition 2.19.0 Mar 6 01:32:05.263939 ignition[789]: Stage: kargs Mar 6 01:32:05.264210 ignition[789]: no configs at "/usr/lib/ignition/base.d" Mar 6 01:32:05.271196 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 6 01:32:05.264229 ignition[789]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 01:32:05.265795 ignition[789]: kargs: kargs passed Mar 6 01:32:05.265862 ignition[789]: Ignition finished successfully Mar 6 01:32:05.289858 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 6 01:32:05.423808 ignition[797]: Ignition 2.19.0 Mar 6 01:32:05.423855 ignition[797]: Stage: disks Mar 6 01:32:05.424212 ignition[797]: no configs at "/usr/lib/ignition/base.d" Mar 6 01:32:05.428957 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 6 01:32:05.424230 ignition[797]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 01:32:05.434732 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 6 01:32:05.426143 ignition[797]: disks: disks passed Mar 6 01:32:05.441195 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 6 01:32:05.426219 ignition[797]: Ignition finished successfully Mar 6 01:32:05.448942 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 6 01:32:05.449149 systemd[1]: Reached target sysinit.target - System Initialization. Mar 6 01:32:05.452203 systemd[1]: Reached target basic.target - Basic System. Mar 6 01:32:05.473923 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 6 01:32:05.501282 systemd-fsck[807]: ROOT: clean, 14/553520 files, 52654/553472 blocks Mar 6 01:32:05.507192 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 6 01:32:05.528732 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 6 01:32:05.737673 kernel: EXT4-fs (vda9): mounted filesystem 6fb83788-0471-4e89-b45f-3a7586a627a9 r/w with ordered data mode. Quota mode: none. Mar 6 01:32:05.739327 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 6 01:32:05.754361 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 6 01:32:05.781784 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 6 01:32:05.796362 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 6 01:32:05.822910 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (815) Mar 6 01:32:05.822975 kernel: BTRFS info (device vda6): first mount of filesystem dcd455b6-671f-4d9f-a5ce-de07977c88a5 Mar 6 01:32:05.822996 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 6 01:32:05.823016 kernel: BTRFS info (device vda6): using free space tree Mar 6 01:32:05.823033 kernel: BTRFS info (device vda6): auto enabling async discard Mar 6 01:32:05.811474 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 6 01:32:05.811631 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 6 01:32:05.811672 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 6 01:32:05.832404 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 6 01:32:05.866711 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 6 01:32:05.867090 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 6 01:32:06.066437 initrd-setup-root[839]: cut: /sysroot/etc/passwd: No such file or directory Mar 6 01:32:06.073366 initrd-setup-root[846]: cut: /sysroot/etc/group: No such file or directory Mar 6 01:32:06.079385 initrd-setup-root[853]: cut: /sysroot/etc/shadow: No such file or directory Mar 6 01:32:06.085048 initrd-setup-root[860]: cut: /sysroot/etc/gshadow: No such file or directory Mar 6 01:32:06.258913 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 6 01:32:06.273907 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 6 01:32:06.281774 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 6 01:32:06.285442 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 6 01:32:06.295184 kernel: BTRFS info (device vda6): last unmount of filesystem dcd455b6-671f-4d9f-a5ce-de07977c88a5 Mar 6 01:32:06.358606 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 6 01:32:06.366627 ignition[927]: INFO : Ignition 2.19.0 Mar 6 01:32:06.366627 ignition[927]: INFO : Stage: mount Mar 6 01:32:06.366627 ignition[927]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 6 01:32:06.366627 ignition[927]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 01:32:06.366627 ignition[927]: INFO : mount: mount passed Mar 6 01:32:06.366627 ignition[927]: INFO : Ignition finished successfully Mar 6 01:32:06.373969 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 6 01:32:06.397836 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 6 01:32:06.507065 systemd-networkd[784]: eth0: Gained IPv6LL Mar 6 01:32:06.769884 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 6 01:32:06.782662 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (940) Mar 6 01:32:06.782709 kernel: BTRFS info (device vda6): first mount of filesystem dcd455b6-671f-4d9f-a5ce-de07977c88a5 Mar 6 01:32:06.788405 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 6 01:32:06.788443 kernel: BTRFS info (device vda6): using free space tree Mar 6 01:32:06.798663 kernel: BTRFS info (device vda6): auto enabling async discard Mar 6 01:32:06.800051 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 6 01:32:06.839106 ignition[958]: INFO : Ignition 2.19.0 Mar 6 01:32:06.839106 ignition[958]: INFO : Stage: files Mar 6 01:32:06.844747 ignition[958]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 6 01:32:06.844747 ignition[958]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 01:32:06.844747 ignition[958]: DEBUG : files: compiled without relabeling support, skipping Mar 6 01:32:06.844747 ignition[958]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 6 01:32:06.844747 ignition[958]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 6 01:32:06.870721 ignition[958]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 6 01:32:06.870721 ignition[958]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 6 01:32:06.870721 ignition[958]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 6 01:32:06.870721 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 6 01:32:06.870721 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 6 01:32:06.870721 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 6 01:32:06.870721 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 6 01:32:06.848703 unknown[958]: wrote ssh authorized keys file for user: core Mar 6 01:32:06.926830 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 6 01:32:07.242792 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 6 01:32:07.242792 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 6 01:32:07.254378 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Mar 6 01:32:07.446484 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Mar 6 01:32:07.729692 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 6 01:32:07.729692 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 6 01:32:07.743140 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-x86-64.raw: attempt #1 Mar 6 01:32:08.053007 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Mar 6 01:32:09.184006 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 6 01:32:09.184006 ignition[958]: INFO : files: op(d): [started] processing unit "containerd.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(d): [finished] processing unit "containerd.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(11): [started] processing unit "coreos-metadata.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(11): op(12): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(11): op(12): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(11): [finished] processing unit "coreos-metadata.service" Mar 6 01:32:09.203291 ignition[958]: INFO : files: op(13): [started] setting preset to disabled for "coreos-metadata.service" Mar 6 01:32:09.279935 ignition[958]: INFO : files: op(13): op(14): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 6 01:32:09.279935 ignition[958]: INFO : files: op(13): op(14): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 6 01:32:09.279935 ignition[958]: INFO : files: op(13): [finished] setting preset to disabled for "coreos-metadata.service" Mar 6 01:32:09.279935 ignition[958]: INFO : files: op(15): [started] setting preset to enabled for "prepare-helm.service" Mar 6 01:32:09.279935 ignition[958]: INFO : files: op(15): [finished] setting preset to enabled for "prepare-helm.service" Mar 6 01:32:09.279935 ignition[958]: INFO : files: createResultFile: createFiles: op(16): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 6 01:32:09.279935 ignition[958]: INFO : files: createResultFile: createFiles: op(16): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 6 01:32:09.279935 ignition[958]: INFO : files: files passed Mar 6 01:32:09.279935 ignition[958]: INFO : Ignition finished successfully Mar 6 01:32:09.271350 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 6 01:32:09.299828 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 6 01:32:09.308118 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 6 01:32:09.316367 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 6 01:32:09.356993 initrd-setup-root-after-ignition[984]: grep: /sysroot/oem/oem-release: No such file or directory Mar 6 01:32:09.316511 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 6 01:32:09.364990 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 6 01:32:09.364990 initrd-setup-root-after-ignition[986]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 6 01:32:09.330915 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 6 01:32:09.379183 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 6 01:32:09.337328 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 6 01:32:09.342160 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 6 01:32:09.380344 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 6 01:32:09.380490 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 6 01:32:09.388201 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 6 01:32:09.391712 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 6 01:32:09.394928 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 6 01:32:09.395947 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 6 01:32:09.419355 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 6 01:32:09.424946 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 6 01:32:09.443337 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 6 01:32:09.447887 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 6 01:32:09.451690 systemd[1]: Stopped target timers.target - Timer Units. Mar 6 01:32:09.458023 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 6 01:32:09.458188 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 6 01:32:09.466432 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 6 01:32:09.471764 systemd[1]: Stopped target basic.target - Basic System. Mar 6 01:32:09.477862 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 6 01:32:09.485816 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 6 01:32:09.493722 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 6 01:32:09.500099 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 6 01:32:09.517650 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 6 01:32:09.531444 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 6 01:32:09.532737 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 6 01:32:09.537259 systemd[1]: Stopped target swap.target - Swaps. Mar 6 01:32:09.538429 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 6 01:32:09.538921 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 6 01:32:09.550112 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 6 01:32:09.669742 ignition[1011]: INFO : Ignition 2.19.0 Mar 6 01:32:09.669742 ignition[1011]: INFO : Stage: umount Mar 6 01:32:09.669742 ignition[1011]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 6 01:32:09.669742 ignition[1011]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 6 01:32:09.550290 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 6 01:32:09.696471 ignition[1011]: INFO : umount: umount passed Mar 6 01:32:09.696471 ignition[1011]: INFO : Ignition finished successfully Mar 6 01:32:09.551387 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 6 01:32:09.551773 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 6 01:32:09.551958 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 6 01:32:09.553129 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 6 01:32:09.553996 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 6 01:32:09.554173 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 6 01:32:09.559237 systemd[1]: Stopped target paths.target - Path Units. Mar 6 01:32:09.565284 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 6 01:32:09.568929 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 6 01:32:09.580216 systemd[1]: Stopped target slices.target - Slice Units. Mar 6 01:32:09.581236 systemd[1]: Stopped target sockets.target - Socket Units. Mar 6 01:32:09.582437 systemd[1]: iscsid.socket: Deactivated successfully. Mar 6 01:32:09.582739 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 6 01:32:09.583332 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 6 01:32:09.583468 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 6 01:32:09.584375 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 6 01:32:09.584726 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 6 01:32:09.585355 systemd[1]: ignition-files.service: Deactivated successfully. Mar 6 01:32:09.585483 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 6 01:32:09.622907 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 6 01:32:09.628617 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 6 01:32:09.628866 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 6 01:32:09.638184 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 6 01:32:09.649795 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 6 01:32:09.650021 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 6 01:32:09.660399 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 6 01:32:09.660774 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 6 01:32:09.672669 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 6 01:32:09.672820 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 6 01:32:09.679920 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 6 01:32:09.682324 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 6 01:32:09.682481 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 6 01:32:09.690209 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 6 01:32:09.690408 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 6 01:32:09.698211 systemd[1]: Stopped target network.target - Network. Mar 6 01:32:09.701964 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 6 01:32:09.702056 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 6 01:32:09.707785 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 6 01:32:09.707856 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 6 01:32:09.713889 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 6 01:32:09.714060 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 6 01:32:09.722627 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 6 01:32:09.722707 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 6 01:32:09.729478 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 6 01:32:09.729637 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 6 01:32:09.736420 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 6 01:32:09.743994 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 6 01:32:09.752626 systemd-networkd[784]: eth0: DHCPv6 lease lost Mar 6 01:32:09.755385 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 6 01:32:09.755627 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 6 01:32:09.763256 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 6 01:32:09.763415 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 6 01:32:09.776115 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 6 01:32:09.776171 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 6 01:32:09.805921 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 6 01:32:09.811680 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 6 01:32:09.811812 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 6 01:32:09.820301 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 6 01:32:09.820388 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 6 01:32:09.827507 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 6 01:32:09.827674 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 6 01:32:10.057051 systemd-journald[195]: Received SIGTERM from PID 1 (systemd). Mar 6 01:32:09.831621 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 6 01:32:09.831695 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 6 01:32:09.839786 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 6 01:32:09.867169 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 6 01:32:09.867485 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 6 01:32:09.875064 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 6 01:32:09.875266 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 6 01:32:09.883149 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 6 01:32:09.883246 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 6 01:32:09.889306 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 6 01:32:09.889372 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 6 01:32:09.893366 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 6 01:32:09.893434 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 6 01:32:09.900871 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 6 01:32:09.900929 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 6 01:32:09.908761 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 6 01:32:09.908832 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 6 01:32:09.935783 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 6 01:32:09.945022 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 6 01:32:09.945108 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 6 01:32:09.949750 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 6 01:32:09.949807 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 01:32:09.956963 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 6 01:32:09.957122 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 6 01:32:09.964416 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 6 01:32:09.988942 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 6 01:32:10.003199 systemd[1]: Switching root. Mar 6 01:32:10.156987 systemd-journald[195]: Journal stopped Mar 6 01:32:15.327387 kernel: SELinux: policy capability network_peer_controls=1 Mar 6 01:32:15.327856 kernel: SELinux: policy capability open_perms=1 Mar 6 01:32:15.327888 kernel: SELinux: policy capability extended_socket_class=1 Mar 6 01:32:15.327900 kernel: SELinux: policy capability always_check_network=0 Mar 6 01:32:15.327913 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 6 01:32:15.327933 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 6 01:32:15.327947 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 6 01:32:15.327959 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 6 01:32:15.328034 kernel: audit: type=1403 audit(1772760730.383:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 6 01:32:15.328050 systemd[1]: Successfully loaded SELinux policy in 114.039ms. Mar 6 01:32:15.328116 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 18.302ms. Mar 6 01:32:15.328136 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 6 01:32:15.328148 systemd[1]: Detected virtualization kvm. Mar 6 01:32:15.328188 systemd[1]: Detected architecture x86-64. Mar 6 01:32:15.328201 systemd[1]: Detected first boot. Mar 6 01:32:15.328212 systemd[1]: Initializing machine ID from VM UUID. Mar 6 01:32:15.328224 zram_generator::config[1074]: No configuration found. Mar 6 01:32:15.328238 systemd[1]: Populated /etc with preset unit settings. Mar 6 01:32:15.328281 systemd[1]: Queued start job for default target multi-user.target. Mar 6 01:32:15.328298 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 6 01:32:15.328311 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 6 01:32:15.328323 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 6 01:32:15.328335 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 6 01:32:15.328347 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 6 01:32:15.328360 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 6 01:32:15.328372 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 6 01:32:15.328391 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 6 01:32:15.328434 systemd[1]: Created slice user.slice - User and Session Slice. Mar 6 01:32:15.328447 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 6 01:32:15.328459 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 6 01:32:15.328472 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 6 01:32:15.328484 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 6 01:32:15.328496 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 6 01:32:15.328508 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 6 01:32:15.328520 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 6 01:32:15.328627 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 6 01:32:15.328648 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 6 01:32:15.328662 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 6 01:32:15.328674 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 6 01:32:15.328752 systemd[1]: Reached target slices.target - Slice Units. Mar 6 01:32:15.328764 systemd[1]: Reached target swap.target - Swaps. Mar 6 01:32:15.328776 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 6 01:32:15.328809 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 6 01:32:15.328841 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 6 01:32:15.328877 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 6 01:32:15.328889 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 6 01:32:15.328901 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 6 01:32:15.328912 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 6 01:32:15.328924 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 6 01:32:15.328936 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 6 01:32:15.328948 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 6 01:32:15.328960 systemd[1]: Mounting media.mount - External Media Directory... Mar 6 01:32:15.328971 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:15.328986 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 6 01:32:15.328999 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 6 01:32:15.329011 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 6 01:32:15.329022 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 6 01:32:15.329034 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 01:32:15.329046 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 6 01:32:15.329057 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 6 01:32:15.329070 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 01:32:15.329082 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 6 01:32:15.329097 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 01:32:15.329108 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 6 01:32:15.329120 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 01:32:15.329132 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 6 01:32:15.329169 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Mar 6 01:32:15.329182 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Mar 6 01:32:15.329213 kernel: fuse: init (API version 7.39) Mar 6 01:32:15.329225 kernel: loop: module loaded Mar 6 01:32:15.329260 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 6 01:32:15.329274 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 6 01:32:15.329286 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 6 01:32:15.329298 kernel: ACPI: bus type drm_connector registered Mar 6 01:32:15.329310 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 6 01:32:15.329347 systemd-journald[1174]: Collecting audit messages is disabled. Mar 6 01:32:15.329415 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 6 01:32:15.329433 systemd-journald[1174]: Journal started Mar 6 01:32:15.329473 systemd-journald[1174]: Runtime Journal (/run/log/journal/57dfbeab1899465b9907bfed1f139555) is 6.0M, max 48.4M, 42.3M free. Mar 6 01:32:15.341657 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:15.354023 systemd[1]: Started systemd-journald.service - Journal Service. Mar 6 01:32:15.357140 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 6 01:32:15.361268 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 6 01:32:15.365402 systemd[1]: Mounted media.mount - External Media Directory. Mar 6 01:32:15.368970 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 6 01:32:15.373053 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 6 01:32:15.377303 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 6 01:32:15.381200 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 6 01:32:15.385901 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 6 01:32:15.390787 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 6 01:32:15.391053 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 6 01:32:15.395791 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 01:32:15.396052 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 01:32:15.400455 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 6 01:32:15.400823 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 6 01:32:15.404985 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 01:32:15.405243 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 01:32:15.409999 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 6 01:32:15.410388 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 6 01:32:15.415874 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 01:32:15.416440 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 01:32:15.421169 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 6 01:32:15.426820 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 6 01:32:15.431923 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 6 01:32:15.454479 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 6 01:32:15.470686 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 6 01:32:15.476716 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 6 01:32:15.482157 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 6 01:32:15.486816 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 6 01:32:15.493037 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 6 01:32:15.496957 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 6 01:32:15.500962 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 6 01:32:15.506074 systemd-journald[1174]: Time spent on flushing to /var/log/journal/57dfbeab1899465b9907bfed1f139555 is 13.697ms for 932 entries. Mar 6 01:32:15.506074 systemd-journald[1174]: System Journal (/var/log/journal/57dfbeab1899465b9907bfed1f139555) is 8.0M, max 195.6M, 187.6M free. Mar 6 01:32:15.528681 systemd-journald[1174]: Received client request to flush runtime journal. Mar 6 01:32:15.509882 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 6 01:32:15.521431 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 6 01:32:15.528770 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 6 01:32:15.538293 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 6 01:32:15.544806 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 6 01:32:15.549339 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 6 01:32:15.555096 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 6 01:32:15.560108 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 6 01:32:15.569443 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 6 01:32:15.582958 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 6 01:32:15.587172 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 6 01:32:15.598122 systemd-tmpfiles[1212]: ACLs are not supported, ignoring. Mar 6 01:32:15.598185 systemd-tmpfiles[1212]: ACLs are not supported, ignoring. Mar 6 01:32:15.606916 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 6 01:32:15.613785 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 6 01:32:15.619497 udevadm[1224]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 6 01:32:15.679010 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 6 01:32:15.689961 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 6 01:32:15.724747 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Mar 6 01:32:15.724811 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Mar 6 01:32:15.735323 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 6 01:32:16.047499 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 6 01:32:16.070893 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 6 01:32:16.105070 systemd-udevd[1239]: Using default interface naming scheme 'v255'. Mar 6 01:32:16.132998 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 6 01:32:16.144785 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 6 01:32:16.159779 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 6 01:32:16.182473 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Mar 6 01:32:16.217495 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 6 01:32:16.233891 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1251) Mar 6 01:32:16.299641 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Mar 6 01:32:16.309431 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 6 01:32:16.318659 kernel: ACPI: button: Power Button [PWRF] Mar 6 01:32:16.323615 systemd-networkd[1245]: lo: Link UP Mar 6 01:32:16.323643 systemd-networkd[1245]: lo: Gained carrier Mar 6 01:32:16.339813 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 6 01:32:16.346773 systemd-networkd[1245]: Enumeration completed Mar 6 01:32:16.346935 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 6 01:32:16.348629 systemd-networkd[1245]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 01:32:16.348634 systemd-networkd[1245]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 6 01:32:16.353425 systemd-networkd[1245]: eth0: Link UP Mar 6 01:32:16.353493 systemd-networkd[1245]: eth0: Gained carrier Mar 6 01:32:16.353852 systemd-networkd[1245]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 6 01:32:16.356447 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Mar 6 01:32:16.356821 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 6 01:32:16.359731 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 6 01:32:16.372659 systemd-networkd[1245]: eth0: DHCPv4 address 10.0.0.56/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 6 01:32:16.384600 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Mar 6 01:32:16.390951 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 6 01:32:16.489605 kernel: mousedev: PS/2 mouse device common for all mice Mar 6 01:32:16.509289 kernel: kvm_amd: TSC scaling supported Mar 6 01:32:16.509355 kernel: kvm_amd: Nested Virtualization enabled Mar 6 01:32:16.509371 kernel: kvm_amd: Nested Paging enabled Mar 6 01:32:16.509384 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Mar 6 01:32:16.509396 kernel: kvm_amd: PMU virtualization is disabled Mar 6 01:32:16.569041 kernel: EDAC MC: Ver: 3.0.0 Mar 6 01:32:16.603506 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 6 01:32:16.709086 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 6 01:32:16.728958 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 6 01:32:16.740417 lvm[1285]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 6 01:32:16.776017 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 6 01:32:16.780317 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 6 01:32:16.793810 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 6 01:32:16.800141 lvm[1289]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 6 01:32:16.836285 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 6 01:32:16.841856 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 6 01:32:16.846399 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 6 01:32:16.846476 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 6 01:32:16.850285 systemd[1]: Reached target machines.target - Containers. Mar 6 01:32:16.855233 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 6 01:32:16.872789 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 6 01:32:16.878984 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 6 01:32:16.882441 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 01:32:16.884323 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 6 01:32:16.891819 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 6 01:32:16.898299 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 6 01:32:16.904493 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 6 01:32:16.919598 kernel: loop0: detected capacity change from 0 to 140768 Mar 6 01:32:16.935461 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 6 01:32:16.942968 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 6 01:32:16.944359 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 6 01:32:16.968635 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 6 01:32:17.010668 kernel: loop1: detected capacity change from 0 to 142488 Mar 6 01:32:17.072606 kernel: loop2: detected capacity change from 0 to 228704 Mar 6 01:32:17.114699 kernel: loop3: detected capacity change from 0 to 140768 Mar 6 01:32:17.137600 kernel: loop4: detected capacity change from 0 to 142488 Mar 6 01:32:17.166645 kernel: loop5: detected capacity change from 0 to 228704 Mar 6 01:32:17.182147 (sd-merge)[1309]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Mar 6 01:32:17.183156 (sd-merge)[1309]: Merged extensions into '/usr'. Mar 6 01:32:17.187520 systemd[1]: Reloading requested from client PID 1297 ('systemd-sysext') (unit systemd-sysext.service)... Mar 6 01:32:17.187695 systemd[1]: Reloading... Mar 6 01:32:17.256638 zram_generator::config[1343]: No configuration found. Mar 6 01:32:17.287636 ldconfig[1293]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 6 01:32:17.394681 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 6 01:32:17.464267 systemd[1]: Reloading finished in 275 ms. Mar 6 01:32:17.488774 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 6 01:32:17.493212 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 6 01:32:17.513730 systemd[1]: Starting ensure-sysext.service... Mar 6 01:32:17.517667 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 6 01:32:17.524085 systemd[1]: Reloading requested from client PID 1381 ('systemctl') (unit ensure-sysext.service)... Mar 6 01:32:17.524097 systemd[1]: Reloading... Mar 6 01:32:17.551093 systemd-tmpfiles[1382]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 6 01:32:17.551782 systemd-tmpfiles[1382]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 6 01:32:17.553452 systemd-tmpfiles[1382]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 6 01:32:17.553985 systemd-tmpfiles[1382]: ACLs are not supported, ignoring. Mar 6 01:32:17.554105 systemd-tmpfiles[1382]: ACLs are not supported, ignoring. Mar 6 01:32:17.562321 systemd-tmpfiles[1382]: Detected autofs mount point /boot during canonicalization of boot. Mar 6 01:32:17.562380 systemd-tmpfiles[1382]: Skipping /boot Mar 6 01:32:17.586624 zram_generator::config[1413]: No configuration found. Mar 6 01:32:17.597239 systemd-tmpfiles[1382]: Detected autofs mount point /boot during canonicalization of boot. Mar 6 01:32:17.597253 systemd-tmpfiles[1382]: Skipping /boot Mar 6 01:32:17.717666 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 6 01:32:17.792136 systemd[1]: Reloading finished in 267 ms. Mar 6 01:32:17.813050 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 6 01:32:17.843075 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 6 01:32:17.848735 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 6 01:32:17.854374 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 6 01:32:17.862799 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 6 01:32:17.878898 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 6 01:32:17.895225 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:17.895884 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 01:32:17.904925 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 01:32:17.913861 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 01:32:17.919806 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 01:32:17.924672 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 01:32:17.924854 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:17.926753 augenrules[1479]: No rules Mar 6 01:32:17.926276 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 6 01:32:17.936320 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 6 01:32:17.941737 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 6 01:32:17.947778 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 01:32:17.948024 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 01:32:17.952886 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 01:32:17.953127 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 01:32:17.958158 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 01:32:17.958438 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 01:32:17.965477 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 6 01:32:17.966237 systemd-resolved[1460]: Positive Trust Anchors: Mar 6 01:32:17.966294 systemd-resolved[1460]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 6 01:32:17.966323 systemd-resolved[1460]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 6 01:32:17.974325 systemd-resolved[1460]: Defaulting to hostname 'linux'. Mar 6 01:32:17.977400 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 6 01:32:17.983635 systemd[1]: Reached target network.target - Network. Mar 6 01:32:17.987089 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 6 01:32:17.991309 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:17.991664 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 01:32:18.000852 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 01:32:18.006161 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 01:32:18.011494 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 01:32:18.015041 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 01:32:18.017187 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 6 01:32:18.021822 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 6 01:32:18.022032 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:18.025017 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 01:32:18.025336 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 01:32:18.030903 systemd-networkd[1245]: eth0: Gained IPv6LL Mar 6 01:32:18.032238 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 01:32:18.032527 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 01:32:18.037964 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 6 01:32:18.045168 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 01:32:18.045480 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 01:32:18.051196 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 6 01:32:18.064917 systemd[1]: Reached target network-online.target - Network is Online. Mar 6 01:32:18.069267 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:18.069524 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 6 01:32:18.079931 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 6 01:32:18.084753 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 6 01:32:18.089096 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 6 01:32:18.094921 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 6 01:32:18.098889 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 6 01:32:18.099050 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 6 01:32:18.099156 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 6 01:32:18.101002 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 6 01:32:18.101313 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 6 01:32:18.107520 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 6 01:32:18.107943 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 6 01:32:18.112782 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 6 01:32:18.113079 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 6 01:32:18.117372 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 6 01:32:18.117809 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 6 01:32:18.124779 systemd[1]: Finished ensure-sysext.service. Mar 6 01:32:18.132649 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 6 01:32:18.132791 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 6 01:32:18.147817 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 6 01:32:18.214918 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 6 01:32:18.218890 systemd[1]: Reached target sysinit.target - System Initialization. Mar 6 01:32:18.222483 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 6 01:32:18.226708 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 6 01:32:19.023032 systemd-resolved[1460]: Clock change detected. Flushing caches. Mar 6 01:32:19.027932 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 6 01:32:19.027970 systemd-timesyncd[1529]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 6 01:32:19.032133 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 6 01:32:19.032176 systemd-timesyncd[1529]: Initial clock synchronization to Fri 2026-03-06 01:32:19.022902 UTC. Mar 6 01:32:19.032198 systemd[1]: Reached target paths.target - Path Units. Mar 6 01:32:19.035110 systemd[1]: Reached target time-set.target - System Time Set. Mar 6 01:32:19.040148 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 6 01:32:19.043588 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 6 01:32:19.047451 systemd[1]: Reached target timers.target - Timer Units. Mar 6 01:32:19.051134 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 6 01:32:19.056820 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 6 01:32:19.061940 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 6 01:32:19.068151 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 6 01:32:19.071715 systemd[1]: Reached target sockets.target - Socket Units. Mar 6 01:32:19.075059 systemd[1]: Reached target basic.target - Basic System. Mar 6 01:32:19.078404 systemd[1]: System is tainted: cgroupsv1 Mar 6 01:32:19.078471 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 6 01:32:19.078496 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 6 01:32:19.080090 systemd[1]: Starting containerd.service - containerd container runtime... Mar 6 01:32:19.085116 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 6 01:32:19.090086 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 6 01:32:19.095949 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 6 01:32:19.101219 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 6 01:32:19.104678 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 6 01:32:19.108715 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:32:19.115980 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 6 01:32:19.122866 jq[1538]: false Mar 6 01:32:19.123178 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 6 01:32:19.133955 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 6 01:32:19.142740 dbus-daemon[1537]: [system] SELinux support is enabled Mar 6 01:32:19.145498 extend-filesystems[1540]: Found loop3 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found loop4 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found loop5 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found sr0 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda1 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda2 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda3 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found usr Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda4 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda6 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda7 Mar 6 01:32:19.145498 extend-filesystems[1540]: Found vda9 Mar 6 01:32:19.145498 extend-filesystems[1540]: Checking size of /dev/vda9 Mar 6 01:32:19.222214 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 6 01:32:19.222335 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1255) Mar 6 01:32:19.145523 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 6 01:32:19.222749 extend-filesystems[1540]: Resized partition /dev/vda9 Mar 6 01:32:19.153436 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 6 01:32:19.233144 extend-filesystems[1567]: resize2fs 1.47.1 (20-May-2024) Mar 6 01:32:19.168021 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 6 01:32:19.174198 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 6 01:32:19.175899 systemd[1]: Starting update-engine.service - Update Engine... Mar 6 01:32:19.245734 jq[1575]: true Mar 6 01:32:19.216322 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 6 01:32:19.223672 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 6 01:32:19.253558 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 6 01:32:19.254013 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 6 01:32:19.260491 systemd[1]: motdgen.service: Deactivated successfully. Mar 6 01:32:19.260945 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 6 01:32:19.276883 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 6 01:32:19.287169 update_engine[1568]: I20260306 01:32:19.271306 1568 main.cc:92] Flatcar Update Engine starting Mar 6 01:32:19.287169 update_engine[1568]: I20260306 01:32:19.275053 1568 update_check_scheduler.cc:74] Next update check in 4m37s Mar 6 01:32:19.270482 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 6 01:32:19.280462 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 6 01:32:19.280966 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 6 01:32:19.288540 extend-filesystems[1567]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 6 01:32:19.288540 extend-filesystems[1567]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 6 01:32:19.288540 extend-filesystems[1567]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 6 01:32:19.305398 extend-filesystems[1540]: Resized filesystem in /dev/vda9 Mar 6 01:32:19.309485 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 6 01:32:19.310056 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 6 01:32:19.324124 (ntainerd)[1585]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 6 01:32:19.327177 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 6 01:32:19.327686 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 6 01:32:19.332507 jq[1584]: true Mar 6 01:32:19.347340 systemd-logind[1561]: Watching system buttons on /dev/input/event1 (Power Button) Mar 6 01:32:19.347370 systemd-logind[1561]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 6 01:32:19.349706 systemd-logind[1561]: New seat seat0. Mar 6 01:32:19.353070 systemd[1]: Started systemd-logind.service - User Login Management. Mar 6 01:32:19.377256 tar[1582]: linux-amd64/LICENSE Mar 6 01:32:19.377322 dbus-daemon[1537]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 6 01:32:19.379356 tar[1582]: linux-amd64/helm Mar 6 01:32:19.394331 systemd[1]: Started update-engine.service - Update Engine. Mar 6 01:32:19.399307 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 6 01:32:19.399504 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 6 01:32:19.399655 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 6 01:32:19.403577 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 6 01:32:19.403720 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 6 01:32:19.408557 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 6 01:32:19.415221 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 6 01:32:19.427591 sshd_keygen[1572]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 6 01:32:19.462601 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 6 01:32:19.468597 bash[1624]: Updated "/home/core/.ssh/authorized_keys" Mar 6 01:32:19.481179 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 6 01:32:19.485216 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 6 01:32:19.491416 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 6 01:32:19.495599 locksmithd[1616]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 6 01:32:19.500596 systemd[1]: issuegen.service: Deactivated successfully. Mar 6 01:32:19.501082 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 6 01:32:19.513074 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 6 01:32:19.533041 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 6 01:32:19.549265 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 6 01:32:19.555926 containerd[1585]: time="2026-03-06T01:32:19.555652324Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Mar 6 01:32:19.561120 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 6 01:32:19.565986 systemd[1]: Reached target getty.target - Login Prompts. Mar 6 01:32:19.581910 containerd[1585]: time="2026-03-06T01:32:19.581573886Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.584276 containerd[1585]: time="2026-03-06T01:32:19.584203920Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 6 01:32:19.584276 containerd[1585]: time="2026-03-06T01:32:19.584252611Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 6 01:32:19.584276 containerd[1585]: time="2026-03-06T01:32:19.584268229Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 6 01:32:19.584522 containerd[1585]: time="2026-03-06T01:32:19.584447014Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 6 01:32:19.584522 containerd[1585]: time="2026-03-06T01:32:19.584494963Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.584626 containerd[1585]: time="2026-03-06T01:32:19.584582737Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 6 01:32:19.584626 containerd[1585]: time="2026-03-06T01:32:19.584619826Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585015 containerd[1585]: time="2026-03-06T01:32:19.584975361Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585015 containerd[1585]: time="2026-03-06T01:32:19.585010686Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585083 containerd[1585]: time="2026-03-06T01:32:19.585030273Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585083 containerd[1585]: time="2026-03-06T01:32:19.585040632Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585178 containerd[1585]: time="2026-03-06T01:32:19.585138585Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585563 containerd[1585]: time="2026-03-06T01:32:19.585497015Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585856 containerd[1585]: time="2026-03-06T01:32:19.585749315Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 6 01:32:19.585968 containerd[1585]: time="2026-03-06T01:32:19.585927969Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 6 01:32:19.586190 containerd[1585]: time="2026-03-06T01:32:19.586086556Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 6 01:32:19.586309 containerd[1585]: time="2026-03-06T01:32:19.586266381Z" level=info msg="metadata content store policy set" policy=shared Mar 6 01:32:19.593186 containerd[1585]: time="2026-03-06T01:32:19.593161363Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 6 01:32:19.593311 containerd[1585]: time="2026-03-06T01:32:19.593294000Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 6 01:32:19.593369 containerd[1585]: time="2026-03-06T01:32:19.593356498Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 6 01:32:19.593420 containerd[1585]: time="2026-03-06T01:32:19.593407843Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 6 01:32:19.593466 containerd[1585]: time="2026-03-06T01:32:19.593454460Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 6 01:32:19.593638 containerd[1585]: time="2026-03-06T01:32:19.593621282Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 6 01:32:19.594081 containerd[1585]: time="2026-03-06T01:32:19.594058258Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 6 01:32:19.594256 containerd[1585]: time="2026-03-06T01:32:19.594239476Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 6 01:32:19.594310 containerd[1585]: time="2026-03-06T01:32:19.594297715Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 6 01:32:19.594358 containerd[1585]: time="2026-03-06T01:32:19.594346276Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 6 01:32:19.594404 containerd[1585]: time="2026-03-06T01:32:19.594393033Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594447 containerd[1585]: time="2026-03-06T01:32:19.594436354Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594507 containerd[1585]: time="2026-03-06T01:32:19.594493220Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594555 containerd[1585]: time="2026-03-06T01:32:19.594543194Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594599 containerd[1585]: time="2026-03-06T01:32:19.594588087Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594641 containerd[1585]: time="2026-03-06T01:32:19.594630396Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594682 containerd[1585]: time="2026-03-06T01:32:19.594671733Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594724 containerd[1585]: time="2026-03-06T01:32:19.594712780Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 6 01:32:19.594914 containerd[1585]: time="2026-03-06T01:32:19.594892446Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.594992 containerd[1585]: time="2026-03-06T01:32:19.594977674Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595055 containerd[1585]: time="2026-03-06T01:32:19.595038939Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595120 containerd[1585]: time="2026-03-06T01:32:19.595106696Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595180 containerd[1585]: time="2026-03-06T01:32:19.595165205Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595229 containerd[1585]: time="2026-03-06T01:32:19.595217522Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595272 containerd[1585]: time="2026-03-06T01:32:19.595261024Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595315 containerd[1585]: time="2026-03-06T01:32:19.595303944Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595366 containerd[1585]: time="2026-03-06T01:32:19.595354017Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595414 containerd[1585]: time="2026-03-06T01:32:19.595402467Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595456 containerd[1585]: time="2026-03-06T01:32:19.595445348Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595511 containerd[1585]: time="2026-03-06T01:32:19.595497696Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595559 containerd[1585]: time="2026-03-06T01:32:19.595546788Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595605 containerd[1585]: time="2026-03-06T01:32:19.595594316Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 6 01:32:19.595661 containerd[1585]: time="2026-03-06T01:32:19.595648938Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595724 containerd[1585]: time="2026-03-06T01:32:19.595711114Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.595875 containerd[1585]: time="2026-03-06T01:32:19.595857317Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 6 01:32:19.596028 containerd[1585]: time="2026-03-06T01:32:19.596008830Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 6 01:32:19.596092 containerd[1585]: time="2026-03-06T01:32:19.596077448Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 6 01:32:19.596135 containerd[1585]: time="2026-03-06T01:32:19.596123504Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 6 01:32:19.596182 containerd[1585]: time="2026-03-06T01:32:19.596168809Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 6 01:32:19.596244 containerd[1585]: time="2026-03-06T01:32:19.596229983Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.596293 containerd[1585]: time="2026-03-06T01:32:19.596281560Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 6 01:32:19.596336 containerd[1585]: time="2026-03-06T01:32:19.596325381Z" level=info msg="NRI interface is disabled by configuration." Mar 6 01:32:19.596378 containerd[1585]: time="2026-03-06T01:32:19.596366488Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 6 01:32:19.596876 containerd[1585]: time="2026-03-06T01:32:19.596720820Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.597112191Z" level=info msg="Connect containerd service" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.597193593Z" level=info msg="using legacy CRI server" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.597202409Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.597284833Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.598076592Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.598596142Z" level=info msg="Start subscribing containerd event" Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.598735422Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 6 01:32:19.598911 containerd[1585]: time="2026-03-06T01:32:19.598744238Z" level=info msg="Start recovering state" Mar 6 01:32:19.599169 containerd[1585]: time="2026-03-06T01:32:19.598940796Z" level=info msg="Start event monitor" Mar 6 01:32:19.599169 containerd[1585]: time="2026-03-06T01:32:19.598964961Z" level=info msg="Start snapshots syncer" Mar 6 01:32:19.599169 containerd[1585]: time="2026-03-06T01:32:19.598980019Z" level=info msg="Start cni network conf syncer for default" Mar 6 01:32:19.599169 containerd[1585]: time="2026-03-06T01:32:19.598942398Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 6 01:32:19.599169 containerd[1585]: time="2026-03-06T01:32:19.598993314Z" level=info msg="Start streaming server" Mar 6 01:32:19.599261 systemd[1]: Started containerd.service - containerd container runtime. Mar 6 01:32:19.599582 containerd[1585]: time="2026-03-06T01:32:19.599493998Z" level=info msg="containerd successfully booted in 0.045093s" Mar 6 01:32:20.022458 tar[1582]: linux-amd64/README.md Mar 6 01:32:20.044420 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 6 01:32:20.284466 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:32:20.288596 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 6 01:32:20.292435 systemd[1]: Startup finished in 12.090s (kernel) + 9.221s (userspace) = 21.312s. Mar 6 01:32:20.368624 (kubelet)[1669]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 01:32:20.956921 kubelet[1669]: E0306 01:32:20.956379 1669 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 01:32:20.961452 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 01:32:20.961928 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 01:32:27.871411 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 6 01:32:27.909147 systemd[1]: Started sshd@0-10.0.0.56:22-10.0.0.1:56382.service - OpenSSH per-connection server daemon (10.0.0.1:56382). Mar 6 01:32:28.324627 sshd[1683]: Accepted publickey for core from 10.0.0.1 port 56382 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:28.329125 sshd[1683]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:28.374030 systemd-logind[1561]: New session 1 of user core. Mar 6 01:32:28.375972 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 6 01:32:28.384471 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 6 01:32:28.418498 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 6 01:32:28.434706 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 6 01:32:28.443270 (systemd)[1689]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 6 01:32:28.649725 systemd[1689]: Queued start job for default target default.target. Mar 6 01:32:28.650876 systemd[1689]: Created slice app.slice - User Application Slice. Mar 6 01:32:28.650913 systemd[1689]: Reached target paths.target - Paths. Mar 6 01:32:28.650928 systemd[1689]: Reached target timers.target - Timers. Mar 6 01:32:28.662120 systemd[1689]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 6 01:32:28.684053 systemd[1689]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 6 01:32:28.684333 systemd[1689]: Reached target sockets.target - Sockets. Mar 6 01:32:28.684363 systemd[1689]: Reached target basic.target - Basic System. Mar 6 01:32:28.684553 systemd[1689]: Reached target default.target - Main User Target. Mar 6 01:32:28.684737 systemd[1689]: Startup finished in 220ms. Mar 6 01:32:28.685123 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 6 01:32:28.687714 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 6 01:32:28.772215 systemd[1]: Started sshd@1-10.0.0.56:22-10.0.0.1:56388.service - OpenSSH per-connection server daemon (10.0.0.1:56388). Mar 6 01:32:28.830501 sshd[1701]: Accepted publickey for core from 10.0.0.1 port 56388 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:28.833971 sshd[1701]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:28.848393 systemd-logind[1561]: New session 2 of user core. Mar 6 01:32:28.858593 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 6 01:32:28.941189 sshd[1701]: pam_unix(sshd:session): session closed for user core Mar 6 01:32:28.953389 systemd[1]: Started sshd@2-10.0.0.56:22-10.0.0.1:56394.service - OpenSSH per-connection server daemon (10.0.0.1:56394). Mar 6 01:32:28.954279 systemd[1]: sshd@1-10.0.0.56:22-10.0.0.1:56388.service: Deactivated successfully. Mar 6 01:32:28.961312 systemd-logind[1561]: Session 2 logged out. Waiting for processes to exit. Mar 6 01:32:28.964870 systemd[1]: session-2.scope: Deactivated successfully. Mar 6 01:32:28.969234 systemd-logind[1561]: Removed session 2. Mar 6 01:32:29.002386 sshd[1706]: Accepted publickey for core from 10.0.0.1 port 56394 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:29.004948 sshd[1706]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:29.014365 systemd-logind[1561]: New session 3 of user core. Mar 6 01:32:29.024616 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 6 01:32:29.083893 sshd[1706]: pam_unix(sshd:session): session closed for user core Mar 6 01:32:29.092364 systemd[1]: Started sshd@3-10.0.0.56:22-10.0.0.1:56398.service - OpenSSH per-connection server daemon (10.0.0.1:56398). Mar 6 01:32:29.093198 systemd[1]: sshd@2-10.0.0.56:22-10.0.0.1:56394.service: Deactivated successfully. Mar 6 01:32:29.099938 systemd[1]: session-3.scope: Deactivated successfully. Mar 6 01:32:29.100427 systemd-logind[1561]: Session 3 logged out. Waiting for processes to exit. Mar 6 01:32:29.104725 systemd-logind[1561]: Removed session 3. Mar 6 01:32:29.135517 sshd[1714]: Accepted publickey for core from 10.0.0.1 port 56398 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:29.139147 sshd[1714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:29.151001 systemd-logind[1561]: New session 4 of user core. Mar 6 01:32:29.161418 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 6 01:32:29.233600 sshd[1714]: pam_unix(sshd:session): session closed for user core Mar 6 01:32:29.247293 systemd[1]: Started sshd@4-10.0.0.56:22-10.0.0.1:56404.service - OpenSSH per-connection server daemon (10.0.0.1:56404). Mar 6 01:32:29.248164 systemd[1]: sshd@3-10.0.0.56:22-10.0.0.1:56398.service: Deactivated successfully. Mar 6 01:32:29.252554 systemd-logind[1561]: Session 4 logged out. Waiting for processes to exit. Mar 6 01:32:29.253556 systemd[1]: session-4.scope: Deactivated successfully. Mar 6 01:32:29.256909 systemd-logind[1561]: Removed session 4. Mar 6 01:32:29.291412 sshd[1722]: Accepted publickey for core from 10.0.0.1 port 56404 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:29.294273 sshd[1722]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:29.304121 systemd-logind[1561]: New session 5 of user core. Mar 6 01:32:29.314337 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 6 01:32:29.404348 sudo[1729]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 6 01:32:29.405111 sudo[1729]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 01:32:29.442656 sudo[1729]: pam_unix(sudo:session): session closed for user root Mar 6 01:32:29.447150 sshd[1722]: pam_unix(sshd:session): session closed for user core Mar 6 01:32:29.461255 systemd[1]: Started sshd@5-10.0.0.56:22-10.0.0.1:56416.service - OpenSSH per-connection server daemon (10.0.0.1:56416). Mar 6 01:32:29.462429 systemd[1]: sshd@4-10.0.0.56:22-10.0.0.1:56404.service: Deactivated successfully. Mar 6 01:32:29.468075 systemd-logind[1561]: Session 5 logged out. Waiting for processes to exit. Mar 6 01:32:29.470960 systemd[1]: session-5.scope: Deactivated successfully. Mar 6 01:32:29.476089 systemd-logind[1561]: Removed session 5. Mar 6 01:32:29.511900 sshd[1731]: Accepted publickey for core from 10.0.0.1 port 56416 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:29.515084 sshd[1731]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:29.533023 systemd-logind[1561]: New session 6 of user core. Mar 6 01:32:29.544408 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 6 01:32:29.613112 sudo[1739]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 6 01:32:29.613898 sudo[1739]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 01:32:29.629505 sudo[1739]: pam_unix(sudo:session): session closed for user root Mar 6 01:32:29.645600 sudo[1738]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Mar 6 01:32:29.646294 sudo[1738]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 01:32:29.679715 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Mar 6 01:32:29.684861 auditctl[1742]: No rules Mar 6 01:32:29.687321 systemd[1]: audit-rules.service: Deactivated successfully. Mar 6 01:32:29.688359 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Mar 6 01:32:29.707842 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 6 01:32:29.780107 augenrules[1761]: No rules Mar 6 01:32:29.781242 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 6 01:32:29.783130 sudo[1738]: pam_unix(sudo:session): session closed for user root Mar 6 01:32:29.786198 sshd[1731]: pam_unix(sshd:session): session closed for user core Mar 6 01:32:29.797430 systemd[1]: Started sshd@6-10.0.0.56:22-10.0.0.1:56420.service - OpenSSH per-connection server daemon (10.0.0.1:56420). Mar 6 01:32:29.798382 systemd[1]: sshd@5-10.0.0.56:22-10.0.0.1:56416.service: Deactivated successfully. Mar 6 01:32:29.803086 systemd-logind[1561]: Session 6 logged out. Waiting for processes to exit. Mar 6 01:32:29.804551 systemd[1]: session-6.scope: Deactivated successfully. Mar 6 01:32:29.807426 systemd-logind[1561]: Removed session 6. Mar 6 01:32:29.842507 sshd[1767]: Accepted publickey for core from 10.0.0.1 port 56420 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:32:29.845692 sshd[1767]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:32:29.855701 systemd-logind[1561]: New session 7 of user core. Mar 6 01:32:29.867322 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 6 01:32:29.930847 sudo[1774]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 6 01:32:29.931292 sudo[1774]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 6 01:32:31.151511 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 6 01:32:31.166106 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:32:31.702299 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 6 01:32:31.703037 (dockerd)[1798]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 6 01:32:32.175672 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:32:32.189966 (kubelet)[1807]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 01:32:34.107352 kubelet[1807]: E0306 01:32:34.106389 1807 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 01:32:34.144017 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 01:32:34.147902 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 01:32:36.610261 dockerd[1798]: time="2026-03-06T01:32:36.609754300Z" level=info msg="Starting up" Mar 6 01:32:37.501576 dockerd[1798]: time="2026-03-06T01:32:37.501316511Z" level=info msg="Loading containers: start." Mar 6 01:32:37.771892 kernel: Initializing XFRM netlink socket Mar 6 01:32:38.063324 systemd-networkd[1245]: docker0: Link UP Mar 6 01:32:38.099522 dockerd[1798]: time="2026-03-06T01:32:38.099247942Z" level=info msg="Loading containers: done." Mar 6 01:32:38.143254 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck318882349-merged.mount: Deactivated successfully. Mar 6 01:32:38.146851 dockerd[1798]: time="2026-03-06T01:32:38.146674085Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 6 01:32:38.147086 dockerd[1798]: time="2026-03-06T01:32:38.147016935Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Mar 6 01:32:38.147369 dockerd[1798]: time="2026-03-06T01:32:38.147318499Z" level=info msg="Daemon has completed initialization" Mar 6 01:32:38.200213 dockerd[1798]: time="2026-03-06T01:32:38.200069698Z" level=info msg="API listen on /run/docker.sock" Mar 6 01:32:38.200343 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 6 01:32:40.509453 containerd[1585]: time="2026-03-06T01:32:40.509274652Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\"" Mar 6 01:32:41.857295 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3638560611.mount: Deactivated successfully. Mar 6 01:32:44.397046 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 6 01:32:44.426271 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:32:44.898164 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:32:44.956922 (kubelet)[2035]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 01:32:45.155126 kubelet[2035]: E0306 01:32:45.153495 2035 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 01:32:45.159938 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 01:32:45.160299 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 01:32:45.603710 containerd[1585]: time="2026-03-06T01:32:45.602698741Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:45.605222 containerd[1585]: time="2026-03-06T01:32:45.604509959Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.9: active requests=0, bytes read=30116186" Mar 6 01:32:45.606340 containerd[1585]: time="2026-03-06T01:32:45.606250872Z" level=info msg="ImageCreate event name:\"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:45.611365 containerd[1585]: time="2026-03-06T01:32:45.611280664Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:45.614114 containerd[1585]: time="2026-03-06T01:32:45.613997863Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.9\" with image id \"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\", size \"30112785\" in 5.102381245s" Mar 6 01:32:45.614114 containerd[1585]: time="2026-03-06T01:32:45.614082682Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\" returns image reference \"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\"" Mar 6 01:32:45.618118 containerd[1585]: time="2026-03-06T01:32:45.618033376Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\"" Mar 6 01:32:50.196277 containerd[1585]: time="2026-03-06T01:32:50.195535199Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:50.198668 containerd[1585]: time="2026-03-06T01:32:50.198447347Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.9: active requests=0, bytes read=26021810" Mar 6 01:32:50.201558 containerd[1585]: time="2026-03-06T01:32:50.201396364Z" level=info msg="ImageCreate event name:\"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:50.216589 containerd[1585]: time="2026-03-06T01:32:50.216146658Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:50.218834 containerd[1585]: time="2026-03-06T01:32:50.218640705Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.9\" with image id \"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\", size \"27678758\" in 4.600520397s" Mar 6 01:32:50.218998 containerd[1585]: time="2026-03-06T01:32:50.218929034Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\" returns image reference \"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\"" Mar 6 01:32:50.228149 containerd[1585]: time="2026-03-06T01:32:50.227862175Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\"" Mar 6 01:32:53.296573 containerd[1585]: time="2026-03-06T01:32:53.296194101Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:53.298633 containerd[1585]: time="2026-03-06T01:32:53.297906524Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.9: active requests=0, bytes read=20162746" Mar 6 01:32:53.299883 containerd[1585]: time="2026-03-06T01:32:53.299457283Z" level=info msg="ImageCreate event name:\"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:53.303615 containerd[1585]: time="2026-03-06T01:32:53.303539067Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:53.304707 containerd[1585]: time="2026-03-06T01:32:53.304631101Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.9\" with image id \"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\", size \"21819712\" in 3.07517057s" Mar 6 01:32:53.304885 containerd[1585]: time="2026-03-06T01:32:53.304734312Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\" returns image reference \"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\"" Mar 6 01:32:53.309588 containerd[1585]: time="2026-03-06T01:32:53.309526531Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\"" Mar 6 01:32:55.396640 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 6 01:32:55.420365 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:32:55.508724 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1028249737.mount: Deactivated successfully. Mar 6 01:32:55.657332 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:32:55.666160 (kubelet)[2069]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 6 01:32:55.830618 kubelet[2069]: E0306 01:32:55.830453 2069 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 6 01:32:55.834596 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 6 01:32:55.835069 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 6 01:32:56.148253 containerd[1585]: time="2026-03-06T01:32:56.148173493Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:56.149355 containerd[1585]: time="2026-03-06T01:32:56.149284427Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.9: active requests=0, bytes read=31828647" Mar 6 01:32:56.150498 containerd[1585]: time="2026-03-06T01:32:56.150438454Z" level=info msg="ImageCreate event name:\"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:56.155393 containerd[1585]: time="2026-03-06T01:32:56.155346071Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.9\" with image id \"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\", repo tag \"registry.k8s.io/kube-proxy:v1.33.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\", size \"31827666\" in 2.845773224s" Mar 6 01:32:56.155393 containerd[1585]: time="2026-03-06T01:32:56.155394931Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\" returns image reference \"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\"" Mar 6 01:32:56.157052 containerd[1585]: time="2026-03-06T01:32:56.156375892Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:56.157300 containerd[1585]: time="2026-03-06T01:32:56.157262766Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Mar 6 01:32:56.620700 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount237351239.mount: Deactivated successfully. Mar 6 01:32:58.667364 containerd[1585]: time="2026-03-06T01:32:58.667131887Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:58.668901 containerd[1585]: time="2026-03-06T01:32:58.668435038Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Mar 6 01:32:58.670586 containerd[1585]: time="2026-03-06T01:32:58.670460101Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:58.675844 containerd[1585]: time="2026-03-06T01:32:58.675727600Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:58.682677 containerd[1585]: time="2026-03-06T01:32:58.680241581Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 2.522890541s" Mar 6 01:32:58.682677 containerd[1585]: time="2026-03-06T01:32:58.680418260Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Mar 6 01:32:58.685120 containerd[1585]: time="2026-03-06T01:32:58.685054001Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 6 01:32:59.368314 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2212067889.mount: Deactivated successfully. Mar 6 01:32:59.373437 containerd[1585]: time="2026-03-06T01:32:59.373358343Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:59.374312 containerd[1585]: time="2026-03-06T01:32:59.374204754Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Mar 6 01:32:59.375641 containerd[1585]: time="2026-03-06T01:32:59.375570013Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:59.378633 containerd[1585]: time="2026-03-06T01:32:59.378569565Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:32:59.379374 containerd[1585]: time="2026-03-06T01:32:59.379317024Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 694.177443ms" Mar 6 01:32:59.379374 containerd[1585]: time="2026-03-06T01:32:59.379365925Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Mar 6 01:32:59.381324 containerd[1585]: time="2026-03-06T01:32:59.381277057Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Mar 6 01:32:59.959555 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2022890267.mount: Deactivated successfully. Mar 6 01:33:01.636641 containerd[1585]: time="2026-03-06T01:33:01.636418622Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:01.639102 containerd[1585]: time="2026-03-06T01:33:01.638483573Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=23718840" Mar 6 01:33:01.640194 containerd[1585]: time="2026-03-06T01:33:01.640140723Z" level=info msg="ImageCreate event name:\"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:01.644275 containerd[1585]: time="2026-03-06T01:33:01.644214003Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:01.645552 containerd[1585]: time="2026-03-06T01:33:01.645499703Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"23716032\" in 2.264175768s" Mar 6 01:33:01.645552 containerd[1585]: time="2026-03-06T01:33:01.645543855Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\"" Mar 6 01:33:05.008346 update_engine[1568]: I20260306 01:33:05.008103 1568 update_attempter.cc:509] Updating boot flags... Mar 6 01:33:05.072870 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2232) Mar 6 01:33:05.168855 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2236) Mar 6 01:33:05.895687 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 6 01:33:05.906042 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:33:06.272873 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 6 01:33:06.273068 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 6 01:33:06.273925 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:33:06.295110 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:33:06.371960 systemd[1]: Reloading requested from client PID 2254 ('systemctl') (unit session-7.scope)... Mar 6 01:33:06.372035 systemd[1]: Reloading... Mar 6 01:33:06.564965 zram_generator::config[2293]: No configuration found. Mar 6 01:33:06.916923 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 6 01:33:06.998265 systemd[1]: Reloading finished in 625 ms. Mar 6 01:33:07.055551 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 6 01:33:07.055903 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 6 01:33:07.056492 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:33:07.060007 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:33:07.281491 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:33:07.288400 (kubelet)[2354]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 6 01:33:07.620085 kubelet[2354]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 6 01:33:07.620085 kubelet[2354]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 6 01:33:07.620085 kubelet[2354]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 6 01:33:07.620085 kubelet[2354]: I0306 01:33:07.619120 2354 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 6 01:33:08.245716 kubelet[2354]: I0306 01:33:08.245603 2354 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 6 01:33:08.245716 kubelet[2354]: I0306 01:33:08.245646 2354 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 6 01:33:08.246188 kubelet[2354]: I0306 01:33:08.246150 2354 server.go:956] "Client rotation is on, will bootstrap in background" Mar 6 01:33:08.298363 kubelet[2354]: E0306 01:33:08.298236 2354 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.56:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 6 01:33:08.303065 kubelet[2354]: I0306 01:33:08.302958 2354 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 6 01:33:08.314670 kubelet[2354]: E0306 01:33:08.314596 2354 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 6 01:33:08.314670 kubelet[2354]: I0306 01:33:08.314673 2354 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 6 01:33:08.327994 kubelet[2354]: I0306 01:33:08.327913 2354 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 6 01:33:08.330463 kubelet[2354]: I0306 01:33:08.330338 2354 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 6 01:33:08.330665 kubelet[2354]: I0306 01:33:08.330420 2354 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 6 01:33:08.330665 kubelet[2354]: I0306 01:33:08.330664 2354 topology_manager.go:138] "Creating topology manager with none policy" Mar 6 01:33:08.331307 kubelet[2354]: I0306 01:33:08.330720 2354 container_manager_linux.go:303] "Creating device plugin manager" Mar 6 01:33:08.331307 kubelet[2354]: I0306 01:33:08.331082 2354 state_mem.go:36] "Initialized new in-memory state store" Mar 6 01:33:08.339552 kubelet[2354]: I0306 01:33:08.339471 2354 kubelet.go:480] "Attempting to sync node with API server" Mar 6 01:33:08.339552 kubelet[2354]: I0306 01:33:08.339529 2354 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 6 01:33:08.340960 kubelet[2354]: I0306 01:33:08.340864 2354 kubelet.go:386] "Adding apiserver pod source" Mar 6 01:33:08.343923 kubelet[2354]: I0306 01:33:08.342916 2354 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 6 01:33:08.350080 kubelet[2354]: E0306 01:33:08.350006 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.56:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 6 01:33:08.350080 kubelet[2354]: E0306 01:33:08.350007 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.56:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 6 01:33:08.351661 kubelet[2354]: I0306 01:33:08.350140 2354 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 6 01:33:08.351661 kubelet[2354]: I0306 01:33:08.351321 2354 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 6 01:33:08.353011 kubelet[2354]: W0306 01:33:08.352937 2354 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 6 01:33:08.362735 kubelet[2354]: I0306 01:33:08.362642 2354 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 6 01:33:08.362953 kubelet[2354]: I0306 01:33:08.362754 2354 server.go:1289] "Started kubelet" Mar 6 01:33:08.364891 kubelet[2354]: I0306 01:33:08.363183 2354 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 6 01:33:08.364891 kubelet[2354]: I0306 01:33:08.364066 2354 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 6 01:33:08.364891 kubelet[2354]: I0306 01:33:08.364167 2354 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 6 01:33:08.364891 kubelet[2354]: I0306 01:33:08.364475 2354 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 6 01:33:08.365363 kubelet[2354]: I0306 01:33:08.365290 2354 server.go:317] "Adding debug handlers to kubelet server" Mar 6 01:33:08.370724 kubelet[2354]: I0306 01:33:08.370611 2354 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 6 01:33:08.371191 kubelet[2354]: E0306 01:33:08.371098 2354 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 01:33:08.371191 kubelet[2354]: I0306 01:33:08.371189 2354 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 6 01:33:08.371896 kubelet[2354]: I0306 01:33:08.371371 2354 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 6 01:33:08.371896 kubelet[2354]: I0306 01:33:08.371443 2354 reconciler.go:26] "Reconciler: start to sync state" Mar 6 01:33:08.372147 kubelet[2354]: E0306 01:33:08.371968 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 6 01:33:08.373014 kubelet[2354]: E0306 01:33:08.372955 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.56:6443: connect: connection refused" interval="200ms" Mar 6 01:33:08.373242 kubelet[2354]: E0306 01:33:08.370397 2354 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.56:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.56:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.189a1c8b661963cd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-06 01:33:08.362683341 +0000 UTC m=+1.065549286,LastTimestamp:2026-03-06 01:33:08.362683341 +0000 UTC m=+1.065549286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 6 01:33:08.374095 kubelet[2354]: I0306 01:33:08.373393 2354 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 6 01:33:08.375134 kubelet[2354]: E0306 01:33:08.375034 2354 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 6 01:33:08.376017 kubelet[2354]: I0306 01:33:08.375947 2354 factory.go:223] Registration of the containerd container factory successfully Mar 6 01:33:08.376017 kubelet[2354]: I0306 01:33:08.375985 2354 factory.go:223] Registration of the systemd container factory successfully Mar 6 01:33:08.525002 kubelet[2354]: E0306 01:33:08.515395 2354 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 01:33:08.759148 kubelet[2354]: E0306 01:33:08.758936 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.56:6443: connect: connection refused" interval="400ms" Mar 6 01:33:08.759148 kubelet[2354]: E0306 01:33:08.759100 2354 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 01:33:08.777224 kubelet[2354]: I0306 01:33:08.776877 2354 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 6 01:33:08.793468 kubelet[2354]: I0306 01:33:08.792170 2354 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 6 01:33:08.793468 kubelet[2354]: I0306 01:33:08.792223 2354 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 6 01:33:08.793468 kubelet[2354]: I0306 01:33:08.792314 2354 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 6 01:33:08.793468 kubelet[2354]: I0306 01:33:08.792331 2354 kubelet.go:2436] "Starting kubelet main sync loop" Mar 6 01:33:08.793468 kubelet[2354]: E0306 01:33:08.792470 2354 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 6 01:33:08.793468 kubelet[2354]: E0306 01:33:08.793426 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.56:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 6 01:33:08.799719 kubelet[2354]: I0306 01:33:08.799645 2354 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 6 01:33:08.800171 kubelet[2354]: I0306 01:33:08.800032 2354 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 6 01:33:08.800171 kubelet[2354]: I0306 01:33:08.800157 2354 state_mem.go:36] "Initialized new in-memory state store" Mar 6 01:33:08.860151 kubelet[2354]: E0306 01:33:08.859962 2354 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 01:33:08.893302 kubelet[2354]: E0306 01:33:08.893160 2354 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 6 01:33:08.960969 kubelet[2354]: I0306 01:33:08.960540 2354 policy_none.go:49] "None policy: Start" Mar 6 01:33:08.960969 kubelet[2354]: I0306 01:33:08.960729 2354 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 6 01:33:08.960969 kubelet[2354]: I0306 01:33:08.960896 2354 state_mem.go:35] "Initializing new in-memory state store" Mar 6 01:33:08.960969 kubelet[2354]: E0306 01:33:08.960906 2354 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 01:33:09.023637 kubelet[2354]: E0306 01:33:09.014915 2354 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 6 01:33:09.071565 kubelet[2354]: I0306 01:33:09.064495 2354 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 6 01:33:09.071565 kubelet[2354]: I0306 01:33:09.066129 2354 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 6 01:33:09.071565 kubelet[2354]: E0306 01:33:09.071085 2354 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 6 01:33:09.074738 kubelet[2354]: I0306 01:33:09.074624 2354 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 6 01:33:09.088942 kubelet[2354]: E0306 01:33:09.087136 2354 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 6 01:33:09.088942 kubelet[2354]: E0306 01:33:09.088880 2354 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 6 01:33:09.162305 kubelet[2354]: E0306 01:33:09.161745 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.56:6443: connect: connection refused" interval="800ms" Mar 6 01:33:09.171858 kubelet[2354]: E0306 01:33:09.170438 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:09.174264 kubelet[2354]: I0306 01:33:09.174181 2354 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 6 01:33:09.174927 kubelet[2354]: E0306 01:33:09.174846 2354 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.56:6443/api/v1/nodes\": dial tcp 10.0.0.56:6443: connect: connection refused" node="localhost" Mar 6 01:33:09.177154 kubelet[2354]: E0306 01:33:09.177098 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:09.183708 kubelet[2354]: E0306 01:33:09.183613 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:09.265842 kubelet[2354]: I0306 01:33:09.265664 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:09.265842 kubelet[2354]: I0306 01:33:09.265743 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:09.266224 kubelet[2354]: I0306 01:33:09.265883 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:09.266224 kubelet[2354]: I0306 01:33:09.265904 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:09.266224 kubelet[2354]: I0306 01:33:09.265920 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:09.266224 kubelet[2354]: I0306 01:33:09.265935 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e944e4cb17af904786c3a2e01e298498-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"e944e4cb17af904786c3a2e01e298498\") " pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:09.266224 kubelet[2354]: I0306 01:33:09.265949 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d9b0c3b0c90cd5a17c634cbdf5c19ab7-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"d9b0c3b0c90cd5a17c634cbdf5c19ab7\") " pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:09.266333 kubelet[2354]: I0306 01:33:09.265984 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d9b0c3b0c90cd5a17c634cbdf5c19ab7-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"d9b0c3b0c90cd5a17c634cbdf5c19ab7\") " pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:09.266333 kubelet[2354]: I0306 01:33:09.266000 2354 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d9b0c3b0c90cd5a17c634cbdf5c19ab7-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"d9b0c3b0c90cd5a17c634cbdf5c19ab7\") " pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:09.282658 kubelet[2354]: E0306 01:33:09.282550 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 6 01:33:09.377613 kubelet[2354]: I0306 01:33:09.377410 2354 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 6 01:33:09.377962 kubelet[2354]: E0306 01:33:09.377912 2354 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.56:6443/api/v1/nodes\": dial tcp 10.0.0.56:6443: connect: connection refused" node="localhost" Mar 6 01:33:09.473375 kubelet[2354]: E0306 01:33:09.472956 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:09.476962 containerd[1585]: time="2026-03-06T01:33:09.476837970Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:d9b0c3b0c90cd5a17c634cbdf5c19ab7,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:09.478041 kubelet[2354]: E0306 01:33:09.477922 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:09.478971 containerd[1585]: time="2026-03-06T01:33:09.478890420Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8747e1f8a49a618fbc1324a8fe2d3754,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:09.485018 kubelet[2354]: E0306 01:33:09.484940 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:09.485669 containerd[1585]: time="2026-03-06T01:33:09.485580129Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:e944e4cb17af904786c3a2e01e298498,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:09.617176 kubelet[2354]: E0306 01:33:09.617090 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.56:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 6 01:33:09.636993 kubelet[2354]: E0306 01:33:09.636668 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.56:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 6 01:33:09.780879 kubelet[2354]: I0306 01:33:09.780517 2354 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 6 01:33:09.781528 kubelet[2354]: E0306 01:33:09.781147 2354 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.56:6443/api/v1/nodes\": dial tcp 10.0.0.56:6443: connect: connection refused" node="localhost" Mar 6 01:33:09.850239 kubelet[2354]: E0306 01:33:09.850068 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.56:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 6 01:33:09.939270 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1979045534.mount: Deactivated successfully. Mar 6 01:33:09.949417 containerd[1585]: time="2026-03-06T01:33:09.949239089Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 01:33:09.954207 containerd[1585]: time="2026-03-06T01:33:09.954091722Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Mar 6 01:33:09.955732 containerd[1585]: time="2026-03-06T01:33:09.955587004Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 01:33:09.957033 containerd[1585]: time="2026-03-06T01:33:09.956987828Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 01:33:09.958637 containerd[1585]: time="2026-03-06T01:33:09.958473420Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 01:33:09.960590 containerd[1585]: time="2026-03-06T01:33:09.960342215Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 6 01:33:09.962889 containerd[1585]: time="2026-03-06T01:33:09.962520591Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 6 01:33:09.963851 kubelet[2354]: E0306 01:33:09.963698 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.56:6443: connect: connection refused" interval="1.6s" Mar 6 01:33:09.965001 containerd[1585]: time="2026-03-06T01:33:09.964936987Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 6 01:33:09.968022 containerd[1585]: time="2026-03-06T01:33:09.967914409Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 490.909036ms" Mar 6 01:33:09.969263 containerd[1585]: time="2026-03-06T01:33:09.969138607Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 490.076548ms" Mar 6 01:33:09.975721 containerd[1585]: time="2026-03-06T01:33:09.975664354Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 489.979829ms" Mar 6 01:33:10.468937 kubelet[2354]: E0306 01:33:10.468086 2354 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.56:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 6 01:33:10.603910 kubelet[2354]: I0306 01:33:10.603308 2354 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 6 01:33:10.606755 kubelet[2354]: E0306 01:33:10.605738 2354 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.56:6443/api/v1/nodes\": dial tcp 10.0.0.56:6443: connect: connection refused" node="localhost" Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.767649885Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.767725927Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.767752166Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.767941350Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.766569207Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.766934818Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:10.768974 containerd[1585]: time="2026-03-06T01:33:10.766962540Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:10.771195 containerd[1585]: time="2026-03-06T01:33:10.770472760Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:10.784634 containerd[1585]: time="2026-03-06T01:33:10.784323061Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:10.787195 containerd[1585]: time="2026-03-06T01:33:10.787038310Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:10.787195 containerd[1585]: time="2026-03-06T01:33:10.787138437Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:10.789945 containerd[1585]: time="2026-03-06T01:33:10.787280882Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:11.328408 kubelet[2354]: E0306 01:33:11.328130 2354 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.56:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 6 01:33:11.397590 containerd[1585]: time="2026-03-06T01:33:11.397031239Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:d9b0c3b0c90cd5a17c634cbdf5c19ab7,Namespace:kube-system,Attempt:0,} returns sandbox id \"185ccfaa21454bd871f7abedd2bfbdeb6978814e7bf74b7bb663def74f92a7ed\"" Mar 6 01:33:11.399638 kubelet[2354]: E0306 01:33:11.398981 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:11.407844 containerd[1585]: time="2026-03-06T01:33:11.407615206Z" level=info msg="CreateContainer within sandbox \"185ccfaa21454bd871f7abedd2bfbdeb6978814e7bf74b7bb663def74f92a7ed\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 6 01:33:11.438635 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2662660492.mount: Deactivated successfully. Mar 6 01:33:11.457930 containerd[1585]: time="2026-03-06T01:33:11.457593876Z" level=info msg="CreateContainer within sandbox \"185ccfaa21454bd871f7abedd2bfbdeb6978814e7bf74b7bb663def74f92a7ed\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"457e2fd40d2b8a4ba4afa6734af0a1ee920e8f2fdf6d2c4586c294494b53ee3f\"" Mar 6 01:33:11.459038 containerd[1585]: time="2026-03-06T01:33:11.458994798Z" level=info msg="StartContainer for \"457e2fd40d2b8a4ba4afa6734af0a1ee920e8f2fdf6d2c4586c294494b53ee3f\"" Mar 6 01:33:11.461013 containerd[1585]: time="2026-03-06T01:33:11.460912236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8747e1f8a49a618fbc1324a8fe2d3754,Namespace:kube-system,Attempt:0,} returns sandbox id \"b4e82e9f00553edd2a76f3a054777b8d2d33c98e4d3c34d8a2976e05989f3131\"" Mar 6 01:33:11.461889 kubelet[2354]: E0306 01:33:11.461657 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:11.467638 containerd[1585]: time="2026-03-06T01:33:11.467516074Z" level=info msg="CreateContainer within sandbox \"b4e82e9f00553edd2a76f3a054777b8d2d33c98e4d3c34d8a2976e05989f3131\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 6 01:33:11.478394 containerd[1585]: time="2026-03-06T01:33:11.478286734Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:e944e4cb17af904786c3a2e01e298498,Namespace:kube-system,Attempt:0,} returns sandbox id \"32bf886f8f86bb3303ead28c40d3372085bb5ef4a68423ab1df0d89c034f64a4\"" Mar 6 01:33:11.479370 kubelet[2354]: E0306 01:33:11.479276 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:11.486751 containerd[1585]: time="2026-03-06T01:33:11.486546789Z" level=info msg="CreateContainer within sandbox \"32bf886f8f86bb3303ead28c40d3372085bb5ef4a68423ab1df0d89c034f64a4\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 6 01:33:11.492331 containerd[1585]: time="2026-03-06T01:33:11.492177258Z" level=info msg="CreateContainer within sandbox \"b4e82e9f00553edd2a76f3a054777b8d2d33c98e4d3c34d8a2976e05989f3131\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"2a7d46588902825113aa12499206996991d4e77ff5373835e5e68798478c301c\"" Mar 6 01:33:11.493222 containerd[1585]: time="2026-03-06T01:33:11.492945068Z" level=info msg="StartContainer for \"2a7d46588902825113aa12499206996991d4e77ff5373835e5e68798478c301c\"" Mar 6 01:33:11.506841 containerd[1585]: time="2026-03-06T01:33:11.506721645Z" level=info msg="CreateContainer within sandbox \"32bf886f8f86bb3303ead28c40d3372085bb5ef4a68423ab1df0d89c034f64a4\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c09e349245e30fd50e08906ae86b5b8a7553a3150edfb929dd9fd2db85d9c491\"" Mar 6 01:33:11.507553 containerd[1585]: time="2026-03-06T01:33:11.507519301Z" level=info msg="StartContainer for \"c09e349245e30fd50e08906ae86b5b8a7553a3150edfb929dd9fd2db85d9c491\"" Mar 6 01:33:11.564855 kubelet[2354]: E0306 01:33:11.564633 2354 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.56:6443: connect: connection refused" interval="3.2s" Mar 6 01:33:11.851307 containerd[1585]: time="2026-03-06T01:33:11.851250664Z" level=info msg="StartContainer for \"c09e349245e30fd50e08906ae86b5b8a7553a3150edfb929dd9fd2db85d9c491\" returns successfully" Mar 6 01:33:11.852092 containerd[1585]: time="2026-03-06T01:33:11.851314864Z" level=info msg="StartContainer for \"457e2fd40d2b8a4ba4afa6734af0a1ee920e8f2fdf6d2c4586c294494b53ee3f\" returns successfully" Mar 6 01:33:11.874733 containerd[1585]: time="2026-03-06T01:33:11.874601181Z" level=info msg="StartContainer for \"2a7d46588902825113aa12499206996991d4e77ff5373835e5e68798478c301c\" returns successfully" Mar 6 01:33:12.215636 kubelet[2354]: I0306 01:33:12.214736 2354 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 6 01:33:12.877037 kubelet[2354]: E0306 01:33:12.877003 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:12.886829 kubelet[2354]: E0306 01:33:12.884347 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:12.892909 kubelet[2354]: E0306 01:33:12.892855 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:12.893145 kubelet[2354]: E0306 01:33:12.893084 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:12.897330 kubelet[2354]: E0306 01:33:12.897303 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:12.899079 kubelet[2354]: E0306 01:33:12.899053 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:13.901173 kubelet[2354]: E0306 01:33:13.900152 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:13.901173 kubelet[2354]: E0306 01:33:13.900348 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:13.905203 kubelet[2354]: E0306 01:33:13.904414 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:13.905203 kubelet[2354]: E0306 01:33:13.904560 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:13.905203 kubelet[2354]: E0306 01:33:13.904943 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:13.905203 kubelet[2354]: E0306 01:33:13.905038 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:14.905307 kubelet[2354]: E0306 01:33:14.904922 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:14.905307 kubelet[2354]: E0306 01:33:14.905194 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:14.905307 kubelet[2354]: E0306 01:33:14.904950 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:14.905307 kubelet[2354]: E0306 01:33:14.905376 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:14.907714 kubelet[2354]: E0306 01:33:14.906393 2354 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 6 01:33:14.907714 kubelet[2354]: E0306 01:33:14.906516 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:15.078259 kubelet[2354]: E0306 01:33:15.078157 2354 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 6 01:33:15.267388 kubelet[2354]: I0306 01:33:15.267153 2354 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 6 01:33:15.267388 kubelet[2354]: E0306 01:33:15.267261 2354 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Mar 6 01:33:15.273164 kubelet[2354]: I0306 01:33:15.272964 2354 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:15.286345 kubelet[2354]: E0306 01:33:15.286273 2354 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:15.286345 kubelet[2354]: I0306 01:33:15.286308 2354 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:15.289104 kubelet[2354]: E0306 01:33:15.289040 2354 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:15.289104 kubelet[2354]: I0306 01:33:15.289079 2354 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:15.291710 kubelet[2354]: E0306 01:33:15.291575 2354 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:15.445647 kubelet[2354]: I0306 01:33:15.445576 2354 apiserver.go:52] "Watching apiserver" Mar 6 01:33:15.472380 kubelet[2354]: I0306 01:33:15.472232 2354 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 6 01:33:15.904703 kubelet[2354]: I0306 01:33:15.904604 2354 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:15.907816 kubelet[2354]: E0306 01:33:15.907719 2354 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:15.908554 kubelet[2354]: E0306 01:33:15.908091 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:17.781537 systemd[1]: Reloading requested from client PID 2642 ('systemctl') (unit session-7.scope)... Mar 6 01:33:17.781598 systemd[1]: Reloading... Mar 6 01:33:18.089965 zram_generator::config[2681]: No configuration found. Mar 6 01:33:18.314203 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 6 01:33:18.378441 kubelet[2354]: I0306 01:33:18.378130 2354 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:18.389330 kubelet[2354]: E0306 01:33:18.388056 2354 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:18.397192 systemd[1]: Reloading finished in 614 ms. Mar 6 01:33:18.479617 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:33:18.509903 systemd[1]: kubelet.service: Deactivated successfully. Mar 6 01:33:18.511344 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:33:18.555630 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 6 01:33:18.799948 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 6 01:33:18.822440 (kubelet)[2736]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 6 01:33:18.906852 kubelet[2736]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 6 01:33:18.906852 kubelet[2736]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 6 01:33:18.906852 kubelet[2736]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 6 01:33:18.906852 kubelet[2736]: I0306 01:33:18.906111 2736 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 6 01:33:18.916280 kubelet[2736]: I0306 01:33:18.915268 2736 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 6 01:33:18.916280 kubelet[2736]: I0306 01:33:18.915286 2736 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 6 01:33:18.916280 kubelet[2736]: I0306 01:33:18.915937 2736 server.go:956] "Client rotation is on, will bootstrap in background" Mar 6 01:33:18.917147 kubelet[2736]: I0306 01:33:18.917094 2736 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 6 01:33:18.920701 kubelet[2736]: I0306 01:33:18.920640 2736 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 6 01:33:18.926330 kubelet[2736]: E0306 01:33:18.926278 2736 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 6 01:33:18.926387 kubelet[2736]: I0306 01:33:18.926334 2736 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 6 01:33:18.929305 sudo[2752]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 6 01:33:18.929923 sudo[2752]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Mar 6 01:33:18.934057 kubelet[2736]: I0306 01:33:18.933426 2736 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 6 01:33:18.934393 kubelet[2736]: I0306 01:33:18.934301 2736 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 6 01:33:18.934921 kubelet[2736]: I0306 01:33:18.934380 2736 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 6 01:33:18.934921 kubelet[2736]: I0306 01:33:18.934925 2736 topology_manager.go:138] "Creating topology manager with none policy" Mar 6 01:33:18.935187 kubelet[2736]: I0306 01:33:18.934939 2736 container_manager_linux.go:303] "Creating device plugin manager" Mar 6 01:33:18.935187 kubelet[2736]: I0306 01:33:18.935060 2736 state_mem.go:36] "Initialized new in-memory state store" Mar 6 01:33:18.935842 kubelet[2736]: I0306 01:33:18.935735 2736 kubelet.go:480] "Attempting to sync node with API server" Mar 6 01:33:18.936673 kubelet[2736]: I0306 01:33:18.935932 2736 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 6 01:33:18.936673 kubelet[2736]: I0306 01:33:18.936014 2736 kubelet.go:386] "Adding apiserver pod source" Mar 6 01:33:18.936673 kubelet[2736]: I0306 01:33:18.936041 2736 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 6 01:33:18.947737 kubelet[2736]: I0306 01:33:18.947683 2736 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 6 01:33:18.948293 kubelet[2736]: I0306 01:33:18.948209 2736 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 6 01:33:18.957897 kubelet[2736]: I0306 01:33:18.957138 2736 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 6 01:33:18.957897 kubelet[2736]: I0306 01:33:18.957203 2736 server.go:1289] "Started kubelet" Mar 6 01:33:18.957897 kubelet[2736]: I0306 01:33:18.957319 2736 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 6 01:33:18.957897 kubelet[2736]: I0306 01:33:18.957372 2736 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 6 01:33:18.958294 kubelet[2736]: I0306 01:33:18.958220 2736 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 6 01:33:18.960386 kubelet[2736]: I0306 01:33:18.960369 2736 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 6 01:33:18.965368 kubelet[2736]: I0306 01:33:18.965106 2736 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 6 01:33:18.965368 kubelet[2736]: I0306 01:33:18.960853 2736 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 6 01:33:18.966446 kubelet[2736]: I0306 01:33:18.966401 2736 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 6 01:33:18.966446 kubelet[2736]: E0306 01:33:18.964367 2736 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 6 01:33:18.966601 kubelet[2736]: I0306 01:33:18.966545 2736 reconciler.go:26] "Reconciler: start to sync state" Mar 6 01:33:18.967554 kubelet[2736]: I0306 01:33:18.967434 2736 factory.go:223] Registration of the systemd container factory successfully Mar 6 01:33:18.967554 kubelet[2736]: I0306 01:33:18.967517 2736 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 6 01:33:18.969526 kubelet[2736]: I0306 01:33:18.969509 2736 server.go:317] "Adding debug handlers to kubelet server" Mar 6 01:33:18.970555 kubelet[2736]: I0306 01:33:18.970493 2736 factory.go:223] Registration of the containerd container factory successfully Mar 6 01:33:18.994971 kubelet[2736]: I0306 01:33:18.994895 2736 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 6 01:33:19.011714 kubelet[2736]: I0306 01:33:19.011669 2736 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 6 01:33:19.011714 kubelet[2736]: I0306 01:33:19.011703 2736 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 6 01:33:19.011964 kubelet[2736]: I0306 01:33:19.011743 2736 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 6 01:33:19.011964 kubelet[2736]: I0306 01:33:19.011751 2736 kubelet.go:2436] "Starting kubelet main sync loop" Mar 6 01:33:19.011964 kubelet[2736]: E0306 01:33:19.011899 2736 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 6 01:33:19.050283 kubelet[2736]: I0306 01:33:19.050145 2736 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.050597 2736 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.050621 2736 state_mem.go:36] "Initialized new in-memory state store" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.050963 2736 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.050975 2736 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.050992 2736 policy_none.go:49] "None policy: Start" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.051011 2736 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.051032 2736 state_mem.go:35] "Initializing new in-memory state store" Mar 6 01:33:19.051193 kubelet[2736]: I0306 01:33:19.051123 2736 state_mem.go:75] "Updated machine memory state" Mar 6 01:33:19.055471 kubelet[2736]: E0306 01:33:19.055452 2736 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 6 01:33:19.055694 kubelet[2736]: I0306 01:33:19.055681 2736 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 6 01:33:19.055861 kubelet[2736]: I0306 01:33:19.055749 2736 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 6 01:33:19.056474 kubelet[2736]: I0306 01:33:19.056405 2736 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 6 01:33:19.065509 kubelet[2736]: E0306 01:33:19.065475 2736 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 6 01:33:19.113963 kubelet[2736]: I0306 01:33:19.113922 2736 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:19.114444 kubelet[2736]: I0306 01:33:19.113957 2736 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:19.115167 kubelet[2736]: I0306 01:33:19.113974 2736 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:19.124388 kubelet[2736]: E0306 01:33:19.124313 2736 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:19.171018 kubelet[2736]: I0306 01:33:19.170737 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d9b0c3b0c90cd5a17c634cbdf5c19ab7-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"d9b0c3b0c90cd5a17c634cbdf5c19ab7\") " pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:19.171018 kubelet[2736]: I0306 01:33:19.170925 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d9b0c3b0c90cd5a17c634cbdf5c19ab7-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"d9b0c3b0c90cd5a17c634cbdf5c19ab7\") " pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:19.171018 kubelet[2736]: I0306 01:33:19.170975 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:19.171018 kubelet[2736]: I0306 01:33:19.171012 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:19.171276 kubelet[2736]: I0306 01:33:19.171041 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e944e4cb17af904786c3a2e01e298498-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"e944e4cb17af904786c3a2e01e298498\") " pod="kube-system/kube-scheduler-localhost" Mar 6 01:33:19.171276 kubelet[2736]: I0306 01:33:19.171069 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d9b0c3b0c90cd5a17c634cbdf5c19ab7-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"d9b0c3b0c90cd5a17c634cbdf5c19ab7\") " pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:19.171276 kubelet[2736]: I0306 01:33:19.171093 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:19.171276 kubelet[2736]: I0306 01:33:19.171120 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:19.171276 kubelet[2736]: I0306 01:33:19.171148 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:19.184485 kubelet[2736]: I0306 01:33:19.183632 2736 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 6 01:33:19.200098 kubelet[2736]: I0306 01:33:19.200005 2736 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Mar 6 01:33:19.200098 kubelet[2736]: I0306 01:33:19.200110 2736 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 6 01:33:19.424942 kubelet[2736]: E0306 01:33:19.424865 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:19.425160 kubelet[2736]: E0306 01:33:19.425103 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:19.425943 kubelet[2736]: E0306 01:33:19.425885 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:20.083581 kubelet[2736]: I0306 01:33:20.075274 2736 apiserver.go:52] "Watching apiserver" Mar 6 01:33:20.295644 kubelet[2736]: I0306 01:33:20.294932 2736 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:20.296927 kubelet[2736]: I0306 01:33:20.296328 2736 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:20.300608 kubelet[2736]: E0306 01:33:20.297043 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:20.308885 kubelet[2736]: E0306 01:33:20.307609 2736 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Mar 6 01:33:20.308885 kubelet[2736]: E0306 01:33:20.308504 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:20.312113 kubelet[2736]: E0306 01:33:20.312052 2736 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 6 01:33:20.312872 kubelet[2736]: E0306 01:33:20.312278 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:20.328733 kubelet[2736]: I0306 01:33:20.328624 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.328607213 podStartE2EDuration="2.328607213s" podCreationTimestamp="2026-03-06 01:33:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:33:20.328564691 +0000 UTC m=+1.491211710" watchObservedRunningTime="2026-03-06 01:33:20.328607213 +0000 UTC m=+1.491254231" Mar 6 01:33:20.359995 kubelet[2736]: I0306 01:33:20.358745 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.358729429 podStartE2EDuration="1.358729429s" podCreationTimestamp="2026-03-06 01:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:33:20.345891684 +0000 UTC m=+1.508538702" watchObservedRunningTime="2026-03-06 01:33:20.358729429 +0000 UTC m=+1.521376447" Mar 6 01:33:20.367136 kubelet[2736]: I0306 01:33:20.367080 2736 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 6 01:33:20.379120 kubelet[2736]: I0306 01:33:20.379017 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.378997877 podStartE2EDuration="1.378997877s" podCreationTimestamp="2026-03-06 01:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:33:20.361739174 +0000 UTC m=+1.524386212" watchObservedRunningTime="2026-03-06 01:33:20.378997877 +0000 UTC m=+1.541644895" Mar 6 01:33:20.385991 sudo[2752]: pam_unix(sudo:session): session closed for user root Mar 6 01:33:21.297730 kubelet[2736]: E0306 01:33:21.297615 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:21.298864 kubelet[2736]: E0306 01:33:21.297839 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:21.298864 kubelet[2736]: E0306 01:33:21.297898 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:22.485938 sudo[1774]: pam_unix(sudo:session): session closed for user root Mar 6 01:33:22.489521 sshd[1767]: pam_unix(sshd:session): session closed for user core Mar 6 01:33:22.496190 systemd[1]: sshd@6-10.0.0.56:22-10.0.0.1:56420.service: Deactivated successfully. Mar 6 01:33:22.499746 systemd-logind[1561]: Session 7 logged out. Waiting for processes to exit. Mar 6 01:33:22.500058 systemd[1]: session-7.scope: Deactivated successfully. Mar 6 01:33:22.503052 systemd-logind[1561]: Removed session 7. Mar 6 01:33:23.164595 kubelet[2736]: I0306 01:33:23.164549 2736 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 6 01:33:23.166055 containerd[1585]: time="2026-03-06T01:33:23.165210916Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 6 01:33:23.167674 kubelet[2736]: I0306 01:33:23.166915 2736 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 6 01:33:24.195031 kubelet[2736]: I0306 01:33:24.194891 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a213513b-dc36-4cf3-92b4-a54c090ff335-xtables-lock\") pod \"kube-proxy-bqps2\" (UID: \"a213513b-dc36-4cf3-92b4-a54c090ff335\") " pod="kube-system/kube-proxy-bqps2" Mar 6 01:33:24.195031 kubelet[2736]: I0306 01:33:24.194982 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vfmc\" (UniqueName: \"kubernetes.io/projected/a213513b-dc36-4cf3-92b4-a54c090ff335-kube-api-access-4vfmc\") pod \"kube-proxy-bqps2\" (UID: \"a213513b-dc36-4cf3-92b4-a54c090ff335\") " pod="kube-system/kube-proxy-bqps2" Mar 6 01:33:24.195031 kubelet[2736]: I0306 01:33:24.195009 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-cgroup\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.195031 kubelet[2736]: I0306 01:33:24.195030 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-xtables-lock\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.195031 kubelet[2736]: I0306 01:33:24.195060 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-kernel\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196240 kubelet[2736]: I0306 01:33:24.195087 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-run\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196240 kubelet[2736]: I0306 01:33:24.195153 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-bpf-maps\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196240 kubelet[2736]: I0306 01:33:24.195179 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-net\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196240 kubelet[2736]: I0306 01:33:24.195234 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hubble-tls\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196240 kubelet[2736]: I0306 01:33:24.195258 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a213513b-dc36-4cf3-92b4-a54c090ff335-lib-modules\") pod \"kube-proxy-bqps2\" (UID: \"a213513b-dc36-4cf3-92b4-a54c090ff335\") " pod="kube-system/kube-proxy-bqps2" Mar 6 01:33:24.196240 kubelet[2736]: I0306 01:33:24.195279 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hostproc\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196411 kubelet[2736]: I0306 01:33:24.195301 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cni-path\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196411 kubelet[2736]: I0306 01:33:24.195331 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-etc-cni-netd\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196411 kubelet[2736]: I0306 01:33:24.195353 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-lib-modules\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196411 kubelet[2736]: I0306 01:33:24.195377 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3f417a9c-3f35-4a89-84f3-1d6547c520e8-clustermesh-secrets\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196411 kubelet[2736]: I0306 01:33:24.195399 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-config-path\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.196411 kubelet[2736]: I0306 01:33:24.195454 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/a213513b-dc36-4cf3-92b4-a54c090ff335-kube-proxy\") pod \"kube-proxy-bqps2\" (UID: \"a213513b-dc36-4cf3-92b4-a54c090ff335\") " pod="kube-system/kube-proxy-bqps2" Mar 6 01:33:24.196526 kubelet[2736]: I0306 01:33:24.195468 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrmbn\" (UniqueName: \"kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-kube-api-access-xrmbn\") pod \"cilium-km4pj\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " pod="kube-system/cilium-km4pj" Mar 6 01:33:24.385077 kubelet[2736]: E0306 01:33:24.384996 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:24.385697 kubelet[2736]: E0306 01:33:24.385544 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:24.386185 containerd[1585]: time="2026-03-06T01:33:24.386081243Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bqps2,Uid:a213513b-dc36-4cf3-92b4-a54c090ff335,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:24.386976 containerd[1585]: time="2026-03-06T01:33:24.386936390Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-km4pj,Uid:3f417a9c-3f35-4a89-84f3-1d6547c520e8,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:24.463957 containerd[1585]: time="2026-03-06T01:33:24.463410549Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:24.463957 containerd[1585]: time="2026-03-06T01:33:24.463654905Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:24.463957 containerd[1585]: time="2026-03-06T01:33:24.463826817Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:24.464166 containerd[1585]: time="2026-03-06T01:33:24.464049853Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:24.466024 containerd[1585]: time="2026-03-06T01:33:24.465472849Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:24.466024 containerd[1585]: time="2026-03-06T01:33:24.465544162Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:24.466024 containerd[1585]: time="2026-03-06T01:33:24.465574830Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:24.469075 containerd[1585]: time="2026-03-06T01:33:24.468915375Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:24.500415 kubelet[2736]: I0306 01:33:24.500280 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-cs6cx\" (UID: \"e2b8bb96-0c8d-467a-ace9-fb9604b61c04\") " pod="kube-system/cilium-operator-6c4d7847fc-cs6cx" Mar 6 01:33:24.500415 kubelet[2736]: I0306 01:33:24.500382 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmfsc\" (UniqueName: \"kubernetes.io/projected/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-kube-api-access-tmfsc\") pod \"cilium-operator-6c4d7847fc-cs6cx\" (UID: \"e2b8bb96-0c8d-467a-ace9-fb9604b61c04\") " pod="kube-system/cilium-operator-6c4d7847fc-cs6cx" Mar 6 01:33:24.557869 containerd[1585]: time="2026-03-06T01:33:24.557328682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bqps2,Uid:a213513b-dc36-4cf3-92b4-a54c090ff335,Namespace:kube-system,Attempt:0,} returns sandbox id \"7dee426ea78c94b8a8cb445850456c9ecd9702e84950b926516d21e3dc592e32\"" Mar 6 01:33:24.558243 kubelet[2736]: E0306 01:33:24.558175 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:24.560653 containerd[1585]: time="2026-03-06T01:33:24.560485907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-km4pj,Uid:3f417a9c-3f35-4a89-84f3-1d6547c520e8,Namespace:kube-system,Attempt:0,} returns sandbox id \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\"" Mar 6 01:33:24.562084 kubelet[2736]: E0306 01:33:24.561730 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:24.563431 containerd[1585]: time="2026-03-06T01:33:24.563309507Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 6 01:33:24.567047 containerd[1585]: time="2026-03-06T01:33:24.566910147Z" level=info msg="CreateContainer within sandbox \"7dee426ea78c94b8a8cb445850456c9ecd9702e84950b926516d21e3dc592e32\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 6 01:33:24.588162 containerd[1585]: time="2026-03-06T01:33:24.588046916Z" level=info msg="CreateContainer within sandbox \"7dee426ea78c94b8a8cb445850456c9ecd9702e84950b926516d21e3dc592e32\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"60ed50b71687c6e1ccb00c81edf87fd54b3b449dcfe6682f7c4af59810326dcd\"" Mar 6 01:33:24.590878 containerd[1585]: time="2026-03-06T01:33:24.589080444Z" level=info msg="StartContainer for \"60ed50b71687c6e1ccb00c81edf87fd54b3b449dcfe6682f7c4af59810326dcd\"" Mar 6 01:33:24.666815 kubelet[2736]: E0306 01:33:24.666734 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:24.667981 containerd[1585]: time="2026-03-06T01:33:24.667872054Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-cs6cx,Uid:e2b8bb96-0c8d-467a-ace9-fb9604b61c04,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:24.735419 containerd[1585]: time="2026-03-06T01:33:24.735293927Z" level=info msg="StartContainer for \"60ed50b71687c6e1ccb00c81edf87fd54b3b449dcfe6682f7c4af59810326dcd\" returns successfully" Mar 6 01:33:24.756134 containerd[1585]: time="2026-03-06T01:33:24.756041559Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:24.756462 containerd[1585]: time="2026-03-06T01:33:24.756348782Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:24.756567 containerd[1585]: time="2026-03-06T01:33:24.756538045Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:24.756952 containerd[1585]: time="2026-03-06T01:33:24.756920389Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:24.886049 containerd[1585]: time="2026-03-06T01:33:24.885461643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-cs6cx,Uid:e2b8bb96-0c8d-467a-ace9-fb9604b61c04,Namespace:kube-system,Attempt:0,} returns sandbox id \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\"" Mar 6 01:33:24.888481 kubelet[2736]: E0306 01:33:24.888438 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:25.317246 kubelet[2736]: E0306 01:33:25.315720 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:26.705897 kubelet[2736]: E0306 01:33:26.705709 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:26.726690 kubelet[2736]: I0306 01:33:26.725666 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-bqps2" podStartSLOduration=2.725645484 podStartE2EDuration="2.725645484s" podCreationTimestamp="2026-03-06 01:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:33:25.32711601 +0000 UTC m=+6.489763048" watchObservedRunningTime="2026-03-06 01:33:26.725645484 +0000 UTC m=+7.888292512" Mar 6 01:33:27.324958 kubelet[2736]: E0306 01:33:27.324907 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:28.554959 kubelet[2736]: E0306 01:33:28.554679 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:28.993109 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3727243172.mount: Deactivated successfully. Mar 6 01:33:29.353263 kubelet[2736]: E0306 01:33:29.345923 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:30.241011 kubelet[2736]: E0306 01:33:30.240970 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:30.348246 kubelet[2736]: E0306 01:33:30.347973 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:31.440302 containerd[1585]: time="2026-03-06T01:33:31.440191242Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:31.441545 containerd[1585]: time="2026-03-06T01:33:31.441492387Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Mar 6 01:33:31.443444 containerd[1585]: time="2026-03-06T01:33:31.443385624Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:31.445561 containerd[1585]: time="2026-03-06T01:33:31.445394012Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 6.882030714s" Mar 6 01:33:31.445561 containerd[1585]: time="2026-03-06T01:33:31.445444256Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Mar 6 01:33:31.448870 containerd[1585]: time="2026-03-06T01:33:31.448726964Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 6 01:33:31.470256 containerd[1585]: time="2026-03-06T01:33:31.470050554Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 6 01:33:31.486344 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount569881819.mount: Deactivated successfully. Mar 6 01:33:31.490513 containerd[1585]: time="2026-03-06T01:33:31.490448631Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\"" Mar 6 01:33:31.492333 containerd[1585]: time="2026-03-06T01:33:31.492245177Z" level=info msg="StartContainer for \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\"" Mar 6 01:33:31.620423 containerd[1585]: time="2026-03-06T01:33:31.620332098Z" level=info msg="StartContainer for \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\" returns successfully" Mar 6 01:33:31.898452 containerd[1585]: time="2026-03-06T01:33:31.898366738Z" level=info msg="shim disconnected" id=445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e namespace=k8s.io Mar 6 01:33:31.898452 containerd[1585]: time="2026-03-06T01:33:31.898428403Z" level=warning msg="cleaning up after shim disconnected" id=445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e namespace=k8s.io Mar 6 01:33:31.898452 containerd[1585]: time="2026-03-06T01:33:31.898444152Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:33:31.921540 containerd[1585]: time="2026-03-06T01:33:31.921442704Z" level=warning msg="cleanup warnings time=\"2026-03-06T01:33:31Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 6 01:33:32.356153 kubelet[2736]: E0306 01:33:32.355613 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:32.360423 containerd[1585]: time="2026-03-06T01:33:32.360350121Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 6 01:33:32.379989 containerd[1585]: time="2026-03-06T01:33:32.379914540Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\"" Mar 6 01:33:32.382057 containerd[1585]: time="2026-03-06T01:33:32.380939438Z" level=info msg="StartContainer for \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\"" Mar 6 01:33:32.487343 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e-rootfs.mount: Deactivated successfully. Mar 6 01:33:32.491393 containerd[1585]: time="2026-03-06T01:33:32.490068000Z" level=info msg="StartContainer for \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\" returns successfully" Mar 6 01:33:32.501388 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 6 01:33:32.502051 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 6 01:33:32.502226 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Mar 6 01:33:32.510231 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 6 01:33:32.543178 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22-rootfs.mount: Deactivated successfully. Mar 6 01:33:32.552461 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 6 01:33:32.555694 containerd[1585]: time="2026-03-06T01:33:32.555445056Z" level=info msg="shim disconnected" id=93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22 namespace=k8s.io Mar 6 01:33:32.555694 containerd[1585]: time="2026-03-06T01:33:32.555495029Z" level=warning msg="cleaning up after shim disconnected" id=93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22 namespace=k8s.io Mar 6 01:33:32.555694 containerd[1585]: time="2026-03-06T01:33:32.555504317Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:33:33.152281 containerd[1585]: time="2026-03-06T01:33:33.152202282Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:33.154027 containerd[1585]: time="2026-03-06T01:33:33.153740788Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Mar 6 01:33:33.155024 containerd[1585]: time="2026-03-06T01:33:33.154976565Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 6 01:33:33.157574 containerd[1585]: time="2026-03-06T01:33:33.157482514Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 1.708622491s" Mar 6 01:33:33.157574 containerd[1585]: time="2026-03-06T01:33:33.157552906Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Mar 6 01:33:33.171313 containerd[1585]: time="2026-03-06T01:33:33.171255166Z" level=info msg="CreateContainer within sandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 6 01:33:33.191033 containerd[1585]: time="2026-03-06T01:33:33.190949275Z" level=info msg="CreateContainer within sandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\"" Mar 6 01:33:33.192027 containerd[1585]: time="2026-03-06T01:33:33.191918949Z" level=info msg="StartContainer for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\"" Mar 6 01:33:33.311658 containerd[1585]: time="2026-03-06T01:33:33.311579936Z" level=info msg="StartContainer for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" returns successfully" Mar 6 01:33:33.365561 kubelet[2736]: E0306 01:33:33.365448 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:33.372477 kubelet[2736]: E0306 01:33:33.371553 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:33.381023 containerd[1585]: time="2026-03-06T01:33:33.380934470Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 6 01:33:33.387859 kubelet[2736]: I0306 01:33:33.385430 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-cs6cx" podStartSLOduration=1.117716094 podStartE2EDuration="9.385411308s" podCreationTimestamp="2026-03-06 01:33:24 +0000 UTC" firstStartedPulling="2026-03-06 01:33:24.89070318 +0000 UTC m=+6.053350219" lastFinishedPulling="2026-03-06 01:33:33.158398415 +0000 UTC m=+14.321045433" observedRunningTime="2026-03-06 01:33:33.383433197 +0000 UTC m=+14.546080215" watchObservedRunningTime="2026-03-06 01:33:33.385411308 +0000 UTC m=+14.548058337" Mar 6 01:33:33.474735 containerd[1585]: time="2026-03-06T01:33:33.474256328Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\"" Mar 6 01:33:33.476449 containerd[1585]: time="2026-03-06T01:33:33.476071568Z" level=info msg="StartContainer for \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\"" Mar 6 01:33:33.657303 containerd[1585]: time="2026-03-06T01:33:33.657229128Z" level=info msg="StartContainer for \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\" returns successfully" Mar 6 01:33:33.711012 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311-rootfs.mount: Deactivated successfully. Mar 6 01:33:33.716374 containerd[1585]: time="2026-03-06T01:33:33.716095445Z" level=info msg="shim disconnected" id=2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311 namespace=k8s.io Mar 6 01:33:33.716374 containerd[1585]: time="2026-03-06T01:33:33.716148063Z" level=warning msg="cleaning up after shim disconnected" id=2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311 namespace=k8s.io Mar 6 01:33:33.716374 containerd[1585]: time="2026-03-06T01:33:33.716157671Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:33:34.384517 kubelet[2736]: E0306 01:33:34.382710 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:34.384517 kubelet[2736]: E0306 01:33:34.382966 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:34.429425 containerd[1585]: time="2026-03-06T01:33:34.429335522Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 6 01:33:34.453816 containerd[1585]: time="2026-03-06T01:33:34.453685097Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\"" Mar 6 01:33:34.455174 containerd[1585]: time="2026-03-06T01:33:34.454594986Z" level=info msg="StartContainer for \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\"" Mar 6 01:33:34.485594 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3831061754.mount: Deactivated successfully. Mar 6 01:33:34.538406 containerd[1585]: time="2026-03-06T01:33:34.538276220Z" level=info msg="StartContainer for \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\" returns successfully" Mar 6 01:33:34.565256 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232-rootfs.mount: Deactivated successfully. Mar 6 01:33:34.573155 containerd[1585]: time="2026-03-06T01:33:34.573020025Z" level=info msg="shim disconnected" id=48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232 namespace=k8s.io Mar 6 01:33:34.573155 containerd[1585]: time="2026-03-06T01:33:34.573118619Z" level=warning msg="cleaning up after shim disconnected" id=48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232 namespace=k8s.io Mar 6 01:33:34.573155 containerd[1585]: time="2026-03-06T01:33:34.573136984Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:33:35.387187 kubelet[2736]: E0306 01:33:35.387061 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:35.409920 containerd[1585]: time="2026-03-06T01:33:35.397868875Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 6 01:33:35.441221 containerd[1585]: time="2026-03-06T01:33:35.441141604Z" level=info msg="CreateContainer within sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\"" Mar 6 01:33:35.443048 containerd[1585]: time="2026-03-06T01:33:35.441952448Z" level=info msg="StartContainer for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\"" Mar 6 01:33:35.529136 containerd[1585]: time="2026-03-06T01:33:35.529020200Z" level=info msg="StartContainer for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" returns successfully" Mar 6 01:33:35.700749 kubelet[2736]: I0306 01:33:35.700625 2736 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Mar 6 01:33:35.908097 kubelet[2736]: I0306 01:33:35.907994 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgfs8\" (UniqueName: \"kubernetes.io/projected/c75a1a18-dc99-4124-9c15-9f73eb54b5da-kube-api-access-cgfs8\") pod \"coredns-674b8bbfcf-6n6g7\" (UID: \"c75a1a18-dc99-4124-9c15-9f73eb54b5da\") " pod="kube-system/coredns-674b8bbfcf-6n6g7" Mar 6 01:33:35.908097 kubelet[2736]: I0306 01:33:35.908038 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/825f9350-e0aa-4afc-8157-f5cb4f64731d-config-volume\") pod \"coredns-674b8bbfcf-hqt2p\" (UID: \"825f9350-e0aa-4afc-8157-f5cb4f64731d\") " pod="kube-system/coredns-674b8bbfcf-hqt2p" Mar 6 01:33:35.908097 kubelet[2736]: I0306 01:33:35.908061 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c75a1a18-dc99-4124-9c15-9f73eb54b5da-config-volume\") pod \"coredns-674b8bbfcf-6n6g7\" (UID: \"c75a1a18-dc99-4124-9c15-9f73eb54b5da\") " pod="kube-system/coredns-674b8bbfcf-6n6g7" Mar 6 01:33:35.908097 kubelet[2736]: I0306 01:33:35.908093 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbssl\" (UniqueName: \"kubernetes.io/projected/825f9350-e0aa-4afc-8157-f5cb4f64731d-kube-api-access-fbssl\") pod \"coredns-674b8bbfcf-hqt2p\" (UID: \"825f9350-e0aa-4afc-8157-f5cb4f64731d\") " pod="kube-system/coredns-674b8bbfcf-hqt2p" Mar 6 01:33:36.071721 kubelet[2736]: E0306 01:33:36.071436 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:36.073083 containerd[1585]: time="2026-03-06T01:33:36.072958794Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-6n6g7,Uid:c75a1a18-dc99-4124-9c15-9f73eb54b5da,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:36.077487 kubelet[2736]: E0306 01:33:36.077272 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:36.078904 containerd[1585]: time="2026-03-06T01:33:36.078141250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hqt2p,Uid:825f9350-e0aa-4afc-8157-f5cb4f64731d,Namespace:kube-system,Attempt:0,}" Mar 6 01:33:36.393876 kubelet[2736]: E0306 01:33:36.393691 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:36.428898 kubelet[2736]: I0306 01:33:36.426411 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-km4pj" podStartSLOduration=5.540548984 podStartE2EDuration="12.426391181s" podCreationTimestamp="2026-03-06 01:33:24 +0000 UTC" firstStartedPulling="2026-03-06 01:33:24.562485658 +0000 UTC m=+5.725132686" lastFinishedPulling="2026-03-06 01:33:31.448327865 +0000 UTC m=+12.610974883" observedRunningTime="2026-03-06 01:33:36.426360478 +0000 UTC m=+17.589007496" watchObservedRunningTime="2026-03-06 01:33:36.426391181 +0000 UTC m=+17.589038199" Mar 6 01:33:37.397185 kubelet[2736]: E0306 01:33:37.397055 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:37.902574 systemd-networkd[1245]: cilium_host: Link UP Mar 6 01:33:37.902972 systemd-networkd[1245]: cilium_net: Link UP Mar 6 01:33:37.903265 systemd-networkd[1245]: cilium_net: Gained carrier Mar 6 01:33:37.903572 systemd-networkd[1245]: cilium_host: Gained carrier Mar 6 01:33:38.397645 systemd-networkd[1245]: cilium_vxlan: Link UP Mar 6 01:33:38.397657 systemd-networkd[1245]: cilium_vxlan: Gained carrier Mar 6 01:33:38.404498 kubelet[2736]: E0306 01:33:38.402987 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:39.268113 systemd-networkd[1245]: cilium_net: Gained IPv6LL Mar 6 01:33:39.270554 systemd-networkd[1245]: cilium_host: Gained IPv6LL Mar 6 01:33:39.489936 kernel: NET: Registered PF_ALG protocol family Mar 6 01:33:40.393350 systemd-networkd[1245]: cilium_vxlan: Gained IPv6LL Mar 6 01:33:46.804074 kubelet[2736]: E0306 01:33:46.803412 2736 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.413s" Mar 6 01:33:48.443219 systemd-networkd[1245]: lxc_health: Link UP Mar 6 01:33:48.467490 systemd-networkd[1245]: lxc_health: Gained carrier Mar 6 01:33:48.762742 systemd-networkd[1245]: lxc30695af5621e: Link UP Mar 6 01:33:48.777882 kernel: eth0: renamed from tmp94280 Mar 6 01:33:48.785124 systemd-networkd[1245]: lxc30695af5621e: Gained carrier Mar 6 01:33:49.050150 systemd-networkd[1245]: lxca8bad5f6a210: Link UP Mar 6 01:33:49.060903 kernel: eth0: renamed from tmpd1607 Mar 6 01:33:49.069479 systemd-networkd[1245]: lxca8bad5f6a210: Gained carrier Mar 6 01:33:50.090277 systemd-networkd[1245]: lxc30695af5621e: Gained IPv6LL Mar 6 01:33:50.279115 systemd-networkd[1245]: lxc_health: Gained IPv6LL Mar 6 01:33:50.394117 kubelet[2736]: E0306 01:33:50.394079 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:50.984337 systemd-networkd[1245]: lxca8bad5f6a210: Gained IPv6LL Mar 6 01:33:53.403230 kubelet[2736]: I0306 01:33:53.403142 2736 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 6 01:33:53.404685 kubelet[2736]: E0306 01:33:53.404090 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:53.410364 containerd[1585]: time="2026-03-06T01:33:53.407238080Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:53.410364 containerd[1585]: time="2026-03-06T01:33:53.407439406Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:53.410364 containerd[1585]: time="2026-03-06T01:33:53.407451459Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:53.410364 containerd[1585]: time="2026-03-06T01:33:53.407618280Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:53.429948 containerd[1585]: time="2026-03-06T01:33:53.429688370Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:33:53.430513 containerd[1585]: time="2026-03-06T01:33:53.430311053Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:33:53.430513 containerd[1585]: time="2026-03-06T01:33:53.430431468Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:53.432526 containerd[1585]: time="2026-03-06T01:33:53.432395415Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:33:53.504936 systemd-resolved[1460]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 01:33:53.508971 systemd-resolved[1460]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 6 01:33:53.556409 containerd[1585]: time="2026-03-06T01:33:53.556190163Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hqt2p,Uid:825f9350-e0aa-4afc-8157-f5cb4f64731d,Namespace:kube-system,Attempt:0,} returns sandbox id \"942809040ba2b58cdd599644f3d4004dbdcb0d2833fb076e4cb3ce982148f481\"" Mar 6 01:33:53.557870 kubelet[2736]: E0306 01:33:53.557566 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:53.561325 containerd[1585]: time="2026-03-06T01:33:53.561296226Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-6n6g7,Uid:c75a1a18-dc99-4124-9c15-9f73eb54b5da,Namespace:kube-system,Attempt:0,} returns sandbox id \"d16071c47e2510fa2fff10aad825df03a16c42b7e23ecff890ae7e20f3fedf83\"" Mar 6 01:33:53.562718 kubelet[2736]: E0306 01:33:53.562253 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:53.567991 containerd[1585]: time="2026-03-06T01:33:53.567682239Z" level=info msg="CreateContainer within sandbox \"942809040ba2b58cdd599644f3d4004dbdcb0d2833fb076e4cb3ce982148f481\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 6 01:33:53.572022 containerd[1585]: time="2026-03-06T01:33:53.571833461Z" level=info msg="CreateContainer within sandbox \"d16071c47e2510fa2fff10aad825df03a16c42b7e23ecff890ae7e20f3fedf83\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 6 01:33:53.598473 containerd[1585]: time="2026-03-06T01:33:53.598364049Z" level=info msg="CreateContainer within sandbox \"d16071c47e2510fa2fff10aad825df03a16c42b7e23ecff890ae7e20f3fedf83\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b04e914bb8bc0a75cf0ecbab05dd848103db349ebf9c64c67575fc98189749e3\"" Mar 6 01:33:53.599784 containerd[1585]: time="2026-03-06T01:33:53.599657937Z" level=info msg="StartContainer for \"b04e914bb8bc0a75cf0ecbab05dd848103db349ebf9c64c67575fc98189749e3\"" Mar 6 01:33:53.604240 containerd[1585]: time="2026-03-06T01:33:53.604122563Z" level=info msg="CreateContainer within sandbox \"942809040ba2b58cdd599644f3d4004dbdcb0d2833fb076e4cb3ce982148f481\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"07d1db91e89be228ee5389f26a6704160bcd7e2d009d24b3a8efbecf359545b1\"" Mar 6 01:33:53.606705 containerd[1585]: time="2026-03-06T01:33:53.606641159Z" level=info msg="StartContainer for \"07d1db91e89be228ee5389f26a6704160bcd7e2d009d24b3a8efbecf359545b1\"" Mar 6 01:33:53.703605 containerd[1585]: time="2026-03-06T01:33:53.703140414Z" level=info msg="StartContainer for \"b04e914bb8bc0a75cf0ecbab05dd848103db349ebf9c64c67575fc98189749e3\" returns successfully" Mar 6 01:33:53.724616 containerd[1585]: time="2026-03-06T01:33:53.724502683Z" level=info msg="StartContainer for \"07d1db91e89be228ee5389f26a6704160bcd7e2d009d24b3a8efbecf359545b1\" returns successfully" Mar 6 01:33:54.082670 kubelet[2736]: E0306 01:33:54.082155 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:54.086071 kubelet[2736]: E0306 01:33:54.085972 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:54.086491 kubelet[2736]: E0306 01:33:54.086430 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:54.097904 kubelet[2736]: I0306 01:33:54.097311 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-hqt2p" podStartSLOduration=30.097299761 podStartE2EDuration="30.097299761s" podCreationTimestamp="2026-03-06 01:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:33:54.096576412 +0000 UTC m=+35.259223431" watchObservedRunningTime="2026-03-06 01:33:54.097299761 +0000 UTC m=+35.259946779" Mar 6 01:33:54.132748 kubelet[2736]: I0306 01:33:54.132425 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-6n6g7" podStartSLOduration=30.132408787 podStartE2EDuration="30.132408787s" podCreationTimestamp="2026-03-06 01:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:33:54.13212434 +0000 UTC m=+35.294771358" watchObservedRunningTime="2026-03-06 01:33:54.132408787 +0000 UTC m=+35.295055804" Mar 6 01:33:55.088347 kubelet[2736]: E0306 01:33:55.088166 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:55.088347 kubelet[2736]: E0306 01:33:55.088292 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:56.090284 kubelet[2736]: E0306 01:33:56.090182 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:33:56.090879 kubelet[2736]: E0306 01:33:56.090328 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:34:03.467411 systemd[1]: Started sshd@7-10.0.0.56:22-10.0.0.1:56922.service - OpenSSH per-connection server daemon (10.0.0.1:56922). Mar 6 01:34:03.513296 sshd[4145]: Accepted publickey for core from 10.0.0.1 port 56922 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:03.515324 sshd[4145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:03.522204 systemd-logind[1561]: New session 8 of user core. Mar 6 01:34:03.528194 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 6 01:34:03.929437 sshd[4145]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:03.934871 systemd[1]: sshd@7-10.0.0.56:22-10.0.0.1:56922.service: Deactivated successfully. Mar 6 01:34:03.937684 systemd-logind[1561]: Session 8 logged out. Waiting for processes to exit. Mar 6 01:34:03.938149 systemd[1]: session-8.scope: Deactivated successfully. Mar 6 01:34:03.940094 systemd-logind[1561]: Removed session 8. Mar 6 01:34:08.949458 systemd[1]: Started sshd@8-10.0.0.56:22-10.0.0.1:56932.service - OpenSSH per-connection server daemon (10.0.0.1:56932). Mar 6 01:34:08.985348 sshd[4174]: Accepted publickey for core from 10.0.0.1 port 56932 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:08.988025 sshd[4174]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:08.994708 systemd-logind[1561]: New session 9 of user core. Mar 6 01:34:09.002246 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 6 01:34:09.156995 sshd[4174]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:09.162463 systemd[1]: sshd@8-10.0.0.56:22-10.0.0.1:56932.service: Deactivated successfully. Mar 6 01:34:09.166041 systemd-logind[1561]: Session 9 logged out. Waiting for processes to exit. Mar 6 01:34:09.166222 systemd[1]: session-9.scope: Deactivated successfully. Mar 6 01:34:09.168305 systemd-logind[1561]: Removed session 9. Mar 6 01:34:14.173188 systemd[1]: Started sshd@9-10.0.0.56:22-10.0.0.1:42040.service - OpenSSH per-connection server daemon (10.0.0.1:42040). Mar 6 01:34:14.215031 sshd[4190]: Accepted publickey for core from 10.0.0.1 port 42040 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:14.217909 sshd[4190]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:14.224467 systemd-logind[1561]: New session 10 of user core. Mar 6 01:34:14.239661 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 6 01:34:14.393635 sshd[4190]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:14.399424 systemd[1]: sshd@9-10.0.0.56:22-10.0.0.1:42040.service: Deactivated successfully. Mar 6 01:34:14.403403 systemd[1]: session-10.scope: Deactivated successfully. Mar 6 01:34:14.404930 systemd-logind[1561]: Session 10 logged out. Waiting for processes to exit. Mar 6 01:34:14.406685 systemd-logind[1561]: Removed session 10. Mar 6 01:34:19.408234 systemd[1]: Started sshd@10-10.0.0.56:22-10.0.0.1:42050.service - OpenSSH per-connection server daemon (10.0.0.1:42050). Mar 6 01:34:19.439181 sshd[4209]: Accepted publickey for core from 10.0.0.1 port 42050 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:19.441552 sshd[4209]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:19.454525 systemd-logind[1561]: New session 11 of user core. Mar 6 01:34:19.463189 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 6 01:34:19.795605 sshd[4209]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:19.811198 systemd[1]: Started sshd@11-10.0.0.56:22-10.0.0.1:42064.service - OpenSSH per-connection server daemon (10.0.0.1:42064). Mar 6 01:34:19.812206 systemd[1]: sshd@10-10.0.0.56:22-10.0.0.1:42050.service: Deactivated successfully. Mar 6 01:34:19.837302 systemd[1]: session-11.scope: Deactivated successfully. Mar 6 01:34:19.841469 systemd-logind[1561]: Session 11 logged out. Waiting for processes to exit. Mar 6 01:34:19.845626 systemd-logind[1561]: Removed session 11. Mar 6 01:34:19.896900 sshd[4226]: Accepted publickey for core from 10.0.0.1 port 42064 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:19.899870 sshd[4226]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:19.913982 systemd-logind[1561]: New session 12 of user core. Mar 6 01:34:19.924146 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 6 01:34:20.111494 sshd[4226]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:20.121154 systemd[1]: Started sshd@12-10.0.0.56:22-10.0.0.1:42076.service - OpenSSH per-connection server daemon (10.0.0.1:42076). Mar 6 01:34:20.121731 systemd[1]: sshd@11-10.0.0.56:22-10.0.0.1:42064.service: Deactivated successfully. Mar 6 01:34:20.128606 systemd[1]: session-12.scope: Deactivated successfully. Mar 6 01:34:20.140146 systemd-logind[1561]: Session 12 logged out. Waiting for processes to exit. Mar 6 01:34:20.142721 systemd-logind[1561]: Removed session 12. Mar 6 01:34:20.175475 sshd[4240]: Accepted publickey for core from 10.0.0.1 port 42076 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:20.177802 sshd[4240]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:20.184425 systemd-logind[1561]: New session 13 of user core. Mar 6 01:34:20.191224 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 6 01:34:20.381466 sshd[4240]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:20.388380 systemd[1]: sshd@12-10.0.0.56:22-10.0.0.1:42076.service: Deactivated successfully. Mar 6 01:34:20.393318 systemd[1]: session-13.scope: Deactivated successfully. Mar 6 01:34:20.394625 systemd-logind[1561]: Session 13 logged out. Waiting for processes to exit. Mar 6 01:34:20.396733 systemd-logind[1561]: Removed session 13. Mar 6 01:34:26.873150 systemd[1]: Started sshd@13-10.0.0.56:22-10.0.0.1:53848.service - OpenSSH per-connection server daemon (10.0.0.1:53848). Mar 6 01:34:26.923355 sshd[4260]: Accepted publickey for core from 10.0.0.1 port 53848 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:26.925723 sshd[4260]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:26.935647 systemd-logind[1561]: New session 14 of user core. Mar 6 01:34:26.944254 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 6 01:34:27.155438 sshd[4260]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:27.162087 systemd[1]: sshd@13-10.0.0.56:22-10.0.0.1:53848.service: Deactivated successfully. Mar 6 01:34:27.165908 systemd-logind[1561]: Session 14 logged out. Waiting for processes to exit. Mar 6 01:34:27.167489 systemd[1]: session-14.scope: Deactivated successfully. Mar 6 01:34:27.169659 systemd-logind[1561]: Removed session 14. Mar 6 01:34:29.013709 kubelet[2736]: E0306 01:34:29.013543 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:34:32.167436 systemd[1]: Started sshd@14-10.0.0.56:22-10.0.0.1:39628.service - OpenSSH per-connection server daemon (10.0.0.1:39628). Mar 6 01:34:32.203191 sshd[4277]: Accepted publickey for core from 10.0.0.1 port 39628 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:32.205331 sshd[4277]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:32.212501 systemd-logind[1561]: New session 15 of user core. Mar 6 01:34:32.220230 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 6 01:34:32.390343 sshd[4277]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:32.396192 systemd[1]: sshd@14-10.0.0.56:22-10.0.0.1:39628.service: Deactivated successfully. Mar 6 01:34:32.399904 systemd-logind[1561]: Session 15 logged out. Waiting for processes to exit. Mar 6 01:34:32.400034 systemd[1]: session-15.scope: Deactivated successfully. Mar 6 01:34:32.402368 systemd-logind[1561]: Removed session 15. Mar 6 01:34:36.901145 kubelet[2736]: E0306 01:34:36.897742 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:34:36.922450 kubelet[2736]: E0306 01:34:36.919347 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:34:37.436931 systemd[1]: Started sshd@15-10.0.0.56:22-10.0.0.1:39636.service - OpenSSH per-connection server daemon (10.0.0.1:39636). Mar 6 01:34:37.529397 sshd[4292]: Accepted publickey for core from 10.0.0.1 port 39636 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:37.532482 sshd[4292]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:37.544293 systemd-logind[1561]: New session 16 of user core. Mar 6 01:34:37.566284 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 6 01:34:38.940100 sshd[4292]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:38.960619 systemd[1]: sshd@15-10.0.0.56:22-10.0.0.1:39636.service: Deactivated successfully. Mar 6 01:34:38.966261 systemd[1]: session-16.scope: Deactivated successfully. Mar 6 01:34:38.968886 systemd-logind[1561]: Session 16 logged out. Waiting for processes to exit. Mar 6 01:34:38.971352 systemd-logind[1561]: Removed session 16. Mar 6 01:34:39.018498 kubelet[2736]: E0306 01:34:39.017711 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:34:43.952166 systemd[1]: Started sshd@16-10.0.0.56:22-10.0.0.1:57722.service - OpenSSH per-connection server daemon (10.0.0.1:57722). Mar 6 01:34:43.999922 sshd[4307]: Accepted publickey for core from 10.0.0.1 port 57722 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:44.002596 sshd[4307]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:44.009365 systemd-logind[1561]: New session 17 of user core. Mar 6 01:34:44.015400 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 6 01:34:44.159088 sshd[4307]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:44.168276 systemd[1]: Started sshd@17-10.0.0.56:22-10.0.0.1:57730.service - OpenSSH per-connection server daemon (10.0.0.1:57730). Mar 6 01:34:44.169307 systemd[1]: sshd@16-10.0.0.56:22-10.0.0.1:57722.service: Deactivated successfully. Mar 6 01:34:44.172700 systemd[1]: session-17.scope: Deactivated successfully. Mar 6 01:34:44.177058 systemd-logind[1561]: Session 17 logged out. Waiting for processes to exit. Mar 6 01:34:44.179457 systemd-logind[1561]: Removed session 17. Mar 6 01:34:44.211112 sshd[4321]: Accepted publickey for core from 10.0.0.1 port 57730 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:44.213528 sshd[4321]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:44.220053 systemd-logind[1561]: New session 18 of user core. Mar 6 01:34:44.232234 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 6 01:34:44.595640 sshd[4321]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:44.610209 systemd[1]: Started sshd@18-10.0.0.56:22-10.0.0.1:57732.service - OpenSSH per-connection server daemon (10.0.0.1:57732). Mar 6 01:34:44.610741 systemd[1]: sshd@17-10.0.0.56:22-10.0.0.1:57730.service: Deactivated successfully. Mar 6 01:34:44.616209 systemd-logind[1561]: Session 18 logged out. Waiting for processes to exit. Mar 6 01:34:44.617337 systemd[1]: session-18.scope: Deactivated successfully. Mar 6 01:34:44.620156 systemd-logind[1561]: Removed session 18. Mar 6 01:34:44.659309 sshd[4334]: Accepted publickey for core from 10.0.0.1 port 57732 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:44.662108 sshd[4334]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:44.670228 systemd-logind[1561]: New session 19 of user core. Mar 6 01:34:44.684695 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 6 01:34:45.410053 sshd[4334]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:45.417925 systemd[1]: Started sshd@19-10.0.0.56:22-10.0.0.1:57746.service - OpenSSH per-connection server daemon (10.0.0.1:57746). Mar 6 01:34:45.418673 systemd[1]: sshd@18-10.0.0.56:22-10.0.0.1:57732.service: Deactivated successfully. Mar 6 01:34:45.433409 systemd[1]: session-19.scope: Deactivated successfully. Mar 6 01:34:45.443701 systemd-logind[1561]: Session 19 logged out. Waiting for processes to exit. Mar 6 01:34:45.450566 systemd-logind[1561]: Removed session 19. Mar 6 01:34:45.473608 sshd[4354]: Accepted publickey for core from 10.0.0.1 port 57746 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:45.475501 sshd[4354]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:45.482541 systemd-logind[1561]: New session 20 of user core. Mar 6 01:34:45.492322 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 6 01:34:45.867374 sshd[4354]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:45.887627 systemd[1]: Started sshd@20-10.0.0.56:22-10.0.0.1:57750.service - OpenSSH per-connection server daemon (10.0.0.1:57750). Mar 6 01:34:45.888350 systemd[1]: sshd@19-10.0.0.56:22-10.0.0.1:57746.service: Deactivated successfully. Mar 6 01:34:45.896326 systemd[1]: session-20.scope: Deactivated successfully. Mar 6 01:34:45.897658 systemd-logind[1561]: Session 20 logged out. Waiting for processes to exit. Mar 6 01:34:45.899907 systemd-logind[1561]: Removed session 20. Mar 6 01:34:45.948935 sshd[4370]: Accepted publickey for core from 10.0.0.1 port 57750 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:45.951041 sshd[4370]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:45.958205 systemd-logind[1561]: New session 21 of user core. Mar 6 01:34:45.970301 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 6 01:34:46.116248 sshd[4370]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:46.121152 systemd[1]: sshd@20-10.0.0.56:22-10.0.0.1:57750.service: Deactivated successfully. Mar 6 01:34:46.123948 systemd[1]: session-21.scope: Deactivated successfully. Mar 6 01:34:46.123993 systemd-logind[1561]: Session 21 logged out. Waiting for processes to exit. Mar 6 01:34:46.125899 systemd-logind[1561]: Removed session 21. Mar 6 01:34:51.131268 systemd[1]: Started sshd@21-10.0.0.56:22-10.0.0.1:57764.service - OpenSSH per-connection server daemon (10.0.0.1:57764). Mar 6 01:34:51.174745 sshd[4387]: Accepted publickey for core from 10.0.0.1 port 57764 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:51.177569 sshd[4387]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:51.185705 systemd-logind[1561]: New session 22 of user core. Mar 6 01:34:51.200374 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 6 01:34:51.363674 sshd[4387]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:51.369208 systemd[1]: sshd@21-10.0.0.56:22-10.0.0.1:57764.service: Deactivated successfully. Mar 6 01:34:51.373635 systemd[1]: session-22.scope: Deactivated successfully. Mar 6 01:34:51.375222 systemd-logind[1561]: Session 22 logged out. Waiting for processes to exit. Mar 6 01:34:51.376997 systemd-logind[1561]: Removed session 22. Mar 6 01:34:53.014014 kubelet[2736]: E0306 01:34:53.013742 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:34:56.381151 systemd[1]: Started sshd@22-10.0.0.56:22-10.0.0.1:49314.service - OpenSSH per-connection server daemon (10.0.0.1:49314). Mar 6 01:34:56.413399 sshd[4405]: Accepted publickey for core from 10.0.0.1 port 49314 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:34:56.415527 sshd[4405]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:34:56.421391 systemd-logind[1561]: New session 23 of user core. Mar 6 01:34:56.432401 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 6 01:34:56.571259 sshd[4405]: pam_unix(sshd:session): session closed for user core Mar 6 01:34:56.576994 systemd[1]: sshd@22-10.0.0.56:22-10.0.0.1:49314.service: Deactivated successfully. Mar 6 01:34:56.580729 systemd-logind[1561]: Session 23 logged out. Waiting for processes to exit. Mar 6 01:34:56.580918 systemd[1]: session-23.scope: Deactivated successfully. Mar 6 01:34:56.582898 systemd-logind[1561]: Removed session 23. Mar 6 01:35:00.014032 kubelet[2736]: E0306 01:35:00.013962 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:01.013637 kubelet[2736]: E0306 01:35:01.013544 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:01.584315 systemd[1]: Started sshd@23-10.0.0.56:22-10.0.0.1:49318.service - OpenSSH per-connection server daemon (10.0.0.1:49318). Mar 6 01:35:01.622970 sshd[4423]: Accepted publickey for core from 10.0.0.1 port 49318 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:35:01.624921 sshd[4423]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:35:01.632997 systemd-logind[1561]: New session 24 of user core. Mar 6 01:35:01.641622 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 6 01:35:01.773481 sshd[4423]: pam_unix(sshd:session): session closed for user core Mar 6 01:35:01.779210 systemd[1]: sshd@23-10.0.0.56:22-10.0.0.1:49318.service: Deactivated successfully. Mar 6 01:35:01.782339 systemd-logind[1561]: Session 24 logged out. Waiting for processes to exit. Mar 6 01:35:01.782384 systemd[1]: session-24.scope: Deactivated successfully. Mar 6 01:35:01.784968 systemd-logind[1561]: Removed session 24. Mar 6 01:35:06.791378 systemd[1]: Started sshd@24-10.0.0.56:22-10.0.0.1:59560.service - OpenSSH per-connection server daemon (10.0.0.1:59560). Mar 6 01:35:06.828222 sshd[4439]: Accepted publickey for core from 10.0.0.1 port 59560 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:35:06.830695 sshd[4439]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:35:06.837356 systemd-logind[1561]: New session 25 of user core. Mar 6 01:35:06.849657 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 6 01:35:07.009589 sshd[4439]: pam_unix(sshd:session): session closed for user core Mar 6 01:35:07.023459 systemd[1]: Started sshd@25-10.0.0.56:22-10.0.0.1:59566.service - OpenSSH per-connection server daemon (10.0.0.1:59566). Mar 6 01:35:07.025287 systemd[1]: sshd@24-10.0.0.56:22-10.0.0.1:59560.service: Deactivated successfully. Mar 6 01:35:07.030631 systemd[1]: session-25.scope: Deactivated successfully. Mar 6 01:35:07.032372 systemd-logind[1561]: Session 25 logged out. Waiting for processes to exit. Mar 6 01:35:07.034753 systemd-logind[1561]: Removed session 25. Mar 6 01:35:07.070534 sshd[4452]: Accepted publickey for core from 10.0.0.1 port 59566 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:35:07.072670 sshd[4452]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:35:07.079111 systemd-logind[1561]: New session 26 of user core. Mar 6 01:35:07.089365 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 6 01:35:09.099278 containerd[1585]: time="2026-03-06T01:35:09.098832373Z" level=info msg="StopContainer for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" with timeout 30 (s)" Mar 6 01:35:09.101689 containerd[1585]: time="2026-03-06T01:35:09.101656095Z" level=info msg="Stop container \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" with signal terminated" Mar 6 01:35:09.164210 systemd[1]: run-containerd-runc-k8s.io-d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526-runc.UI5iB9.mount: Deactivated successfully. Mar 6 01:35:09.193539 containerd[1585]: time="2026-03-06T01:35:09.193353758Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 6 01:35:09.211492 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37-rootfs.mount: Deactivated successfully. Mar 6 01:35:09.217414 containerd[1585]: time="2026-03-06T01:35:09.217296423Z" level=info msg="StopContainer for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" with timeout 2 (s)" Mar 6 01:35:09.218274 containerd[1585]: time="2026-03-06T01:35:09.218208415Z" level=info msg="Stop container \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" with signal terminated" Mar 6 01:35:09.230485 containerd[1585]: time="2026-03-06T01:35:09.230425565Z" level=info msg="shim disconnected" id=36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37 namespace=k8s.io Mar 6 01:35:09.230905 containerd[1585]: time="2026-03-06T01:35:09.230641989Z" level=warning msg="cleaning up after shim disconnected" id=36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37 namespace=k8s.io Mar 6 01:35:09.230905 containerd[1585]: time="2026-03-06T01:35:09.230663810Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:09.235324 systemd-networkd[1245]: lxc_health: Link DOWN Mar 6 01:35:09.235362 systemd-networkd[1245]: lxc_health: Lost carrier Mar 6 01:35:09.264739 containerd[1585]: time="2026-03-06T01:35:09.264632991Z" level=warning msg="cleanup warnings time=\"2026-03-06T01:35:09Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 6 01:35:09.275645 containerd[1585]: time="2026-03-06T01:35:09.275558020Z" level=info msg="StopContainer for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" returns successfully" Mar 6 01:35:09.283476 containerd[1585]: time="2026-03-06T01:35:09.283421927Z" level=info msg="StopPodSandbox for \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\"" Mar 6 01:35:09.286911 containerd[1585]: time="2026-03-06T01:35:09.283694326Z" level=info msg="Container to stop \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 6 01:35:09.289438 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4-shm.mount: Deactivated successfully. Mar 6 01:35:09.320703 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526-rootfs.mount: Deactivated successfully. Mar 6 01:35:09.343721 containerd[1585]: time="2026-03-06T01:35:09.343554494Z" level=info msg="shim disconnected" id=d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526 namespace=k8s.io Mar 6 01:35:09.343721 containerd[1585]: time="2026-03-06T01:35:09.343640795Z" level=warning msg="cleaning up after shim disconnected" id=d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526 namespace=k8s.io Mar 6 01:35:09.343721 containerd[1585]: time="2026-03-06T01:35:09.343653309Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:09.371052 containerd[1585]: time="2026-03-06T01:35:09.370730328Z" level=info msg="shim disconnected" id=21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4 namespace=k8s.io Mar 6 01:35:09.373153 containerd[1585]: time="2026-03-06T01:35:09.372328171Z" level=warning msg="cleaning up after shim disconnected" id=21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4 namespace=k8s.io Mar 6 01:35:09.373153 containerd[1585]: time="2026-03-06T01:35:09.372358358Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:09.401484 containerd[1585]: time="2026-03-06T01:35:09.401361900Z" level=info msg="StopContainer for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" returns successfully" Mar 6 01:35:09.403193 containerd[1585]: time="2026-03-06T01:35:09.403164599Z" level=info msg="StopPodSandbox for \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\"" Mar 6 01:35:09.403839 containerd[1585]: time="2026-03-06T01:35:09.403559155Z" level=info msg="Container to stop \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 6 01:35:09.403839 containerd[1585]: time="2026-03-06T01:35:09.403580175Z" level=info msg="Container to stop \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 6 01:35:09.403839 containerd[1585]: time="2026-03-06T01:35:09.403592608Z" level=info msg="Container to stop \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 6 01:35:09.403839 containerd[1585]: time="2026-03-06T01:35:09.403602676Z" level=info msg="Container to stop \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 6 01:35:09.403839 containerd[1585]: time="2026-03-06T01:35:09.403612665Z" level=info msg="Container to stop \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 6 01:35:09.411725 containerd[1585]: time="2026-03-06T01:35:09.411643300Z" level=info msg="TearDown network for sandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" successfully" Mar 6 01:35:09.411725 containerd[1585]: time="2026-03-06T01:35:09.411698624Z" level=info msg="StopPodSandbox for \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" returns successfully" Mar 6 01:35:09.475031 containerd[1585]: time="2026-03-06T01:35:09.474954747Z" level=info msg="shim disconnected" id=bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058 namespace=k8s.io Mar 6 01:35:09.475645 containerd[1585]: time="2026-03-06T01:35:09.475401992Z" level=warning msg="cleaning up after shim disconnected" id=bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058 namespace=k8s.io Mar 6 01:35:09.475645 containerd[1585]: time="2026-03-06T01:35:09.475431728Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:09.489319 kubelet[2736]: I0306 01:35:09.487047 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmfsc\" (UniqueName: \"kubernetes.io/projected/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-kube-api-access-tmfsc\") pod \"e2b8bb96-0c8d-467a-ace9-fb9604b61c04\" (UID: \"e2b8bb96-0c8d-467a-ace9-fb9604b61c04\") " Mar 6 01:35:09.489319 kubelet[2736]: I0306 01:35:09.487113 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-cilium-config-path\") pod \"e2b8bb96-0c8d-467a-ace9-fb9604b61c04\" (UID: \"e2b8bb96-0c8d-467a-ace9-fb9604b61c04\") " Mar 6 01:35:09.494485 kubelet[2736]: I0306 01:35:09.494442 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "e2b8bb96-0c8d-467a-ace9-fb9604b61c04" (UID: "e2b8bb96-0c8d-467a-ace9-fb9604b61c04"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 6 01:35:09.497738 kubelet[2736]: I0306 01:35:09.497692 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-kube-api-access-tmfsc" (OuterVolumeSpecName: "kube-api-access-tmfsc") pod "e2b8bb96-0c8d-467a-ace9-fb9604b61c04" (UID: "e2b8bb96-0c8d-467a-ace9-fb9604b61c04"). InnerVolumeSpecName "kube-api-access-tmfsc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 6 01:35:09.509547 containerd[1585]: time="2026-03-06T01:35:09.509451280Z" level=info msg="TearDown network for sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" successfully" Mar 6 01:35:09.509547 containerd[1585]: time="2026-03-06T01:35:09.509521361Z" level=info msg="StopPodSandbox for \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" returns successfully" Mar 6 01:35:09.590357 kubelet[2736]: I0306 01:35:09.589128 2736 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tmfsc\" (UniqueName: \"kubernetes.io/projected/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-kube-api-access-tmfsc\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.590357 kubelet[2736]: I0306 01:35:09.589169 2736 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e2b8bb96-0c8d-467a-ace9-fb9604b61c04-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.690056 kubelet[2736]: I0306 01:35:09.689922 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3f417a9c-3f35-4a89-84f3-1d6547c520e8-clustermesh-secrets\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690056 kubelet[2736]: I0306 01:35:09.690028 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-bpf-maps\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690056 kubelet[2736]: I0306 01:35:09.690054 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-lib-modules\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690382 kubelet[2736]: I0306 01:35:09.690080 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-config-path\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690382 kubelet[2736]: I0306 01:35:09.690107 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-xtables-lock\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690382 kubelet[2736]: I0306 01:35:09.690135 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrmbn\" (UniqueName: \"kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-kube-api-access-xrmbn\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690382 kubelet[2736]: I0306 01:35:09.690140 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.690382 kubelet[2736]: I0306 01:35:09.690165 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-cgroup\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690382 kubelet[2736]: I0306 01:35:09.690183 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-run\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690615 kubelet[2736]: I0306 01:35:09.690196 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hostproc\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690615 kubelet[2736]: I0306 01:35:09.690210 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-etc-cni-netd\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690615 kubelet[2736]: I0306 01:35:09.690225 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-kernel\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690615 kubelet[2736]: I0306 01:35:09.690242 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-net\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690615 kubelet[2736]: I0306 01:35:09.690298 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cni-path\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.690615 kubelet[2736]: I0306 01:35:09.690319 2736 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hubble-tls\") pod \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\" (UID: \"3f417a9c-3f35-4a89-84f3-1d6547c520e8\") " Mar 6 01:35:09.691054 kubelet[2736]: I0306 01:35:09.690349 2736 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-lib-modules\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.693003 kubelet[2736]: I0306 01:35:09.690193 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693003 kubelet[2736]: I0306 01:35:09.692882 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693003 kubelet[2736]: I0306 01:35:09.692964 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693003 kubelet[2736]: I0306 01:35:09.692985 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693003 kubelet[2736]: I0306 01:35:09.693006 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hostproc" (OuterVolumeSpecName: "hostproc") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693213 kubelet[2736]: I0306 01:35:09.693025 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693213 kubelet[2736]: I0306 01:35:09.693044 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693213 kubelet[2736]: I0306 01:35:09.693064 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.693213 kubelet[2736]: I0306 01:35:09.693083 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cni-path" (OuterVolumeSpecName: "cni-path") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 6 01:35:09.696253 kubelet[2736]: I0306 01:35:09.696227 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 6 01:35:09.696494 kubelet[2736]: I0306 01:35:09.696279 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f417a9c-3f35-4a89-84f3-1d6547c520e8-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 6 01:35:09.696916 kubelet[2736]: I0306 01:35:09.696730 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 6 01:35:09.700968 kubelet[2736]: I0306 01:35:09.700907 2736 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-kube-api-access-xrmbn" (OuterVolumeSpecName: "kube-api-access-xrmbn") pod "3f417a9c-3f35-4a89-84f3-1d6547c520e8" (UID: "3f417a9c-3f35-4a89-84f3-1d6547c520e8"). InnerVolumeSpecName "kube-api-access-xrmbn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 6 01:35:09.791618 kubelet[2736]: I0306 01:35:09.791506 2736 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-bpf-maps\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.791618 kubelet[2736]: I0306 01:35:09.791578 2736 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.791618 kubelet[2736]: I0306 01:35:09.791594 2736 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-xtables-lock\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.791618 kubelet[2736]: I0306 01:35:09.791608 2736 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xrmbn\" (UniqueName: \"kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-kube-api-access-xrmbn\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.791618 kubelet[2736]: I0306 01:35:09.791624 2736 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.791618 kubelet[2736]: I0306 01:35:09.791635 2736 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cilium-run\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791647 2736 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hostproc\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791659 2736 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791672 2736 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791683 2736 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791695 2736 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3f417a9c-3f35-4a89-84f3-1d6547c520e8-cni-path\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791706 2736 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3f417a9c-3f35-4a89-84f3-1d6547c520e8-hubble-tls\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.792131 kubelet[2736]: I0306 01:35:09.791717 2736 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3f417a9c-3f35-4a89-84f3-1d6547c520e8-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Mar 6 01:35:09.822285 kubelet[2736]: I0306 01:35:09.822182 2736 scope.go:117] "RemoveContainer" containerID="36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37" Mar 6 01:35:09.825737 containerd[1585]: time="2026-03-06T01:35:09.825617143Z" level=info msg="RemoveContainer for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\"" Mar 6 01:35:09.848537 containerd[1585]: time="2026-03-06T01:35:09.848470691Z" level=info msg="RemoveContainer for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" returns successfully" Mar 6 01:35:09.849124 kubelet[2736]: I0306 01:35:09.849034 2736 scope.go:117] "RemoveContainer" containerID="36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37" Mar 6 01:35:09.849478 containerd[1585]: time="2026-03-06T01:35:09.849403182Z" level=error msg="ContainerStatus for \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\": not found" Mar 6 01:35:09.868529 kubelet[2736]: E0306 01:35:09.867935 2736 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\": not found" containerID="36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37" Mar 6 01:35:09.868529 kubelet[2736]: I0306 01:35:09.868001 2736 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37"} err="failed to get container status \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\": rpc error: code = NotFound desc = an error occurred when try to find container \"36a64d18e1a490b49c485362d9bedd14450879f59496b57ecb0c4ccac9e6ff37\": not found" Mar 6 01:35:09.868529 kubelet[2736]: I0306 01:35:09.868052 2736 scope.go:117] "RemoveContainer" containerID="d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526" Mar 6 01:35:09.871641 containerd[1585]: time="2026-03-06T01:35:09.871472834Z" level=info msg="RemoveContainer for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\"" Mar 6 01:35:09.881130 containerd[1585]: time="2026-03-06T01:35:09.880879756Z" level=info msg="RemoveContainer for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" returns successfully" Mar 6 01:35:09.881629 kubelet[2736]: I0306 01:35:09.881590 2736 scope.go:117] "RemoveContainer" containerID="48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232" Mar 6 01:35:09.884064 containerd[1585]: time="2026-03-06T01:35:09.883964169Z" level=info msg="RemoveContainer for \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\"" Mar 6 01:35:09.895499 containerd[1585]: time="2026-03-06T01:35:09.895359679Z" level=info msg="RemoveContainer for \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\" returns successfully" Mar 6 01:35:09.896104 kubelet[2736]: I0306 01:35:09.895955 2736 scope.go:117] "RemoveContainer" containerID="2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311" Mar 6 01:35:09.897838 containerd[1585]: time="2026-03-06T01:35:09.897632548Z" level=info msg="RemoveContainer for \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\"" Mar 6 01:35:09.908840 containerd[1585]: time="2026-03-06T01:35:09.908675848Z" level=info msg="RemoveContainer for \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\" returns successfully" Mar 6 01:35:09.909471 kubelet[2736]: I0306 01:35:09.909351 2736 scope.go:117] "RemoveContainer" containerID="93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22" Mar 6 01:35:09.912075 containerd[1585]: time="2026-03-06T01:35:09.911911604Z" level=info msg="RemoveContainer for \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\"" Mar 6 01:35:09.919441 containerd[1585]: time="2026-03-06T01:35:09.919383186Z" level=info msg="RemoveContainer for \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\" returns successfully" Mar 6 01:35:09.920536 kubelet[2736]: I0306 01:35:09.919753 2736 scope.go:117] "RemoveContainer" containerID="445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e" Mar 6 01:35:09.922281 containerd[1585]: time="2026-03-06T01:35:09.921948364Z" level=info msg="RemoveContainer for \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\"" Mar 6 01:35:09.929335 containerd[1585]: time="2026-03-06T01:35:09.929198786Z" level=info msg="RemoveContainer for \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\" returns successfully" Mar 6 01:35:09.930963 kubelet[2736]: I0306 01:35:09.930717 2736 scope.go:117] "RemoveContainer" containerID="d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526" Mar 6 01:35:09.932047 containerd[1585]: time="2026-03-06T01:35:09.931917646Z" level=error msg="ContainerStatus for \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\": not found" Mar 6 01:35:09.932586 kubelet[2736]: E0306 01:35:09.932325 2736 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\": not found" containerID="d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526" Mar 6 01:35:09.932586 kubelet[2736]: I0306 01:35:09.932386 2736 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526"} err="failed to get container status \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\": rpc error: code = NotFound desc = an error occurred when try to find container \"d26a33933c59730608b942099c3768650e2e58bd3bb0d130b6660de3bb92c526\": not found" Mar 6 01:35:09.932586 kubelet[2736]: I0306 01:35:09.932419 2736 scope.go:117] "RemoveContainer" containerID="48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232" Mar 6 01:35:09.932934 containerd[1585]: time="2026-03-06T01:35:09.932716894Z" level=error msg="ContainerStatus for \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\": not found" Mar 6 01:35:09.933000 kubelet[2736]: E0306 01:35:09.932942 2736 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\": not found" containerID="48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232" Mar 6 01:35:09.933000 kubelet[2736]: I0306 01:35:09.932971 2736 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232"} err="failed to get container status \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\": rpc error: code = NotFound desc = an error occurred when try to find container \"48c6071740697d14d6ba4bb75c3efc5e1e82e52cc24d011fd31f2eacf1414232\": not found" Mar 6 01:35:09.933000 kubelet[2736]: I0306 01:35:09.932995 2736 scope.go:117] "RemoveContainer" containerID="2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311" Mar 6 01:35:09.933268 containerd[1585]: time="2026-03-06T01:35:09.933214429Z" level=error msg="ContainerStatus for \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\": not found" Mar 6 01:35:09.933453 kubelet[2736]: E0306 01:35:09.933407 2736 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\": not found" containerID="2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311" Mar 6 01:35:09.934200 kubelet[2736]: I0306 01:35:09.933464 2736 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311"} err="failed to get container status \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\": rpc error: code = NotFound desc = an error occurred when try to find container \"2e2f747b8ae76fd3d61b360f5a9c2cd9c78b5d4a9b905df1b33c0fd007911311\": not found" Mar 6 01:35:09.934200 kubelet[2736]: I0306 01:35:09.933491 2736 scope.go:117] "RemoveContainer" containerID="93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22" Mar 6 01:35:09.934200 kubelet[2736]: E0306 01:35:09.933986 2736 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\": not found" containerID="93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22" Mar 6 01:35:09.934200 kubelet[2736]: I0306 01:35:09.934011 2736 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22"} err="failed to get container status \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\": rpc error: code = NotFound desc = an error occurred when try to find container \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\": not found" Mar 6 01:35:09.934200 kubelet[2736]: I0306 01:35:09.934029 2736 scope.go:117] "RemoveContainer" containerID="445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e" Mar 6 01:35:09.934396 kubelet[2736]: E0306 01:35:09.934344 2736 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\": not found" containerID="445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e" Mar 6 01:35:09.934396 kubelet[2736]: I0306 01:35:09.934372 2736 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e"} err="failed to get container status \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\": rpc error: code = NotFound desc = an error occurred when try to find container \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\": not found" Mar 6 01:35:09.934447 containerd[1585]: time="2026-03-06T01:35:09.933658417Z" level=error msg="ContainerStatus for \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"93a6c99cb7462c38f89ccc6fe8fb524c7ca03cbd262f395eea034b75e7856c22\": not found" Mar 6 01:35:09.934447 containerd[1585]: time="2026-03-06T01:35:09.934239559Z" level=error msg="ContainerStatus for \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"445d1ae5878e9faf3565cafc41ac78c16c5439203f3e5461eae199581ae74b2e\": not found" Mar 6 01:35:10.155105 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4-rootfs.mount: Deactivated successfully. Mar 6 01:35:10.156045 systemd[1]: var-lib-kubelet-pods-e2b8bb96\x2d0c8d\x2d467a\x2dace9\x2dfb9604b61c04-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dtmfsc.mount: Deactivated successfully. Mar 6 01:35:10.156390 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058-rootfs.mount: Deactivated successfully. Mar 6 01:35:10.156903 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058-shm.mount: Deactivated successfully. Mar 6 01:35:10.157214 systemd[1]: var-lib-kubelet-pods-3f417a9c\x2d3f35\x2d4a89\x2d84f3\x2d1d6547c520e8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dxrmbn.mount: Deactivated successfully. Mar 6 01:35:10.157465 systemd[1]: var-lib-kubelet-pods-3f417a9c\x2d3f35\x2d4a89\x2d84f3\x2d1d6547c520e8-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 6 01:35:10.157753 systemd[1]: var-lib-kubelet-pods-3f417a9c\x2d3f35\x2d4a89\x2d84f3\x2d1d6547c520e8-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 6 01:35:11.004901 sshd[4452]: pam_unix(sshd:session): session closed for user core Mar 6 01:35:11.013212 systemd[1]: Started sshd@26-10.0.0.56:22-10.0.0.1:59580.service - OpenSSH per-connection server daemon (10.0.0.1:59580). Mar 6 01:35:11.015166 systemd[1]: sshd@25-10.0.0.56:22-10.0.0.1:59566.service: Deactivated successfully. Mar 6 01:35:11.018004 kubelet[2736]: I0306 01:35:11.017395 2736 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f417a9c-3f35-4a89-84f3-1d6547c520e8" path="/var/lib/kubelet/pods/3f417a9c-3f35-4a89-84f3-1d6547c520e8/volumes" Mar 6 01:35:11.018618 systemd[1]: session-26.scope: Deactivated successfully. Mar 6 01:35:11.019448 kubelet[2736]: I0306 01:35:11.019387 2736 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2b8bb96-0c8d-467a-ace9-fb9604b61c04" path="/var/lib/kubelet/pods/e2b8bb96-0c8d-467a-ace9-fb9604b61c04/volumes" Mar 6 01:35:11.020740 systemd-logind[1561]: Session 26 logged out. Waiting for processes to exit. Mar 6 01:35:11.023655 systemd-logind[1561]: Removed session 26. Mar 6 01:35:11.066227 sshd[4620]: Accepted publickey for core from 10.0.0.1 port 59580 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:35:11.068527 sshd[4620]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:35:11.076026 systemd-logind[1561]: New session 27 of user core. Mar 6 01:35:11.081611 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 6 01:35:11.724284 sshd[4620]: pam_unix(sshd:session): session closed for user core Mar 6 01:35:11.743321 systemd[1]: Started sshd@27-10.0.0.56:22-10.0.0.1:59582.service - OpenSSH per-connection server daemon (10.0.0.1:59582). Mar 6 01:35:11.745441 systemd[1]: sshd@26-10.0.0.56:22-10.0.0.1:59580.service: Deactivated successfully. Mar 6 01:35:11.774711 systemd-logind[1561]: Session 27 logged out. Waiting for processes to exit. Mar 6 01:35:11.775743 systemd[1]: session-27.scope: Deactivated successfully. Mar 6 01:35:11.781240 systemd-logind[1561]: Removed session 27. Mar 6 01:35:11.862316 sshd[4634]: Accepted publickey for core from 10.0.0.1 port 59582 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:35:11.865726 sshd[4634]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:35:11.877934 systemd-logind[1561]: New session 28 of user core. Mar 6 01:35:11.884383 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 6 01:35:11.907898 kubelet[2736]: I0306 01:35:11.907683 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-etc-cni-netd\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.907898 kubelet[2736]: I0306 01:35:11.907831 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6f57fdfc-e150-4480-b655-fe48245fcc71-cilium-ipsec-secrets\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908084 kubelet[2736]: I0306 01:35:11.907907 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-cilium-run\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908084 kubelet[2736]: I0306 01:35:11.907932 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6f57fdfc-e150-4480-b655-fe48245fcc71-clustermesh-secrets\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908084 kubelet[2736]: I0306 01:35:11.907956 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv9d8\" (UniqueName: \"kubernetes.io/projected/6f57fdfc-e150-4480-b655-fe48245fcc71-kube-api-access-xv9d8\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908316 kubelet[2736]: I0306 01:35:11.908125 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-hostproc\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908316 kubelet[2736]: I0306 01:35:11.908183 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-cilium-cgroup\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908316 kubelet[2736]: I0306 01:35:11.908200 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-cni-path\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908316 kubelet[2736]: I0306 01:35:11.908215 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6f57fdfc-e150-4480-b655-fe48245fcc71-cilium-config-path\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908316 kubelet[2736]: I0306 01:35:11.908236 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-host-proc-sys-net\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908517 kubelet[2736]: I0306 01:35:11.908284 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-lib-modules\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908517 kubelet[2736]: I0306 01:35:11.908412 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6f57fdfc-e150-4480-b655-fe48245fcc71-hubble-tls\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908517 kubelet[2736]: I0306 01:35:11.908478 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-xtables-lock\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908633 kubelet[2736]: I0306 01:35:11.908543 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-host-proc-sys-kernel\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.908633 kubelet[2736]: I0306 01:35:11.908572 2736 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6f57fdfc-e150-4480-b655-fe48245fcc71-bpf-maps\") pod \"cilium-74zgq\" (UID: \"6f57fdfc-e150-4480-b655-fe48245fcc71\") " pod="kube-system/cilium-74zgq" Mar 6 01:35:11.943004 kubelet[2736]: E0306 01:35:11.940141 2736 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 6 01:35:11.953329 sshd[4634]: pam_unix(sshd:session): session closed for user core Mar 6 01:35:11.972209 systemd[1]: Started sshd@28-10.0.0.56:22-10.0.0.1:59588.service - OpenSSH per-connection server daemon (10.0.0.1:59588). Mar 6 01:35:11.972889 systemd[1]: sshd@27-10.0.0.56:22-10.0.0.1:59582.service: Deactivated successfully. Mar 6 01:35:11.977748 systemd-logind[1561]: Session 28 logged out. Waiting for processes to exit. Mar 6 01:35:11.977912 systemd[1]: session-28.scope: Deactivated successfully. Mar 6 01:35:11.980925 systemd-logind[1561]: Removed session 28. Mar 6 01:35:12.018893 sshd[4645]: Accepted publickey for core from 10.0.0.1 port 59588 ssh2: RSA SHA256:po+n4m2L0Y6JnDj1VTc5p26N9zFlj54R7gCeXzXqR3M Mar 6 01:35:12.022418 sshd[4645]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 6 01:35:12.046737 systemd-logind[1561]: New session 29 of user core. Mar 6 01:35:12.067355 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 6 01:35:12.122595 kubelet[2736]: E0306 01:35:12.122487 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:12.124609 containerd[1585]: time="2026-03-06T01:35:12.124135180Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-74zgq,Uid:6f57fdfc-e150-4480-b655-fe48245fcc71,Namespace:kube-system,Attempt:0,}" Mar 6 01:35:12.179498 containerd[1585]: time="2026-03-06T01:35:12.179150650Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 6 01:35:12.179498 containerd[1585]: time="2026-03-06T01:35:12.179232523Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 6 01:35:12.179498 containerd[1585]: time="2026-03-06T01:35:12.179254975Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:35:12.179498 containerd[1585]: time="2026-03-06T01:35:12.179472020Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 6 01:35:12.286270 containerd[1585]: time="2026-03-06T01:35:12.285984422Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-74zgq,Uid:6f57fdfc-e150-4480-b655-fe48245fcc71,Namespace:kube-system,Attempt:0,} returns sandbox id \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\"" Mar 6 01:35:12.287397 kubelet[2736]: E0306 01:35:12.287275 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:12.298368 containerd[1585]: time="2026-03-06T01:35:12.298094963Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 6 01:35:12.325411 containerd[1585]: time="2026-03-06T01:35:12.325314540Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"ed82c7a95e790c2872c2c5c166beee048c885fb842a6b67c3e03e05aec84b4a0\"" Mar 6 01:35:12.328616 containerd[1585]: time="2026-03-06T01:35:12.326962771Z" level=info msg="StartContainer for \"ed82c7a95e790c2872c2c5c166beee048c885fb842a6b67c3e03e05aec84b4a0\"" Mar 6 01:35:12.455169 containerd[1585]: time="2026-03-06T01:35:12.455091862Z" level=info msg="StartContainer for \"ed82c7a95e790c2872c2c5c166beee048c885fb842a6b67c3e03e05aec84b4a0\" returns successfully" Mar 6 01:35:12.537189 containerd[1585]: time="2026-03-06T01:35:12.536526276Z" level=info msg="shim disconnected" id=ed82c7a95e790c2872c2c5c166beee048c885fb842a6b67c3e03e05aec84b4a0 namespace=k8s.io Mar 6 01:35:12.537189 containerd[1585]: time="2026-03-06T01:35:12.536632494Z" level=warning msg="cleaning up after shim disconnected" id=ed82c7a95e790c2872c2c5c166beee048c885fb842a6b67c3e03e05aec84b4a0 namespace=k8s.io Mar 6 01:35:12.537189 containerd[1585]: time="2026-03-06T01:35:12.536648164Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:12.869314 kubelet[2736]: E0306 01:35:12.869078 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:12.878686 containerd[1585]: time="2026-03-06T01:35:12.878546373Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 6 01:35:12.900673 containerd[1585]: time="2026-03-06T01:35:12.900560738Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"2bb0a426fd92f5a3fab778bb022c78d404db4492385c91b23370a984615e1cf7\"" Mar 6 01:35:12.902083 containerd[1585]: time="2026-03-06T01:35:12.901984199Z" level=info msg="StartContainer for \"2bb0a426fd92f5a3fab778bb022c78d404db4492385c91b23370a984615e1cf7\"" Mar 6 01:35:12.993076 containerd[1585]: time="2026-03-06T01:35:12.993004623Z" level=info msg="StartContainer for \"2bb0a426fd92f5a3fab778bb022c78d404db4492385c91b23370a984615e1cf7\" returns successfully" Mar 6 01:35:13.054004 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2bb0a426fd92f5a3fab778bb022c78d404db4492385c91b23370a984615e1cf7-rootfs.mount: Deactivated successfully. Mar 6 01:35:13.062890 containerd[1585]: time="2026-03-06T01:35:13.062694335Z" level=info msg="shim disconnected" id=2bb0a426fd92f5a3fab778bb022c78d404db4492385c91b23370a984615e1cf7 namespace=k8s.io Mar 6 01:35:13.063080 containerd[1585]: time="2026-03-06T01:35:13.062904928Z" level=warning msg="cleaning up after shim disconnected" id=2bb0a426fd92f5a3fab778bb022c78d404db4492385c91b23370a984615e1cf7 namespace=k8s.io Mar 6 01:35:13.063080 containerd[1585]: time="2026-03-06T01:35:13.062922872Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:13.877986 kubelet[2736]: E0306 01:35:13.875411 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:13.882724 containerd[1585]: time="2026-03-06T01:35:13.882545071Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 6 01:35:13.920565 containerd[1585]: time="2026-03-06T01:35:13.920484140Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"08cea0dd6da8c0dc0c4a7230730771035358d7bd9a6e6aaf90e45c7159146e7e\"" Mar 6 01:35:13.923374 containerd[1585]: time="2026-03-06T01:35:13.921375298Z" level=info msg="StartContainer for \"08cea0dd6da8c0dc0c4a7230730771035358d7bd9a6e6aaf90e45c7159146e7e\"" Mar 6 01:35:14.046411 containerd[1585]: time="2026-03-06T01:35:14.039154274Z" level=info msg="StartContainer for \"08cea0dd6da8c0dc0c4a7230730771035358d7bd9a6e6aaf90e45c7159146e7e\" returns successfully" Mar 6 01:35:14.087151 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-08cea0dd6da8c0dc0c4a7230730771035358d7bd9a6e6aaf90e45c7159146e7e-rootfs.mount: Deactivated successfully. Mar 6 01:35:14.096489 containerd[1585]: time="2026-03-06T01:35:14.096331963Z" level=info msg="shim disconnected" id=08cea0dd6da8c0dc0c4a7230730771035358d7bd9a6e6aaf90e45c7159146e7e namespace=k8s.io Mar 6 01:35:14.096489 containerd[1585]: time="2026-03-06T01:35:14.096408476Z" level=warning msg="cleaning up after shim disconnected" id=08cea0dd6da8c0dc0c4a7230730771035358d7bd9a6e6aaf90e45c7159146e7e namespace=k8s.io Mar 6 01:35:14.096489 containerd[1585]: time="2026-03-06T01:35:14.096429595Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:14.880987 kubelet[2736]: E0306 01:35:14.880914 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:14.886530 containerd[1585]: time="2026-03-06T01:35:14.886286488Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 6 01:35:14.908340 containerd[1585]: time="2026-03-06T01:35:14.908196657Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"adf04e6574e0371bb8fc894afd8c37f3379c7c793b917024f0f87dbed988ee4a\"" Mar 6 01:35:14.914262 containerd[1585]: time="2026-03-06T01:35:14.914130251Z" level=info msg="StartContainer for \"adf04e6574e0371bb8fc894afd8c37f3379c7c793b917024f0f87dbed988ee4a\"" Mar 6 01:35:15.011989 containerd[1585]: time="2026-03-06T01:35:15.011842068Z" level=info msg="StartContainer for \"adf04e6574e0371bb8fc894afd8c37f3379c7c793b917024f0f87dbed988ee4a\" returns successfully" Mar 6 01:35:15.052462 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-adf04e6574e0371bb8fc894afd8c37f3379c7c793b917024f0f87dbed988ee4a-rootfs.mount: Deactivated successfully. Mar 6 01:35:15.060934 containerd[1585]: time="2026-03-06T01:35:15.060603330Z" level=info msg="shim disconnected" id=adf04e6574e0371bb8fc894afd8c37f3379c7c793b917024f0f87dbed988ee4a namespace=k8s.io Mar 6 01:35:15.061213 containerd[1585]: time="2026-03-06T01:35:15.061068609Z" level=warning msg="cleaning up after shim disconnected" id=adf04e6574e0371bb8fc894afd8c37f3379c7c793b917024f0f87dbed988ee4a namespace=k8s.io Mar 6 01:35:15.061213 containerd[1585]: time="2026-03-06T01:35:15.061092023Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 6 01:35:15.887345 kubelet[2736]: E0306 01:35:15.887231 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:15.895647 containerd[1585]: time="2026-03-06T01:35:15.895254413Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 6 01:35:15.928058 containerd[1585]: time="2026-03-06T01:35:15.927531657Z" level=info msg="CreateContainer within sandbox \"5b0e42ae751bc09a8f811299556a1147aef94e9a06221325f555a0e319ed0608\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"4955546a2a363f210113369c9ac8060cd9b11502c108465b327f8d69f81d4ec1\"" Mar 6 01:35:15.935833 containerd[1585]: time="2026-03-06T01:35:15.935109468Z" level=info msg="StartContainer for \"4955546a2a363f210113369c9ac8060cd9b11502c108465b327f8d69f81d4ec1\"" Mar 6 01:35:16.057451 containerd[1585]: time="2026-03-06T01:35:16.057349342Z" level=info msg="StartContainer for \"4955546a2a363f210113369c9ac8060cd9b11502c108465b327f8d69f81d4ec1\" returns successfully" Mar 6 01:35:16.893839 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Mar 6 01:35:16.895666 kubelet[2736]: E0306 01:35:16.895583 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:16.918062 kubelet[2736]: I0306 01:35:16.917580 2736 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-74zgq" podStartSLOduration=5.917564881 podStartE2EDuration="5.917564881s" podCreationTimestamp="2026-03-06 01:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-06 01:35:16.916656682 +0000 UTC m=+118.079303729" watchObservedRunningTime="2026-03-06 01:35:16.917564881 +0000 UTC m=+118.080211899" Mar 6 01:35:18.013208 kubelet[2736]: E0306 01:35:18.013077 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:18.123812 kubelet[2736]: E0306 01:35:18.123638 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:18.988534 containerd[1585]: time="2026-03-06T01:35:18.988491301Z" level=info msg="StopPodSandbox for \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\"" Mar 6 01:35:18.989130 containerd[1585]: time="2026-03-06T01:35:18.988601937Z" level=info msg="TearDown network for sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" successfully" Mar 6 01:35:18.989130 containerd[1585]: time="2026-03-06T01:35:18.988617726Z" level=info msg="StopPodSandbox for \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" returns successfully" Mar 6 01:35:18.989213 containerd[1585]: time="2026-03-06T01:35:18.989177705Z" level=info msg="RemovePodSandbox for \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\"" Mar 6 01:35:18.989213 containerd[1585]: time="2026-03-06T01:35:18.989205587Z" level=info msg="Forcibly stopping sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\"" Mar 6 01:35:18.989316 containerd[1585]: time="2026-03-06T01:35:18.989263385Z" level=info msg="TearDown network for sandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" successfully" Mar 6 01:35:18.993811 containerd[1585]: time="2026-03-06T01:35:18.993659019Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 6 01:35:18.993928 containerd[1585]: time="2026-03-06T01:35:18.993753034Z" level=info msg="RemovePodSandbox \"bf3346eda08b09acee7cfcc1fc6c91c86406509df66a285c66c7d01a1f92d058\" returns successfully" Mar 6 01:35:18.994471 containerd[1585]: time="2026-03-06T01:35:18.994448704Z" level=info msg="StopPodSandbox for \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\"" Mar 6 01:35:18.994471 containerd[1585]: time="2026-03-06T01:35:18.994518464Z" level=info msg="TearDown network for sandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" successfully" Mar 6 01:35:18.994621 containerd[1585]: time="2026-03-06T01:35:18.994530276Z" level=info msg="StopPodSandbox for \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" returns successfully" Mar 6 01:35:18.996666 containerd[1585]: time="2026-03-06T01:35:18.994963888Z" level=info msg="RemovePodSandbox for \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\"" Mar 6 01:35:18.996666 containerd[1585]: time="2026-03-06T01:35:18.994990198Z" level=info msg="Forcibly stopping sandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\"" Mar 6 01:35:18.996666 containerd[1585]: time="2026-03-06T01:35:18.995046753Z" level=info msg="TearDown network for sandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" successfully" Mar 6 01:35:18.999223 containerd[1585]: time="2026-03-06T01:35:18.999088797Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 6 01:35:18.999223 containerd[1585]: time="2026-03-06T01:35:18.999141265Z" level=info msg="RemovePodSandbox \"21116f272330e81f1384da8c2f256772705cce8bb18990b9344f650640dae1c4\" returns successfully" Mar 6 01:35:21.031193 systemd-networkd[1245]: lxc_health: Link UP Mar 6 01:35:21.061063 systemd-networkd[1245]: lxc_health: Gained carrier Mar 6 01:35:22.127894 kubelet[2736]: E0306 01:35:22.126035 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:22.925707 kubelet[2736]: E0306 01:35:22.925353 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:23.017057 systemd-networkd[1245]: lxc_health: Gained IPv6LL Mar 6 01:35:23.924436 kubelet[2736]: E0306 01:35:23.924365 2736 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 6 01:35:25.229046 systemd[1]: run-containerd-runc-k8s.io-4955546a2a363f210113369c9ac8060cd9b11502c108465b327f8d69f81d4ec1-runc.LmE4hT.mount: Deactivated successfully. Mar 6 01:35:25.573487 kubelet[2736]: E0306 01:35:25.572990 2736 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:59330->127.0.0.1:35431: write tcp 127.0.0.1:59330->127.0.0.1:35431: write: broken pipe Mar 6 01:35:27.730236 sshd[4645]: pam_unix(sshd:session): session closed for user core Mar 6 01:35:27.735569 systemd[1]: sshd@28-10.0.0.56:22-10.0.0.1:59588.service: Deactivated successfully. Mar 6 01:35:27.739151 systemd-logind[1561]: Session 29 logged out. Waiting for processes to exit. Mar 6 01:35:27.739198 systemd[1]: session-29.scope: Deactivated successfully. Mar 6 01:35:27.741525 systemd-logind[1561]: Removed session 29.